Martie G. Haselton, Daniel Nettle, and Damian R. Murray
Despite widespread claims to the contrary, the human mind is not worse than rational…but may often be better than rational.
Cosmides and Tooby, 1994, p. 329
On the surface, cognitive biases appear to be somewhat puzzling when viewed through an evolutionary lens. Because they depart from standards of logic and accuracy, they appear to be design flaws instead of examples of good engineering. Cognitive traits can be evaluated according to any number of performance criteria—logical sufficiency, accuracy, speed of processing, and so on. The value of a criterion depends on the question the scientist is asking. To the evolutionary psychologist, however, the evaluative task is not whether the cognitive feature is accurate or logical, but rather how well it solves a particular problem, and how solving this problem contributed to fitness ancestrally. Viewed in this way, if a cognitive bias positively impacted fitness, it is not a design flaw—it is a design feature. This chapter discusses the many biases that are probably not the result of mere constraints on the design of the mind or other mysterious irrationalities, but rather are adaptations that can be studied and better understood from an evolutionary perspective.
By cognitive bias, we mean cases in which human cognition reliably produces representations that are systematically distorted compared to some aspect of objective reality. We note that the term bias is used in the literature in a number of different ways (see, e.g., Johnson, Blumstein, Fowler, & Haselton, 2013; Marshall, Trimmer, Houston, & McNamara, 2013; Nettle & Bateson, 2012). We do not seek to make commitments about these definitions here; rather, we use bias throughout this chapter in the relatively noncommittal sense defined above.
An evolutionary psychological perspective predicts that the mind is equipped with function-specific mechanisms adapted for special purposes—mechanisms with special design for solving problems such as mating, which are separate, at least in part, from those involved in solving problems of food choice, predator avoidance, and social exchange (e.g., Kenrick, Neuberg, Griskevicius, Becker, & Schaller, 2010). In the evaluation of cognitive biases, demonstrating domain specificity in solving a particular problem is a part of building a case that the trait has been shaped by selection to perform that function. The evolved function of the eye, for instance, is to facilitate sight because it does this well (it exhibits proficiency), the features of the eye have the common and unique effect of facilitating sight (it exhibits specificity), and there are no plausible alternative hypotheses that account for the eye's features.
Some design features that appear to be flaws when viewed in one way are revealed to be adaptations when viewed differently. If one were to only consider the idea that selection favors the maximization of direct reproductive success, for example, the fact that human females lose reproductive capability many years before death would appear a design flaw. However, there is evidence that women in traditional societies can enhance their inclusive fitness by transferring investment to their daughters' daughters as soon as the latter are of reproductive age (Voland & Beise, 2002). Viewed in this light, female menopause might be very well designed (Hawkes, 2003).
In sum, there may be many evolutionary reasons for apparent design flaws, and a close examination often provides insight into the evolutionary forces that shaped them and their functions. Analogous logic may be applied to understanding cognitive biases. We argue that cognitive biases can arise for at least three reasons (see Table 41.1).
Table 41.1 Evolutionary Taxonomy of Cognitive Biases
Type of Bias | Examples |
Artifact: Apparent biases and errors are artifacts of research strategies; they result from the application of inappropriate normative standards or placement of humans in unnatural settings. | Some instances of base-rate neglect in statistical prediction Some instances of confirmation bias |
Error Management Bias: Selection favored bias toward the less costly error; although error rates are increased, net costs are reduced. | Auditory looming Xenophobia Sexual overperception by men Commitment underperception by women |
First, selection may favor useful shortcuts that tend to work in most circumstances, though they fall short of some normative standards (heuristics); second, apparent biases can arise if the task at hand is not one for which the mind is designed (artifacts); and third, biases can arise if biased response patterns to adaptive problems resulted in lower error costs than unbiased response patterns (error management biases). As well as being interesting in their own right, the investigation of biases offers the capacity to reveal the contours of the evolved mind by revealing the problems it appears to have been designed to solve: Whereas “accurate” perceptions do little to constrain hypotheses about cognitive design, discovering bias can often reveal it.
Since the original edition of this Handbook, the volume of work investigating error management biases has grown rapidly. Therefore, we discuss heuristics and artifacts only briefly and focus on newer work on error management biases (for a more detailed evolutionary discussion of heuristics and artifacts, see Haselton et al., 2009). We do not intend the three categories of bias to be fully exhaustive or mutually exclusive; we offer them instead as a useful way of organizing research on cognitive bias and gaining insight into why biases exist.
Perhaps the most commonly invoked explanation for bias is as a necessary by-product of processing limitations—because information processing time and ability are limited, humans must use shortcuts or rules of thumb that are prone to breakdown in systematic ways. Kahneman and Tversky (1973) demonstrated that human judgments often departed substantially from normative standards based on probability theory or simple logic. In judging the sequences of coin flips, for example, people assessed the sequence HTHTTH to be more likely than the sequence HHHTTT or HHHHTH. As Tversky and Kahneman (1974) pointed out, while in some sense representative, the first type of sequence is improbable—it contains too many alternations and too few runs. The “gambler's fallacy” is the expression of a similar intuition. The more bets lost, the more the gambler feels a win is now due, even though each new turn is independent of the last (Tversky & Kahneman, 1974).
Tversky and Kahneman attributed these and other biases to the operation of mental shortcuts: “People rely on a limited number of heuristic principles which reduce the complex tasks of assessing probabilities and predicting values to simpler judgmental operations” (1974, p. 1124). The gambler's fallacy and the conjunction fallacy are attributed to one of the most commonly invoked heuristics, representativeness, or the way in which A resembles or is representative of B. According to this account, alternating heads and tails are more representative of randomness than are series containing runs.
The notion that biases result from the use of simplifying heuristics has logical appeal. As expressed by Arkes (1991), “the extra effort required to use a more sophisticated strategy is a cost that often outweighs the potential benefit of enhanced accuracy” (pp. 486–487). This cost can affect the evolution of cognitive mechanisms at two levels. There may be costs in evolutionary terms, since the development of certain brain circuits will either increase the length of ontogeny or move potential energetic allocation away from the development of other mechanisms. There may also be costs in real time, since decisions using complex algorithms will often take longer or require more attentional resources than decisions using simpler alternatives. Adaptive decisions often need to be made fast, and this may well constrain the type of strategies that are optimal. Evidence from a variety of sources demonstrates that people do indeed solve problems differently when under time pressure or when their motivations to be accurate are reduced.
One example of the effects of motivation is the fact that the social perceptions of individuals occupying positions of higher power in social hierarchies are often less accurate than those lower in the hierarchy (Fiske, 1993). Those higher in power are more likely to endorse stereotypes about others than to attend to individuating information specific to the target being evaluated, which presumably enhances accuracy (Goodwin, Gubin, Fiske, & Yzerbyt, 2000). Individuals assigned more decision-making power in reviewing internship applications attend more to stereotype-consistent information and less to stereotype-inconsistent information (Goodwin et al., 2000). Similarly, in a study of two student groups competing for university funding, individuals reporting more personal power judged their opponents' attitudes less accurately (Ebenbach & Keltner, 1998). A common interpretation of findings such as these is that lower-power individuals occupy a more precarious social position and they must therefore allocate more time and energy to social judgments; more powerful individuals enjoy the luxury of allocating their cognitive efforts elsewhere (Galinsky, Magee, Inesi, & Gruenfeld, 2006; Keltner, Gruenfeld, & Anderson, 2003).
Overall, there is ample evidence of cognitive bias and error in humans. Some of these biases may result from the use of shortcuts, which are often effective. For these effects, however, it is important to note that a “processing limitations” explanation is not complete. Of all possible equally economical cognitive shortcuts, why were these particular ones favored by selection? In the error management biases section that follows, we suggest that the direction and content of biases is not arbitrary. Selection has sculpted the ways that limited computational power is deployed so as best to serve the fitness interests of humans over evolutionary time.
One criticism of classic heuristics and biases research (e.g., Tversky & Kahneman, 1974) is that the strategies for identifying bias and evaluating cognitive performance might not be fully appropriate. If problems presented in the laboratory are not those for which the human mind is designed, we should not be surprised that people's responses appear to be systematically irrational.
One type of artifact arises from evolutionarily novel problem formats. Gigerenzer (1997) proposed that tasks intended to assess human statistical prediction should present information in frequency (rather than probability) format, given that natural frequencies, such as the number of times an event has occurred in a given time period, are more readily observable in nature. In contrast, probabilities (in the sense of a number between 0 and 1) are mathematical abstractions beyond sensory input data, and information about the base rates of occurrence is lost when probabilities are computed (Cosmides & Tooby, 1996). Bayesian calculations involving frequencies are therefore computationally simpler than equivalent calculations involving probabilities, relative frequencies, or percentages. Whereas probability calculations need to reintroduce information about base rates, frequency calculations do not since this part of the computation is already “done” within the frequency representation itself (Hoffrage, Lindsey, Hertwig, & Gigerenzer, 2001).
According to this perspective, humans will possess the ability to estimate the likelihood of events given certain cues. If this skill is a part of human reasoning, however, tasks involving probability input are less likely to reveal it than are tasks involving natural frequencies. Indeed, frequency formats do improve performance in tasks like the famous “Linda problem.” Whereas a probability format produces violations of the conjunction rule in between 50 and 90% of respondents, frequency formats decrease the rate of error to between 0 and 25% (Fiedler, 1988; Hertwig & Gigerenzer, 1999; Tversky & Kahneman, 1983; but see Mellers, Hertwig, & Kahneman, 2001). More recent research suggests that probability formats pose serious problems for medical doctors: Three quarters of doctors surveyed misinterpreted the meaning and application of “survival rates,” and journals frequently publish papers in which these probability statistics are misused in interpreting results (Gigerenzer & Wegwarth, 2013).
A second artifact can arise from evolutionarily novel problem content. The perspective on cognitive design we have described suggests that researchers should not necessarily expect good performance in tasks involving abstract rules of logic. Falsification-based logic is sufficiently difficult for humans that university courses in logic, statistics, and research design attempt to teach it to students (with only mixed success). Wason (1983) empirically confirmed this in the laboratory using a task that required subjects to determine whether a conditional rule (if p then q) had been broken. He demonstrated that subjects recognized that confirmatory evidence (the presence of p) was relevant to the decision, but they often failed to check for falsifications of the rule (the absence of q). Research using the Wason task revealed a variety of apparent content effects (Wason & Shapiro, 1971; Johnson-Laird, Legrenzi, & Legrenzi, 1972), in which performance dramatically changed for the better.
In a series of now-classic experiments, Cosmides (1989) demonstrated that a number of the content effects could be attributed to a cheater-detection algorithm. When the content of the conditional rule involves social exchange (if you take the benefit [p], then you pay the cost [q]), people are spontaneously induced to look not only for benefits taken (p) but also costs not paid (not q), and performance dramatically increases from 25% correct (Wason, 1983) to 75% correct (Cosmides, 1989; also see Cosmides, Barrett, & Tooby, 2010, for a more recent update that replicates these findings and helps to rule out alternative explanations proposed by critics).
The conclusion to be drawn from these studies is not that humans are good at using abstract rules of logic. Rather, it is that humans have evolved problem-solving mechanisms tailored to problems recurrently present over evolutionary history. When problems are framed in ways congruent with these adaptive problems (such as social contract violation), humans can be shown to use appropriate reasoning strategies.
Like biases resulting from the application of heuristics, biases in this third set—error management biases—are genuine biases. In this case, however, biases are not the result of shortcuts in the design of the mind. Instead, the biases themselves serve evolved functions.
Error management theory (EMT; Haselton & Buss, 2000; Haselton & Nettle, 2006; Johnson et al., 2013) applies the principles of signal detection theory (Green & Swets, 1966) to judgment tasks in order to make predictions about evolved cognitive design. An error management framework views cognitive mechanisms not so much as “truth seekers” (as has been previously thought; e.g., Fodor, 2001), but as adaptation executors (e.g., Tooby & Cosmides, 1990). The central tenet of this framework is that cognitive mechanisms can generally produce two types of errors: false positives (taking an action that would have been better not to take), and false negatives (failing to take an action that would have been better to take).
An optimal mechanism would make no errors of either type. However, most real-world judgment tasks are probabilistic and include an irreducible amount of uncertainty. Auditory judgment, for example, is rendered uncertain by the presence of ambient noise, and some error is likely to occur however good the mechanism.
Crucially, the fitness costs of making each type of error are rarely equal. Fleeing from an area that contains no predator results in a small inconvenience cost, but it is much less costly than the failure to flee from a predator that really is close by. EMT predicts that an optimal decision rule will minimize not the total error rate, but the net effect of error on fitness. Where one error is consistently more damaging to fitness than the other, EMT predicts that a bias toward making the less costly error will evolve—this is because it is better to make more errors overall as long as they are relatively cheap. Overall, then, EMT predicts that biases will evolve in human judgments and evaluations that fit all of the following criteria: (a) they involve some degree of noise or uncertainty, (b) they have consequences for fitness and reproductive success, and (c) they are consistently associated with asymmetrical costs (where more asymmetry leads to larger biases). For mathematical formalism of this logic and the expectations of EMT, see Haselton and Nettle (2006) and Johnson et al. (2013). (For a related account, see Higgins, 1997.)
Within this framework, many ostensible faults in human judgment and evaluation may reflect the operation of mechanisms designed to make inexpensive, frequent errors rather than occasional disastrous ones (Haselton & Nettle, 2006; Johnson et al., 2013). In the decade since the publication of the first edition of this volume, the scope of EMT research has expanded, with streams of research documenting functionally biased judgments across a variety of fitness-relevant domains. In this section, we highlight key examples across these domains (for reviews containing additional examples, see Haselton & Galperin, 2013; Haselton et al., 2009; Haselton & Nettle, 2006; Johnson et al., 2013).
Error management biases can be generally sorted into three broad categories: biases pertaining to judgments of threat, biases pertaining to evaluations of interpersonal relationships, and biases pertaining to evaluations of the self (following Haselton & Nettle, 2006). Table 41.2 provides examples within each of these categories, the hypothesized costs of each type of error in a given domain, and the expected outcome for each.
Table 41.2 A Selection of Adaptive Biases
Category and Domain | False Positive (FP) | Costs of FP | False Negative (FN) | Costs of FN | Result |
Protective: Approaching Sounds | Ready too early | Low | Struck by source | High | Bias toward underestimating time to arrival |
Protective: Foods | Reject a food type that is in fact safe | Low | Ingest toxin or pathogen | High | Bias toward acquiring permanent aversion on the basis of one piece of evidence of toxicity |
Protective: Diseased persons | Avoid noninfectious person | Usually low | Become infected | Often very high | Tendency to avoid persons with physical afflictions, even if noninfectious |
Protective: Physically threatening persons | Avoid altercation with safe person | Usually low | Suffer physical injury | Often high | Tendency to overestimate physical formidability of potentially threatening persons |
Social: Men's inference of female sexual interest | Inferring sexual interest where there is none | Rejection—relatively low | Inferring no interest when there is interest | Missed reproductive opportunity—high | Sexual overperception by men |
Social: Women's inference of commitment | Inferring interest to commit where there is none | Desertion—high | Inferring unwillingness to commit where there is willingness | Delayed start to reproduction—relatively low | Underperception of commitment by women |
Social: Social exchange | Attempt to free-ride and get caught | Potential ostracism, especially in collectivist social situations—high | Cooperate when one could free-ride | Give up a unnecessary benefit in exchange—relatively low | Bias toward cooperation |
Self and Future: Beliefs about future achievements | Believe you can achieve things when you cannot | Low (if costs of failure are low) | Believe you cannot achieve things when, in fact, you could | High (if benefit of success is high) | Optimistic bias (where benefits of success exceed costs of failure); overconfidence bias |
Several biases might protect people from threats to physical safety or to health. We begin by discussing the former.
People tend to judge a sound that is rising in intensity to be closer and approaching more rapidly than an equidistant sound that is falling in intensity. In a series of experiments involving speakers moving on cables, people showed biased perceptions of the proximity of moving sound sources, as well as a general tendency to underestimate the distance of sound sources (Neuhoff, 2001). People judged an approaching sound source to be closer by than a receding one, when in fact the sounds were located at distances equally far away. There is a clear error management interpretation of this effect: It is better to be ready for an approaching object too early than too late (Neuhoff, 2001).
Recent work has shown that individuals in poorer physical condition—measured by both heart rate recovery time and physical strength—have larger auditory looming biases than individuals in better physical condition (Neuhoff, Long, & Worthington, 2012). An error management interpretation of this relationship is that individuals with reduced motor capacity require a larger “margin of safety.” In another recent study, people exposed to an infant cry showed larger auditory looming biases than people not exposed to the cry. And, conversely, people exposed to an infant laugh showed smaller auditory looming biases than people not exposed to the laugh (Neuhoff, Hamilton, Gittleson, & Mejia, 2014). Female participants showed larger shifts in auditory looming biases in response to these infant stimuli—a pattern also found in a follow-up study (Neuhoff et al., 2014). Because throughout evolutionary history infants likely required direct care from mothers more so than from fathers (e.g., due to breastfeeding), these effects suggest that self-protective biases like auditory looming are tuned to threats associated with care for vulnerable offspring. At a more general level, adaptively patterned variation in auditory looming demonstrates that error management biases are not fixed, but are responsive to cues of variation in threat.
Might there be analogous phenomena in the perception of visual threats? In one recent study, people judged the speed of an approaching spider, ladybug, or rubber ball (Witt & Sugovic, 2013). Although all objects moved at the same speed, spiders were judged to be moving more quickly than the other objects. Further, when people were given the task of “blocking” the spider, they judged the spider as approaching them faster when they used a smaller paddle relative to a larger paddle, demonstrating that the bias was enhanced when avoiding the spider was more difficult.
There are asymmetric costs of injury from underestimating the height of a cliff, and perhaps erroneously judging it safe to jump, than from overestimating it and finding a different means of navigation. Consistent with this idea, people tend to judge the height of a vertical surface as greater when looking from the top than from the bottom (Jackson & Cormack, 2007; Stefanucci & Proffitt, 2009).
A similar example involves the perceived steepness of hills. In one series of studies, people consistently overestimated the steepness of hills—both real and computer-simulated (Proffitt, Bhalla, Gossweiler, & Midgett, 1995). Failing to properly descend a steep hill is far more costly than failing to properly ascend it. An error management perspective therefore predicts that this bias toward overestimating slopes will be greater when people view from the top than from the bottom, which was exactly what is found (Proffitt et al., 1995). Making the situation even more precarious increases the bias even more—people standing on skateboards at the top of hills perceive greater steepness than those standing on flat earth (Stefanucci, Proffitt, Clore, & Parekh, 2008).
Lasting aversion to a food is reliably acquired, in humans and other species, following a single incidence of sickness after ingestion of the food (Garcia, Hankins, & Rusiniak, 1976; Rozin & Kalat, 1971). Given one data point (sickness following the food type on one occasion), the system treats the food as if it is always illness-inducing. There are again two possible errors here. The false positive may be inconvenient, but the false negative is more likely to be fatal. The system appears biased toward overresponsiveness to avoid illness.
Similar logic predicts an aversion to individuals who have superficial cues that might connote the threat of infectious disease. The error management account is similar to that for food aversions: The false negative (failing to avoid someone with a contagious disease) is highly costly, whereas the false positive (avoiding contact with a noncontagious person) may have small social or interpersonal costs, but is unlikely to have significant negative fitness consequences. Given the fact that infectious disease has represented one of the key selective forces throughout human history (e.g., Inhorn & Brown, 1990), disease avoidance mechanisms are often expected to be biased to avoid many individuals or objects that are in fact safe.
The significant bias toward false positives in assessing cues of disease threat has far-reaching social and societal implications, and may lie at the root of many forms of stigmatization and prejudice, including racism, ageism, homophobia, and anti-fat prejudice (e.g., Kurzban & Leary, 2001). Hypersensitivity to disease threat leads to stigmatization or avoidance of individuals who pose no risk of disease transmission whatsoever, yet display cues that were associated with disease threat ancestrally. Individuals with noninfectious morphological anomalies, such as prominent birthmarks, activate avoidant responses (Zebrowitz & Montepare, 2006). Similarly, individuals with clearly noninfectious physical disabilities are also implicitly associated with disease (Park, Faulkner, & Schaller, 2003), as are obese individuals (Park, Schaller, & Crandall, 2007). Individuals displaying symptoms of HIV/AIDS are also implicitly associated with the threat of infectious disease, despite knowledge that this ailment is not infectious through superficial contact. These individuals are frequently regarded as disgusting (e.g., Herek, 1999), and they, along with their families, are often ostracized from their communities (Gerbert, Sumser, & Maguire, 1991). Other patently noninfectious afflictions that result in social distancing include cancer (Greene & Banerjee, 2006) and physical disfigurements (Houston & Bull, 1994).
The strength of these implicit associations is predicted by the extent to which individuals perceive themselves to be vulnerable to infectious disease. Individuals who tend to be more worried about disease threat have stronger implicit associations between infectious disease and both obesity and physical disability, and also have more negative attitudes toward obese and physically disabled people (Lieberman, Tybur, & Latner, 2012; Park et al., 2003; Park et al., 2007). Moreover, making a threat of disease temporarily salient amplifies these prejudicial cognitions (Park et al., 2003; Park et al., 2007). Other evidence suggests that prejudicial cognitions regarding elderly people are greater among people who feel more chronically vulnerable to disease (Duncan & Schaller, 2009). Studies have also documented links between perceived vulnerability to disease and overperceptions of unusual morphological features. For example, individuals higher in disease concerns set a lower threshold for categorizing someone as obese, and situationally priming disease threat leads to over-remembering seeing obese targets (Miller & Maner, 2012).
Members of other cultural groups may also be implicitly associated with disease threat. Human immune systems are attuned to local disease threats. Contact with unfamiliar outgroups might have historically increased the risk of contracting dangerous pathogens unfamiliar to locally adapted immune systems (Diamond, 1999). An error management perspective predicts that the benefits of exaggerated avoidance of outgroup members (e.g., xenophobia) may have historically outweighed its costs (Kurzban & Leary, 2001). Indeed, individuals who tend to be particularly worried about infectious disease tend to hold more negative attitudes toward unfamiliar ethnic groups (Faulkner, Schaller, Park & Duncan, 2004), and making the threat of disease temporarily salient increases opposition toward policies allowing immigration of unfamiliar outgroups (Faulkner et al., 2004). Ethnocentric and xenophobic attitudes are also higher for women during the first trimester of pregnancy, when the immune system is temporarily compromised (Navarrete, Fessler, & Eng, 2007). Cross-culturally, individuals in countries with higher levels of infectious disease are more likely to report that they would not want “people of a different race” as neighbors (Schaller & Murray, 2010).
Infectious disease is not the only threat posed by others, particularly for individuals who are physically vulnerable. A recent series of studies manipulated or measured vulnerability to harm and showed that vulnerable individuals overestimated the formidability of potentially threatening individuals. One study found that when people were told that a man was holding a gun, they perceived that person to be taller and more muscular than when they were told he was holding a drill, handsaw, or caulking gun (Fessler, Holbrook, & Snyder, 2012). Similarly, men who were temporarily physically incapacitated (either by being bound to a chair or by standing on a balance board) estimated an image of an angry man to be significantly taller and more muscular than did men who were not incapacitated (Fessler & Holbrook, 2013a). The presence of weapons also appears to influence dispositional judgments: Men who were pictured holding potentially harmful tools in nonviolent situations (such as gardening shears) were judged to be more anger prone than when pictured holding innocuous tools (such as a watering can; Holbrook et al., 2014). In another study, parents with dependent children perceived a potentially threatening criminal to be more physically formidable than did nonparents (Fessler, Holbrook, Pollack, & Hahn-Holbrook, 2014).
Other situational variables can make potentially threatening individuals appear less physically formidable: Men who were in the presence of companions judged a solitary foe as smaller and less muscular than men who made these judgments alone (Fessler & Holbrook, 2013b).
A second cluster of error management biases concern our perceptions of the intentions or dispositions of others.
Courtship communications are often ambiguous. Does a smile convey mere friendliness, or does it mean more? For men, error management logic predicts a bias toward overestimating a potential mate's sexual interest. This is because, all else equal, the reproductive costs of underestimating a woman's sexual interest and failing to pursue her—thereby missing out on an opportunity to reproduce—were likely to have been greater than the costs of pursuing a disinterested woman (Haselton & Buss, 2000). Men who were more successful in mating with greater numbers of women would have outreproduced other men, passing along this possible overperception bias to their descendants.
For women, a different logic applies. Because of women's necessarily heavy investment in each child produced, and the necessarily long interval between births, finding high-quality partners—not more numerous partners—probably had a greater impact on women's reproductive success (Buss, 1994; Symons, 1979; Trivers, 1972). Therefore, error management logic predicts that men, but not women, possess the sexual overperception bias. Many sources of evidence support the sexual overperception hypothesis (see Haselton & Galperin, 2013, Table 11.1, for a review). For instance, in the earliest demonstration of the phenomenon, male and female strangers engaged in a get-to-know-you conversation in the lab and were viewed by a second pair of male and female strangers through one-way glass (Abbey, 1982). Both the male participant in the conversation (target) and the male observer rated the female target as more flirtatious and sexually interested than the female observer and female target. In this study and similar later studies, the difference between male and female ratings of women's sexual interest was present when men's ratings were compared to the target woman's self-ratings and when compared to ratings made by third-party women assessing the interaction (Abbey, 1982; Haselton & Buss, 2000).
Similar results are found in surveys of men's and women's misperception experiences. In one study, for example, undergraduate women from the United States reported more instances within the last year in which men overestimated their sexual interest than in which men underestimated it, suggesting that these men overperceived women's sexual interest in naturalistic situations outside of the lab (Haselton, 2003). Men in the same study reported roughly equal numbers of overperception and underperception errors on the part of women, providing no evidence of a bias in women. These patterns were closely replicated in a study of Norwegian undergraduates—a replication that is particularly noteworthy because Norway has a more gender-equal culture than does the United States (Bendixen, 2014). In related studies of opposite-sex friendships, men estimated their female friends' sexual interest to be greater than those women reported it to be (Bleske-Rechek et al., 2012; Koenig, Kirkpatrick, & Ketelaar, 2007). Similarly, a recent speed-dating study similarly found that men estimated greater sexual interest in their female partners than their female partners reported (Perilloux, Easton, & Buss, 2012). This study also assessed variation in this apparent bias and found that men who were higher in short-term mating orientation and who were higher in self-rated attractiveness had a larger bias. Further, men's apparent bias was greater when they interacted with relatively more attractive women. In some birds, insects, and mammals (Alcock, 1993, Chapter 13; Domjan, Huber-McDonald, & Holloway, 1992), males sometimes attempt to copulate with objects that only vaguely resemble females of their species, such as beer bottles or crude female models, suggesting similar behavioral biases in other species.
The reverse asymmetry might have applied to ancestral women as they estimated men's intentions to commit to long-term relationships (Haselton & Buss, 2000). Inferring long-term commitment interest in a man in whom it was absent could have resulted in abandonment after the woman had already conceived a child, a high-cost error potentially associated with reduced offspring survival (e.g., Hurtado & Hill, 1992). Underestimating a man's commitment could also be costly, including delays in reproduction, but these costs might have been lower on average than costs associated with desertion (Haselton & Buss, 2000). Women might therefore possess a bias toward underestimating men's interest in commitment. Consistent with this idea, several studies have shown that women rate men's commitment given various courtship behaviors, such as giving gifts and verbal affirmations of love, lower than men rate it (Haselton & Buss, 2000). In contrast, women and men tend to agree on the level of commitment indicated by women's enaction of the same behaviors (Haselton & Buss, 2000). A recent study similar to earlier research documented apparent commitment skepticism in women who were prior to the age of menopause, but not women past the age of menopause, possibly because women past the age of reproduction would not have faced the same reproductive costs of overestimating men's commitment (Cyrus, Schwarz, & Hassebrauck, 2011).
Further evidence for commitment underperception in women was found in a study of face-to-face interactions between previously unacquainted male–female dyads (Henningsen & Henningsen, 2010). Dyads engaged in a 5-minute conversation and afterwards filled out questionnaires about their own and their partner's perceived level of interest in a committed long-term relationship. Consistent with the commitment underperception hypothesis, women estimated lower levels of commitment interest than men reported for themselves. In contrast, men's estimates of women's commitment were not significantly different from women's reports of their commitment interest, providing no evidence of bias in men's judgments of women.
Humans appear to possess a bias toward inferring that members of competing coalitions (or outgroups) are less generous and kind (Brewer, 1979) and more dangerous and ill-tempered (Quillian & Pager, 2001) than are members of their own group. This bias might have been adaptive for reasons that extend beyond those related to the threat of disease transmission, presented above. For ancestral humans, the costs of falsely assuming peacefulness on the part of an aggressor were likely to outweigh the comparatively low costs of elevated vigilance about aggression, particularly for inferences about outgroup members who are not part of an individual's regular social circle. In one study, when participants were exposed to a cue indicating increased risk of injury—ambient darkness in the laboratory—they endorsed racial and ethnic stereotypes connoting violence more so than people completing the task in a brightly lit room (Schaller, Park, & Mueller, 2003). Darkness had no effect on other negative stereotypes of outgroup others (such as laziness or ignorance) (Schaller et al., 2003; for a related recent study see Stroessner, Scholer, & Marx, 2015).
Behavioral economists have puzzled over the fact that people cooperate in economic games with economic incentive structures favoring defection (Camerer & Thaler, 1995; Caporael, Dawes, Orbell, & van der Kragt, 1989; Henrich et al., 2001; Sally, 1994). In the one-shot Prisoner's Dilemma game, for example, participants are expected to defect rather than to cooperate. If partner A cooperates while B defects, partner A suffers a greater loss than if he or she had defected. The interaction is not repeated, so there is no incentive to signal cooperativeness, nor is there prior information about reputation that might serve to provide clues about the partner's cooperative disposition. Yet cooperation often occurs, as it does in other one-shot economic tasks.
One possibility is that cooperation in one-shot games results from the operation of a social exchange bias that manages the costs of errors in social exchange (Yamagishi, Terai, Kiyonari, Mifune, & Kanazawa, 2007). According to this logic, the costs of falsely believing one can defect without negative social consequences are often higher than cooperating when one could safely defect. This asymmetry holds when the costs of “unneeded” cooperation are relatively low (e.g., a low dollar amount is lost) or when the social costs of failing to cooperate (potential ostracism) are high. The costs of ostracism may be particularly high in interdependent social contexts, in which cooperation is either highly valued or especially necessary (Yamagishi, Jin, & Kiyonari, 1999). In Japanese collectivist samples where exchanges are relatively closed to outsiders, cooperation in one-shot experiments is higher than in the more individualist United States samples (Yamagishi et al., 1999). Also consistent with the social exchange bias hypothesis, when people are led to think of the game as an exchange relationship (by making forecasts about their exchange partner's behavior) they cooperate more than when they are not led to think this (Yamagishi et al., 2007; see also Savitsky, Epley, & Gilovich, 2001, and Williams, Case, & Govan, 2003, for related predictions). Similar predictions were tested using evolutionary modeling and showed that one-shot cooperation can evolve due to the asymmetric costs of sometimes mistaking a repeated interaction for one-shot relative to mistaking one-shot for repeated interaction (Delton, Krasnow, Cosmides, & Tooby, 2011).
Note that this bias can be conceptualized as some combination of error management, as in the social exchange bias account, and an artifact of modern living, since in an ancestral environment the probability of re-encountering individuals would have been high and social reputation effects very potent. Thus, people may be predisposed to expect negative consequences of non-prosocial behavior even when, objectively, such consequences are unlikely to follow. The bias toward prosociality has been the subject of competing explanations that take quite different explanatory stances (Bowles & Gintis, 2002; Gintis, Bowles, Boyd, & Fehr, 2003; Henrich & Boyd, 2001; Price, Cosmides, & Tooby, 2002), although these explanations might not be mutually exclusive.
The third cluster of biases concerns judgment about the self and personal efficacy. Here we briefly discuss the representative example of the “positive illusions” (for a more complete review, see Haselton & Nettle, 2006).
These are a well-known cluster of findings in judgment tasks concerning the self (Taylor & Brown, 1988). Individuals display unrealistically positive perceptions of their own qualities (Alicke, 1985), their likelihood of achieving positive outcomes in the future (Weinstein, 1980), and their degree of control over processes in the environment (Alloy & Abramson, 1979; Rudski, 2000). Two classes of evolutionary explanation have been proposed for such tendencies. One explanation is that individuals may have been selected to optimize the impression of their qualities that they display to observers. Given that observers will not be able to accurately assess such qualities directly, individuals may display behaviors that strategically enhance the qualities conveyed (Sedikides, Gaertner, & Toguchi, 2003).
An alternative explanation involves error management. Nettle (2004) outlines such an explanation, building upon the interpretation of the positive illusions given by Taylor and Brown (1988). In evaluating a prospective course of action, there are two possible errors. One may judge that the action is worthwhile when in fact it achieves nothing to promote fitness (or would not have ancestrally), or judge that a behavior is not worthwhile when in fact it enhances fitness to do it (or would have ancestrally). The former error (a false positive) leads to behaviors that are actually useless, whereas the latter (a false negative) leads to passivity. The costs of the false positive and false negative errors may not be symmetrical—that is, trying and failing may not matter very much, whereas failing to try could be very costly, especially in competitive contexts. Thus, evolution can be expected to produce mechanisms biased toward positive illusion in domains where there is uncertainty about outcomes, and the cost of trying and failing is reliably less than that of not trying where success was possible (Nettle, 2004). Recent neuroscientific research suggests that these biases have deep cognitive roots: Individuals tend to encode undesirable information in a distorted manner, which leads to the relative enhancement of desirable information (Sharot, Korn, & Dolan, 2011; Sharot, Riccardi, Raio, & Phelps, 2007). Note that the error management account does not predict blanket optimism, but optimism where fitness gains were potentially high relative to the cost of passivity.
A similar argument can be made for a different type of positive illusion: overconfidence. Although overconfidence can sometimes lead to costly decisions and behaviors, its motivational benefits—in the form of increased ambition and persistence—might outweigh these costs. Evolutionary models are consistent with the notion that biased representations of personal success probabilities can be favored under certain circumstances (Johnson & Fowler, 2011; but see Johnson & Fowler, 2013, and Marshall et al., 2013).
Other evolutionary models suggest that nation states that are overconfident in warlike behaviors are more likely to be successful than accurate or underconfident states (Johnson, Weidmann, & Cederman, 2011).
Research on cognitive and social bias has been dominated by the failure and bleak implications of heuristics (see Krueger & Funder, 2004). A Newsweek magazine account of the heuristics and biases literature summarized it as showing that “most people…are woefully muddled information processors who often stumble on ill-chosen short-cuts to reach bad conclusions” (cited in Gigerenzer, Todd, & the ABC Research Group, 1999, p. 27). In reflecting back on the history of social psychology, Aronson (1999) noted that “odious behavior (‘sin’) is at the heart of [the] most powerful research in social psychology” (p. 104). Browsing journals in social psychology, behavioral economics, and social cognition reveals a proliferation of seemingly foolish bias effects (see Haselton et al., 2009; Krueger & Funder, 2004).
Adopting an evolutionary perspective turns this focus on its head. Natural selection is the force responsible for creating the intricate designs with an improbably perfect match to their environments. Complex visual systems with specialized features tailored to species' differing ecologies have evolved several times, independently (Goldsmith, 1990). Reproductive adaptations allow animals to reproduce small copies of themselves, developmentally intact, complete with miniature versions of the adaptations that will enable their own reproduction. Natural selection is similarly responsible for the intricacy of the human mind. How could natural selection produce systems that equip the brain that are prone to fail as a rule and succeed only in exceptional cases?
The conceptual tide has now turned. There has been a shift toward explanations for bias invoking adaptive function, as well as a demonstration that simple mechanisms (heuristics) can function well in their proper domains. This reconceptualization has stimulated new developments in psychological theory and empirical research. Documenting content effects in biases—where bias effects emerge, recede, or reverse depending on the content of the judgment at hand—suggests that the mind does in fact contain computationally distinct mechanisms governing reasoning in functionally distinct domains. Results demonstrating the presence of adaptive biases where they might logically be expected in one sex but not in the other, and protective biases in response to stimuli that were ancestrally dangerous (but their conspicuous absence in response to modern threats), are key pieces of evidence in the debate about domain specificity. On the empirical side, these newer breeds of explanation cannot reasonably be dismissed as just-so stories. Although controversy about their interpretation remains, researchers from many different perspectives have tested competing predictions about classic effects and contributed their findings to the body of knowledge in psychology. The adaptive bias explanation we have featured in this chapter, error management theory, has also stimulated investigation on particular biases that were predicted a priori (e.g., women's commitment skepticism, auditory looming and navigation biases, and overestimation of physical formidability of threatening targets).
Recent investigations have also begun to document conditions that moderate certain error management biases. Further investigations of the sexual overperception bias, for example, have shown that, along with the sex of the perceiver predicting overperception (the classic EMT finding), the perceiver's levels of interest predicted even greater apparent sexual overperception (e.g., Koenig et al., 2007). Similarly, in studies of overperception and commitment skepticism in face-to-face interactions, whereas sexual overperception emerged only among men who were sexually interested in the women they were interacting with, commitment skepticism was reduced or eliminated among women who were sexually interested in the men they were interacting with (Henningsen & Henningsen, 2010). Investigations such as these provide an increasingly nuanced understanding of biases predicted by error management theory.
Many questions remain. Some scholars have noted that a cognitive bias is not actually necessary to manage error costs—a wholly accurate cognitive evaluation coupled with behavioral bias could be equally effective or superior to a cognitive bias (McKay & Dennett, 2009; McKay & Efferson, 2010). Consider sexual overperception. A man does not need to have a biased belief that a woman is sexually interested in order to approach her. He might think to himself, “My chances are low, but why not try?”
This is a plausible alternative design for managing error costs. Although EMT was originally advanced to explain cognitive biases, the core logic of the theory is neutral in predicting whether a bias must be built into belief or occur further along in the decision chain, leading more directly to biased actions. The question of whether solutions to error management problems are sometimes rooted in biased belief is an open question that can be answered only on a case-by-case basis with empirical research (Haselton & Buss, 2009). However, as sexual misperception biases, perceptual auditory looming biases, navigation biases, and many others demonstrate, there is abundant evidence that people's beliefs are indeed biased. Therefore, the argument that, in theory, error management adaptations need not involve biased beliefs does not render true cognitive biases nonexistent or impossible. The state of the evidence clearly indicates otherwise. The fascinating puzzle that remains is an explanation for why humans often seem to have biased beliefs when a behavioral bias might suffice. One possibility is that the functional thinking that has guided error management theory will need to be more fully integrated with an understanding of the proximate mechanisms that give rise to biases (Marshall et al., 2013). Such an integration could reveal that the easiest or most effective way for an evolved brain to deliver behavioral biases is via cognitive biases (Haselton & Buss, 2009).
In sum, the notion that human judgment is fundamentally flawed appears to have been flawed itself. When we observe humans in adaptively relevant environments, we can observe impressive design of human judgment that is free of irrational biases. Because of trade-offs in error costs, true biases might also prove to be more functional than one would think at first. Some genuine cognitive biases might be functional features designed by the wisdom of natural selection.