Cognitive bias

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Jojalozzo (talk | contribs) at 00:11, 18 May 2014 (→‎Practical significance: refspam). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

A cognitive bias is a pattern of deviation in judgment, whereby inferences about other people and situations may be drawn in an illogical fashion.[1] Individuals create their own “subjective social reality” from their perception of the input.[2] An individual’s construction of social reality, not the objective input, may dictate their behaviour in the social world.[3] Thus, cognitive biases may sometimes lead to perceptual distortion, inaccurate judgment, illogical interpretation, or what is broadly called irrationality.[4][5][6]

Some cognitive biases are presumably adaptive. Cognitive biases may lead to more effective actions in a given context.[7] Furthermore, cognitive biases enable faster decisions when timeliness is more valuable than accuracy, as illustrated in heuristics.[8] Other cognitive biases are a “by-product” of human processing limitations,[9] resulting from a lack of appropriate mental mechanisms (bounded rationality), or simply from a limited capacity for information processing.[10]

A continually evolving list of cognitive biases has been identified over the last six decades of research on human judgment and decision-making in cognitive science, social psychology, and behavioral economics. Cognitive biases are important to study because “systematic errors” highlight the “psychological processes that underlie perception and judgement” (Tversky & Kahneman,1999, p. 582). Moreover, Kahneman and Tversky (1996) argue cognitive biases have efficient practical implications for areas including clinical judgment.[11]

Overview

Daniel Kahneman

Bias arises from various processes that are sometimes difficult to distinguish. These include

  • mental noise
  • the mind's limited information processing capacity[13]
  • emotional and moral motivations[14]
  • social influence[15]

The notion of cognitive biases was introduced by Amos Tversky and Daniel Kahneman in 1972[16] and grew out of their experience of people's innumeracy, or inability to reason intuitively with the greater orders of magnitude. Tversky, Kahneman and colleagues demonstrated several replicable ways in which human judgments and decisions differ from rational choice theory. Tversky and Kahneman explained human differences in judgement and decision making in terms of heuristics. Heuristics involve mental shortcuts which provide swift estimates about the possibility of uncertain occurrences (Baumeister & Bushman, 2010, p. 141). Heuristics are simple for the brain to compute but sometimes introduce “severe and systematic errors” (Tversky & Kahneman, 1974, p. 1125).[17]

For example, the representativeness heuristic is defined as the tendency to “judge the frequency or likelihood” of an occurrence by the extent of which the event “resembles the typical case” (Baumeister & Bushman, 2010, p. 141). The “Linda Problem” illustrates the representativeness heuristic (Tversky & Kahneman, 1983[18] ). Participants were given a description of "Linda" that suggests Linda might well be a feminist (e.g., she is said to be concerned about discrimination and social justice issues). They were then asked whether they thought Linda was more likely to be a “(a) bank teller” or a “(b) bank teller and active in the feminist movement”. A majority chose answer “(b)”. This error (mathematically, answer (b) cannot be more likely than answer (a)) is an example of the “conjunction fallacy”; Tversky and Kahneman argued that respondents chose (b) because it seemed more "representative" or typical of persons who might fit the description of Linda. The representativeness heuristic may lead to errors such as activating stereotypes and inaccurate judgements of others (Haselton et al., 2005, p. 726).

Alternatively, critics of Kahneman and Tversky such as Gerd Gigerenzer argue that heuristics should not lead us to conceive of human thinking as riddled with irrational cognitive biases, but rather to conceive rationality as an adaptive tool that is not identical to the rules of formal logic or the probability calculus.[19] Nevertheless, experiments such as the “Linda problem” grew into the heuristics and biases research program which spread beyond academic psychology into other disciplines including medicine and political science.

Types

Biases can be distinguished on a number of dimensions. For example, there are biases specific to groups (such as the risky shift) as well as biases at the individual level.

Some biases affect decision-making, where the desirability of options has to be considered (e.g., sunk costs fallacy). Others such as illusory correlation affect judgment of how likely something is, or of whether one thing is the cause of another. A distinctive class of biases affect memory,[20] such as consistency bias (remembering one's past attitudes and behavior as more similar to one's present attitudes).

Some biases reflect a subject's motivation,[21] for example, the desire for a positive self-image leading to Egocentric bias[22] and the avoidance of unpleasant cognitive dissonance. Other biases are due to the particular way the brain perceives, forms memories and makes judgments. This distinction is sometimes described as "Hot cognition" versus "Cold Cognition", as motivated reasoning can involve a state of arousal.

Among the "cold" biases,

  • some involve a decision or judgement being affected by irrelevant information (for example the framing effect where the same problem receives different responses depending on how it is described; or the distinction bias where choices presented together have different outcomes than those presented separately)
  • others give excessive weight to an unimportant but salient feature of the problem (e.g., anchoring)

The fact that some biases reflect motivation, and in particular the motivation to have positive attitudes to oneself[22] accounts for the fact that many biases are self-serving or self-directed (e.g. illusion of asymmetric insight, self-serving bias, projection bias). There are also biases in how subjects evaluate in-groups or out-groups; evaluating in-groups as more diverse and "better" in many respects, even when those groups are arbitrarily-defined (ingroup bias, outgroup homogeneity bias).

Some cognitive biases belong to the subgroup of attentional biases which refer to the paying of increased attention to certain stimuli. It has been shown, for example, that people addicted to alcohol and other drugs pay more attention to drug-related stimuli. Common psychological tests to measure those biases are the Stroop Task[23][24] and the Dot Probe Task.

The following is a list of the more commonly studied cognitive biases:

Name Description
Fundamental Attribution Error (FAE) Also known as the correspondence bias (Baumeister & Bushman, 2010) is the tendency for people to over-emphasize personality-based explanations for behaviours observed in others. At the same time, individuals under-emphasize the role and power of situational influences on the same behaviour. Jones and Harris’ (1967)[25] classic study illustrates the FAE. Despite being made aware that the target’s speech direction (pro-Castro/anti-Castro) was assigned to the writer, participants ignored the situational pressures and attributed pro-Castro attitudes to the writer when the speech represented such attitudes.
Confirmation bias The tendency to search for or interpret information in a way that confirms one's preconceptions. In addition, individuals may discredit information that does not support their views.[26] The confirmation bias is related to the concept of cognitive dissonance. Whereby, individuals may reduce inconsistency by searching for information which re-confirms their views (Jermias, 2001, p. 146).[27]
Self-serving bias The tendency to claim more responsibility for successes than failures. It may also manifest itself as a tendency for people to evaluate ambiguous information in a way beneficial to their interests.
Belief bias When one's evaluation of the logical strength of an argument is biased by their belief in the truth or falsity of the conclusion.
Framing Using a too-narrow approach and description of the situation or issue.
Hindsight bias Sometimes called the "I-knew-it-all-along" effect, is the inclination to see past events as being predictable.

A 2012 Psychological Bulletin article suggests that at least 8 seemingly unrelated biases can be produced by the same information-theoretic generative mechanism.[28] It is shown that noisy deviations in the memory-based information processes that convert objective evidence (observations) into subjective estimates (decisions) can produce regressive conservatism, the belief revision (Bayesian conservatism), illusory correlations, illusory superiority (better-than-average effect) and worse-than-average effect, subadditivity effect, exaggerated expectation, overconfidence, and the hard–easy effect.

Practical significance

Many social institutions rely on individuals to make rational judgments.

The present securities regulation regime largely assumes that all investors act as perfectly, rational persons. In truth, actual investors face cognitive limitations from biases, heuristics, and framing effects.

A fair jury trial, for example, requires that the jury ignore irrelevant features of the case, weigh the relevant features appropriately, consider different possibilities open-mindedly and resist fallacies such as appeal to emotion. The various biases demonstrated in these psychological experiments suggest that people will frequently fail to do all these things.[29] However, they fail to do so in systematic, directional ways that are predictable.[30]

Cognitive biases are also related to the persistence of superstition, to large social issues such as prejudice, and they also work as a hindrance in the acceptance of scientific non-intuitive knowledge by the public.[31]

Reducing cognitive bias

Similar to Gigerenzer (1996),[32] Haselton et al. (2005) state the content and direction of cognitive biases are not "arbitrary" (p. 730).[9] Moreover, cognitive biases can be controlled. Debiasing is a technique which aims to decrease biases by encouraging individuals to use controlled processing compared to automatic processing (Baumeister & Bushman, 2010, p. 155).[33] In relation to reducing the FAE, monetary incentives[34] and informing participants they will be held accountable for their attributions [35] have been linked to the increase of accurate attributions.

Cognitive bias modification refers to the process of modifying cognitive biases in healthy people and also refers to a growing area of psychological (non-pharmaceutical) therapies for anxiety, depression and addiction called CBMT. Cognitive Bias Modification Therapy (CBMT) is sub-group of therapies within a growing area of psychological therapies based on modifying cognitive processes with or without accompanying medication and talk therapy, sometimes referred to as Applied Cognitive Processing Therapies (ACPT). Although Cognitive Bias Modification can refer to modifying cognitive processes in healthy individuals, CBMT is a growing area of evidence-based psychological therapy, in which cognitive processes are modified to relieve suffering[36][37] from serious Depression,[38] Anxiety,[39] and Addiction.[40] CBMT techniques are technology assisted therapies that are delivered via a computer with or without clinician support. CBM combines evidence and theory from the cognitive model of anxiety,[41] cognitive neuroscience[42] and attentional models.[43]

See also

References

  1. ^ Haselton, M. G., Nettle, D., & Andrews, P. W. (2005). The evolution of cognitive bias. In D. M. Buss (Ed.), The Handbook of Evolutionary Psychology: Hoboken, NJ, US: John Wiley & Sons Inc. pp. 724–746.{{cite book}}: CS1 maint: multiple names: authors list (link)
  2. ^ Bless, H., Fiedler, K., & Strack, F. (2004). Social cognition: How individuals construct social reality. Hove and New York: Psychology Press. p. 2.{{cite book}}: CS1 maint: multiple names: authors list (link)
  3. ^ Bless, H., Fiedler, K., & Strack, F. (2004). Social cognition: How individuals construct social reality. Hove and New York: Psychology Press.{{cite book}}: CS1 maint: multiple names: authors list (link)
  4. ^ Kahneman, D.; Tversky, A. (1972). "Subjective probability: A judgment of representativeness". Cognitive Psychology. 3 (3): 430–454. doi:10.1016/0010-0285(72)90016-3.
  5. ^ Baron, J. (2007). Thinking and deciding (4th ed.). New York, NY: Cambridge University Press.
  6. ^ Ariely, D. (2008). Predictably irrational: The hidden forces that shape our decisions. New York, NY: HarperCollins.
  7. ^ For instance: Gigerenzer, G. & Goldstein, D. G. (1996). "Reasoning the fast and frugal way: Models of bounded rationality". Psychological Review. 103: 650–669. doi:10.1037/0033-295X.103.4.650.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  8. ^ Tversky, A., & Kahneman, D. (1974). "Judgement under uncertainty: Heuristics and biases". Sciences. 185: 1124–1131. doi:10.1126/science.185.4157.1124.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  9. ^ a b Haselton, M. G., Nettle, D., & Andrews, P. W. (2005). The evolution of cognitive bias. In D. M. Buss (Ed.), The Handbook of Evolutionary Psychology: Hoboken, NJ, US: John Wiley & Sons Inc. pp. 724–746.{{cite book}}: CS1 maint: multiple names: authors list (link)
  10. ^ Bless, H., Fiedler, K., & Strack, F. (2004). Social cognition: How individuals construct social reality. Hove and New York: Psychology Press.{{cite book}}: CS1 maint: multiple names: authors list (link)
  11. ^ Kahneman, D., & Tversky, A. (1996). "On the reality of cognitive illusions". Psychological Review. 103 (3): 582–591. doi:10.1037/0033-295X.103.3.582.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  12. ^ Kahneman, D., Slovic, P., & Tversky, A. (1982). Judgment under uncertainty: Heuristics and biases (1st ed.). Cambridge University Press.
  13. ^ Simon, H. A. (1955). A behavioral model of rational choice. The Quarterly Journal of Economics, 69(1), 99 -118. doi:10.2307/1884852
  14. ^ Pfister, H.-R., & Böhm, G. (2008). The multiplicity of emotions: A framework of emotional functions in decision making" Judgment and Decision Making 3, 5–17.
  15. ^ Wang, X. T., Simons, F., & Brédart, S. (2001). Social cues and verbal framing in risky choice. Journal of Behavioral Decision Making, 14(1), 1–15. doi:10.1002/1099-0771(200101)14:1<1::AID-BDM361>3.0.CO;2-N
  16. ^ Kahneman, Daniel; Shane Frederick (2002). "Representativeness Revisited: Attribute Substitution in Intuitive Judgment". In Thomas Gilovich, Dale Griffin, Daniel Kahneman (ed.). Heuristics and Biases: The Psychology of Intuitive Judgment. Cambridge: Cambridge University Press. pp. 51–52. ISBN 978-0-521-79679-8.{{cite book}}: CS1 maint: multiple names: editors list (link)
  17. ^ Tversky, A., & Kahneman, D. (1974). "Judgement under uncertainty: Heuristics and biases". Sciences. 185: 1124–1131. doi:10.1126/science.185.4157.1124.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  18. ^ Tversky, A., & Kahneman, D. (1983). "Extensional versus intuitive reasoning: The conjunction fallacy in probability judgement". Psychological Review. 90: 293–315. doi:10.1037/0033-295X.90.4.293.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  19. ^ Gigerenzer, G. (2006). "Bounded and Rational". In Stainton, R. J. (ed.). Contemporary Debates in Cognitive Science. Blackwell. p. 129. ISBN 1-4051-1304-9.
  20. ^ Schacter, D.L. (1999). "The Seven Sins of Memory: Insights From Psychology and Cognitive Neuroscience". American Psychologist. 54 (3): 182–203. doi:10.1037/0003-066X.54.3.182. PMID 10199218.
  21. ^ Kunda, Z. (1990). "The Case for Motivated Reasoning". Psychological Bulletin. 108 (3): 480–498. doi:10.1037/0033-2909.108.3.480. PMID 2270237.
  22. ^ a b Hoorens, V. (1993). "Self-enhancement and Superiority Biases in Social Comparison". In Stroebe, W. and Hewstone, Miles (ed.). European Review of Social Psychology 4. Wiley.{{cite book}}: CS1 maint: multiple names: editors list (link)
  23. ^ Jensen AR, Rohwer WD (1966). "The Stroop color-word test: a review". Acta psychologica. 25 (1): 36–93. doi:10.1016/0001-6918(66)90004-7. PMID 5328883.
  24. ^ MacLeod CM (March 1991). "Half a century of research on the Stroop effect: an integrative review". Psychological Bulletin. 109 (2): 163–203. doi:10.1037/0033-2909.109.2.163. PMID 2034749.
  25. ^ Jones, E. E., & Harris, V. A (1967). "The attribution of attitudes". Journal of Experimental Social Psychology. 3: 1–24. doi:10.1016/0022-1031(67)90034-0.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  26. ^ Mahoney, M. J. (1977). "Publication prejudices: An experimental study of confirmatory bias in the peer review system". Cognitive Therapy and Research,. 1 (2): 161–175. doi:10.1007/bf01173636.{{cite journal}}: CS1 maint: extra punctuation (link)
  27. ^ Jermias, J. (2001). "Cognitive dissonance and resistance to change: The influence of commitment confirmation and feedback on judgement usefulness of accounting systems". Accounting, Organizations and Society. 26: 141–160. doi:10.1016/s0361-3682(00)00008-8.
  28. ^ Martin Hilbert (2012) "Toward a synthesis of cognitive biases: How noisy information processing can bias human decision making"" Psychological Bulletin 138(2), 211–237; free access to the study here: martinhilbert.net/HilbertPsychBull.pdf
  29. ^ Sutherland, Stuart (2007) Irrationality: The Enemy Within Second Edition (First Edition 1994) Pinter & Martin. ISBN 978-1-905177-07-3
  30. ^ Ariely, Dan (2008). Predictably Irrational: The Hidden Forces That Shape Our Decisions. HarperCollins. p. 304. ISBN 978-0-06-135323-9. {{cite book}}: Invalid |ref=harv (help)
  31. ^ Günter Radden, H. Cuyckens (2003). Motivation in language: studies in honor of Günter Radden. John Benjamins. p. 275. ISBN 978-1-58811-426-6.
  32. ^ Gigerenzer, G. (1996). "On narrow norms and vague heuristics: A reply to Kahneman and Tversky (1996)". Psychological Review. 103 (3): 592–596. doi:10.1037/0033-295x.103.3.592.
  33. ^ Baumeister, R. F. & Bushman, B. J. (2010). Social psychology and human nature: International Edition. Belmont, USA: Wadsworth.{{cite book}}: CS1 maint: multiple names: authors list (link)
  34. ^ Vonk, R. (1999). "Effects of outcome dependency on correspondence bias". Personality and Social Psychology Bulletin. 25: 382–389. doi:10.1177/0146167299025003009.
  35. ^ Tetlock, P. E. (1985). "Accountability: A social check on the fundamental attribution error". Social Psychology Quarterly. 48: 227–236. doi:10.2307/3033683.
  36. ^ MacLeod, C., Mathews, A., & Tata, P. (1986). Attentional Bias in Emotional Disorders" Journal of Abnormal Psychology 95(1), 15–20.
  37. ^ Bar-Haim, Y., Lamy, D., Pergamin, L., Bakermans-Kranenburg, M. J., & van, I. M. H. (2007). Threat-related attentional bias in anxious and nonanxious individuals: a meta-analytic study. Psychol Bull, 133(1), 1–24. doi:10.1037/0033-2909.133.1.1
  38. ^ Holmes, E. A., Lang, T. J., & Shah, D. M. (2009). Developing interpretation bias modification as a "cognitive vaccine" for depressed mood: imagining positive events makes you feel better than thinking about them verbally. J Abnorm Psychol, 118(1), 76–88. doi:10.1037/a0012590
  39. ^ Hakamata, Y., Lissek, S., Bar-Haim, Y., Britton, J. C., Fox, N. A., Leibenluft, E., ... Pine, D. S. (2010). Attention bias modification treatment: a meta-analysis toward the establishment of novel treatment for anxiety" Biol Psychiatry 68(11), 982–990. doi:10.1016/j.biopsych.2010.07.021
  40. ^ Eberl, C., Wiers, R. W., Pawelczack, S., Rinck, M., Becker, E. S., & Lindenmeyer, J. (2013). Approach bias modification in alcohol dependence: Do clinical effects replicate and for whom does it work best?" Developmental Cognitive Neuroscience 4(0), 38–51. doi:10.1016/j.dcn.2012.11.002
  41. ^ Clark, D. A., & Beck, A. T. (2009). Cognitive Therapy of Anxiety Disorders: Science and Practice. London: Guildford.
  42. ^ Browning, M., Holmes, E. A., Murphy, S. E., Goodwin, G. M., & Harmer, C. J. (2010). Lateral prefrontal cortex mediates the cognitive modification of attentional bias" Biol Psychiatry 67(10), 919–925. doi:10.1016/j.biopsych.2009.10.031
  43. ^ Eysenck, M. W., Derakshan, N., Santos, R., & Calvo, M. G. (2007). Anxiety and cognitive performance: Attentional control theory. Emotion, 7(2), 336–353. doi:10.1037/1528-3542.7.2.336

Further reading

  • Eiser, J.R. and Joop van der Pligt (1988) Attitudes and Decisions London: Routledge. ISBN 978-0-415-01112-9
  • Fine, Cordelia (2006) A Mind of its Own: How your brain distorts and deceives Cambridge, UK: Icon Books. ISBN 1-84046-678-2
  • Gilovich, Thomas (1993). How We Know What Isn't So: The Fallibility of Human Reason in Everyday Life. New York: The Free Press. ISBN 0-02-911706-2
  • Haselton, M.G., Nettle, D. & Andrews, P.W. (2005). The evolution of cognitive bias. In D.M. Buss (Ed.), Handbook of Evolutionary Psychology, (pp. 724–746). Hoboken: Wiley. Full text
  • Heuer, Richards J. Jr. (1999) Psychology of Intelligence Analysis. Central Intelligence Agency. http://www.au.af.mil/au/awc/awcgate/psych-intel/art5.html
  • Kahneman D., Slovic P., and Tversky, A. (Eds.) (1982) Judgment Under Uncertainty: Heuristics and Biases. New York: Cambridge University Press ISBN 978-0-521-28414-1
  • Kahneman, Daniel (2011) Thinking, Fast and Slow. New York: Farrar, Straus and Giroux ISBN 978-0-374-27563-1
  • Kida, Thomas (2006) Don't Believe Everything You Think: The 6 Basic Mistakes We Make in Thinking New York: Prometheus. ISBN 978-1-59102-408-8
  • Nisbett, R., and Ross, L. (1980) Human Inference: Strategies and shortcomings of human judgement. Englewood Cliffs, NJ: Prentice-Hall ISBN 978-0-13-445130-5
  • Piatelli-Palmarini, Massimo (1994) Inevitable Illusions: How Mistakes of Reason Rule Our Minds New York: John Wiley & Sons. ISBN 0-471-15962-X
  • Stanovich, Keith (2009). What Intelligence Tests Miss: The Psychology of Rational Thought. New Haven (CT): Yale University Press. ISBN 978-0-300-12385-2. {{cite book}}: Unknown parameter |laydate= ignored (help); Unknown parameter |laysummary= ignored (help)
  • Sutherland, Stuart (2007) Irrationality: The Enemy Within Second Edition (First Edition 1994) Pinter & Martin. ISBN 978-1-905177-07-3
  • Tavris, Carol and Elliot Aronson (2007) Mistakes Were Made (But Not by Me): Why We Justify Foolish Beliefs, Bad Decisions and Hurtful Acts Orlando, Florida: Harcourt Books. ISBN 978-0-15-101098-1
  • Funder, David C.; Joachim I. Krueger (June 2004). "Towards a balanced social psychology: Causes, consequences, and cures for the problem-seeking approach to social behavior and cognition" (PDF). Behavioral and Brain Sciences. 27 (3): 313–376. PMID 15736870. Retrieved 3 May 2011.

External links