Motivated reasoning

From Wikipedia, the free encyclopedia

Motivated reasoning is the phenomenon in cognitive science and social psychology in which emotional biases lead to justifications or decisions based on their desirability rather than an accurate reflection of the evidence. It is the "tendency to find arguments in favor of conclusions we want to believe to be stronger than arguments for conclusions we do not want to believe".[1] People can therefore draw self-serving conclusions not just because they want to but because the conclusions seemed more plausible given their beliefs and expectancies.

Motivated reasoning is similar to confirmation bias, where evidence that confirms a belief (which might be a logical belief, rather than an emotional one) is either sought after more or given more credibility than evidence that disconfirms a belief. It stands in contrast to critical thinking where beliefs are approached in a skeptical and unbiased fashion.

It can lead to forming and clinging to false beliefs despite substantial evidence to the contrary. The desired outcome acts as a filter that affects evaluation of scientific evidence and of other people.[2]


Early research on how humans evaluated and integrated information supported a cognitive approach consistent with Bayesian probability, in which individuals weighted new information using rational calculations.[3] More recent theories endorse these cognitive processes as only partial explanations of motivated reasoning but have also introduced motivational[1] or affective processes.[4] These processes illuminate the mechanisms of the bias inherent in cases of motivated reasoning. To further complicate the issue, the first neuro-imaging study designed to test the neural circuitry of individuals engaged in motivated reasoning found that motivated reasoning "was not associated with neural activity in regions previously linked with cold reasoning tasks [Bayesian reasoning] and conscious (explicit) emotion regulation".[5] This section focuses on two theories that elucidate the mechanisms involved in motivated reasoning.

Ziva Kunda's theory[edit]

One review of the research by Kunda develops the following theoretical model to explain the mechanism by which motivated reasoning results in bias.[1] The model is summarized as follows:

Motivation to arrive at a desired conclusion provides a level of arousal, which acts as an initial trigger for the operation of cognitive processes. In order for someone to participate in motivated reasoning, either consciously or subconsciously, that individual first needs to be motivated. Motivation then affects reasoning by influencing the knowledge structures (beliefs, memories, information) that are accessed and the cognitive processes used.

Lodge and Taber's theory[edit]

In comparison, Milton Lodge and Charles Taber[6] introduce an empirically supported model in which affect is intricately tied to cognition, and information processing is biased toward support for positions that the individual already holds.

This model has three components:

  1. On-line processing in which when called on to make an evaluation, people instantly draw on stored information which is marked with affect;
  2. Affect is automatically activated along with the cognitive node to which it is tied;[7]
  3. A "heuristic mechanism" for evaluating new information triggers a reflection on "How do I feel?" about this topic. The result of this process results in a bias towards maintaining existing affect, even in the face of other, disconfirming information.

This theory of motivated reasoning is fully developed and tested in Lodge and Taber's The Rationalizing Voter (2013).[8] David Redlawsk (2002) found that the timing of when disconfirming information was introduced played a role in determining bias. When subjects encountered incongruity during an information search, the automatic assimilation and update process was interrupted. This results in one of two outcomes: subjects may enhance attitude strength in a desire to support existing affect (resulting in degradation in decision quality and potential bias) or, subjects may counter-argue existing beliefs in an attempt to integrate the new data.[6] This second outcome is consistent with the research on how processing occurs when one is tasked with accuracy goals.

To summarize, the two models differ in that M. Lodge and C. Taber identify a primary role for affect in guiding cognitive processes and in maintaining bias. In contrast, Kunda identifies a primary role for cognitive strategies such as memory processes, and the use of rules in determining biased information selection. At least one study in neuroscience does not support the use of cognitive processes in motivated reasoning, lending greater support to affective processing as a key mechanism in supporting bias.[citation needed]


Motivated reasoning can be classified into 2 categories: 1) those in which the motive is to arrive at an accurate conclusion, irrespective of the individual's beliefs, and 2) those in which the motive is to arrive at a particular, directional conclusion. The mechanisms in play differ based in these two scenarios.

Accuracy-oriented motivated reasoning[edit]

Several work on accuracy-driven reasoning suggest that when people are motivated to be accurate, they expend more cognitive effort, attend to relevant information more carefully, process it more deeply, often using more complex rules.

Kunda asserts that accuracy goals delay the process of coming to a premature conclusion, in that accuracy goals increase both the quantity and quality of processing—particularly in leading to more complex inferential cognitive processing procedures. When researchers manipulated test subjects’ motivation to be accurate by informing them that the target task was highly important or that they would be expected to defend their judgments, it was found that subjects utilized deeper processing and that there was less biasing of information. This was true when accuracy motives were present at the initial processing and encoding of information.[9][10] In reviewing a line of research on accuracy goals and bias, Kunda concludes, "several different kinds of biases have been shown to weaken in the presence of accuracy goals".[1] However, accuracy goals do not always eliminate biases and improve reasoning. Some biases(biases resulting from using availability heuristic) might be resistant to accuracy manipulations. For accuracy to reduce bias, the following conditions must be present:

  1. Subjects must possess appropriate reasoning strategies.
  2. They must view these as superior to other strategies.
  3. They must be capable of using these strategies at will.

These last two conditions introduce the construct that accuracy goals include a conscious process of utilizing cognitive strategies in motivated reasoning. This construct is called into question by later neuroscience research that concludes that motivated reasoning is qualitatively distinct from reasoning (in instances when there is no strong emotional stake in the outcomes).[5]

Goal-oriented motivated reasoning[edit]

Directional goals enhance the accessibility of knowledge structures (memories, beliefs, information) that are consistent with desired conclusions. According to Kunda, such goals can lead to biased memory search and belief construction mechanism.[1] Several studies serve as evidence for the effect of directional goals in selection and construction of beliefs about oneself, other people and the world. Cognitive dissonance research provides extensive evidence that people may bias their self-characterizations when motivated to do so. For example, in one study, subjects altered their self-view by viewing themselves as more extroverted when induced to believe that extroversion was beneficial.

The motivation to achieve directional goals could also influence which rules (procedural structures, such as inferential rules) are accessed to guide the search for information. Studies also suggest that evaluation of scientific evidence may be biased by whether the conclusions are in-line with the reader's beliefs.

In spite of goal-oriented motivated reasoning, people are not at the liberty to conclude whatever they want to conclude merely because they want to.[1] People tend to draw conclusions only if they can muster up the evidence necessary to support it. They search memory for those beliefs and rules that could support their desired conclusion or they could create new beliefs to logically support their desired goals.


Climate change[edit]

Coal-fired power station Neurath in Grevenbroich, North Rhine-Westphalia, Germany.

The topic of climate change is a prime example where motivated reasoning to believe in climate change is shown. Climate change is becoming an increasingly obvious issue in the US specifically[citation needed]. Though there are many facts and evidence showing it[citation needed], many still like to debate if the issue really is what it seems to be. Many deny climate change, say it is a hoax, blame the government, mind control, conspiracy theories, etc. “A significant segment of the American public has fixed beliefs, either because they are not politically engaged, or because they hold strong beliefs that are unlikely to change”.[11] With the hundreds of facts given to people about climate change, the thousands of power plants and melting ice caps, the phenomenon of motivated reasoning keeps people stubborn in their belief that climate change is not real.

Social media[edit]

Social media is used for many different purposes and ways of spreading opinions. It is the number one place people go to get information and most of that information is complete opinion and bias. The way this applies to motivated reasoning is the way it spreads, “However, motivated reasoning suggests that informational uses of social media are conditioned by various social and cultural ways of thinking”.[12] All ideas and opinions are shared and makes it very easy for motivated reasoning and biases to come through when searching for an answer or just facts on the internet or any news source.


As stated above, neuroscience research suggests that "motivated reasoning is qualitatively distinct from reasoning when people do not have a strong emotional stake in the conclusions reached."[5] However, if there is a strong emotion attached during their previous round of motivated reasoning and that emotion is again present when the individual's conclusion is reached, a strong emotional stake is then attached to the conclusion. Any new information in regards to that conclusion will cause motivated reasoning to reoccur. This can create pathways within the neural network that further ingrains the reasoned beliefs of that individual along similar neural networks where logical reasoning occurs. This causes the strong emotion to reoccur when confronted with contradictory information, time and time again. This is what is referred to by Lodge and Taber as the affective contagion.[8] But instead of "infecting" other individuals, the emotion "infects" the individuals reasoning pathways and conclusions.

Social science research suggests that reasoning away contradictions is psychologically easier than revising feelings. As previously discussed, emotions are shown to color how "facts" are perceived. Feelings come first, and evidence is used in service of those feelings. Evidence that supports what is already believed is accepted. Evidence which contradicts those beliefs is not.[13] An example of motivated reasoning in the public sphere is the fact that many people continued to believe that Barack Obama was not born in the United States in the face of ample evidence that he was.[14]


The outcomes of motivated reasoning derive from "a biased set of cognitive processes—that is, strategies for accessing, constructing, and evaluating beliefs. The motivation to be accurate enhances use of those beliefs and strategies that are considered most appropriate, whereas the motivation to arrive at particular conclusions enhances use of those that are considered most likely to yield the desired conclusion."[1] Careful or "reflective" reasoning has been linked to both overcoming and reinforcing motivated reasoning, suggesting that reflection is not a panacea, but a tool that can be used for rational or arational purposes depending on other factors.[15] For example, when people are presented with and forced to think analytically about something complex that they lack adequate knowledge of (i.e. being presented with a new study on meteorology whilst having no degree in the subject), there is no directional shift in thinking, and their extant conclusions are more likely to be supported with motivated reasoning. Conversely, if they are presented with a more simplistic test of analytical thinking that confronts their beliefs (i.e. seeing implausible headlines as false), motivated reasoning is less likely to occur and a directional shift in thinking may result.[16]

Research on motivated reasoning tested accuracy goals (i.e., reaching correct conclusions) and directional goals (i.e., reaching preferred conclusions). Factors such as these affect perceptions; and results confirm that motivated reasoning affects decision-making and estimates.[17] These results have far reaching consequences because, when confronted with a small amount of information contrary to an established belief, an individual is motivated to reason away the new information, contributing to the hostile media effect.[18] If this pattern continues over an extended period of time, the individual becomes more entrenched in their beliefs. However, recent studies have shown that motivated reasoning can be overcome. "When the amount of incongruency is relatively small, the heightened negative affect does not necessarily override the motivation to maintain [belief]."[This quote needs a citation] However, there is evidence of a theoretical "tipping point" where the amount of incongruent information that is received by the motivated reasoner can turn certainty into anxiety. This anxiety of being incorrect may lead to a change of opinion.[13]

See also[edit]


  1. ^ a b c d e f g Kunda, Z. (1990). "The case for motivated reasoning". Psychological Bulletin. 108 (3): 480–498. doi:10.1037/0033-2909.108.3.480. PMID 2270237.
  2. ^ O'Leary, Ann. "Teaching Tip Sheet: Motivated Reasoning," Archived 2017-12-01 at the Wayback Machine American Psychological Association (APA); citing Sternlicht, H. C. (1977). "The range of periodontal therapy--past and present". Texas Dental Journal. 95 (10): 6–13. PMID 270237.
  3. ^ Gerber, Alan; Green, Donald (1999). "Misperceptions about Perceptual Bias". Annual Review of Political Science. 18 (11): 189–2100. doi:10.1146/annurev.polisci.2.1.189.
  4. ^ Redlawsk, D (2002). "Hot Cognition or Cool Consideration". The Journal of Politics. 64 (4): 1021–1044. doi:10.1111/1468-2508.00161. S2CID 15895299.
  5. ^ a b c Westen, D.; Blagov, P. S.; Harenski, K.; Kilts, C.; Hamann, S. (2006). "Neural Bases of Motivated Reasoning: An fMRI Study of Emotional Constraints on Partisan Political Judgment in the 2004 U.S. Presidential Election". Journal of Cognitive Neuroscience. 18 (11): 1947–1958. CiteSeerX doi:10.1162/jocn.2006.18.11.1947. PMID 17069484. S2CID 8625992.
  6. ^ a b Lodge and Taber (2000). "Three steps toward a theory of motivated political reasoning". Elements of Reason: Cognition, Choice and Bounds of Rationality. Cambridge University Press. ISBN 978-0-521-65329-9.[page needed]
  7. ^ Fazio, Russell (1995). Richard E. Petty; Jon A. Krosnick (eds.). Attitude Strength: Antecedents and Consequences. Mahwah, NJ. ISBN 978-0-8058-1086-8.[page needed]
  8. ^ a b Milton Lodge; Charles Taber (2013). The Rationalizing Voter. New York: Cambridge University Press.
  9. ^ Tetlock, P (1983). "Accountability and the perseverance of first impressions". Social Psychology Quarterly. 46 (4): 285–292. doi:10.2307/3033716. JSTOR 3033716.
  10. ^ Tetlock, P (1985). "Accountability: A social check on the fundamental attribution error". Social Psychology Quarterly. 48 (3): 227–236. doi:10.2307/3033683. JSTOR 3033683.
  11. ^ Brulle, Robert J.; Carmichael, Jason; Jenkins, J. Craig (September 2012). "Shifting public opinion on climate change: an empirical assessment of factors influencing concern over climate change in the U.S., 2002–2010". Climatic Change. 114 (2): 169–188. Bibcode:2012ClCh..114..169B. doi:10.1007/s10584-012-0403-y. hdl:10.1007/s10584-012-0403-y. S2CID 8220644.
  12. ^ Diehl, Trevor; Huber, Brigitte; Gil de Zúñiga, Homero; Liu, James (17 August 2021). "Social Media and Beliefs about Climate Change: A Cross-National Analysis of News Use, Political Ideology, and Trust in Science". International Journal of Public Opinion Research. 33 (2): 197–213. doi:10.1093/ijpor/edz040.
  13. ^ a b Redlawsk, D. P.; Civettini, A. J. W.; Emmerson, K. M. (2010). "The Affective Tipping Point: Do Motivated Reasoners Ever "Get It"?". Political Psychology. 31 (4): 563. doi:10.1111/j.1467-9221.2010.00772.x.
  14. ^ Redlawsk, David P. (April 22, 2011). "A Matter of Motivated 'Reasoning'". New York Times.
  15. ^ Byrd, Nick (2022). "Bounded Reflectivism and Epistemic Identity". Metaphilosophy. doi:10.1111/meta.12534.
  16. ^ Pennycook, Gordon; Rand, David G. (July 2019). "Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning". Cognition. 188: 39–50. doi:10.1016/j.cognition.2018.06.011. PMID 29935897. S2CID 49418191.
  17. ^ Nir, Lilach (2011). "Motivated Reasoning and Public Opinion Perception". Public Opinion Quarterly. 75 (3): 504–532. doi:10.1093/poq/nfq076.
  18. ^ Tsang, Stephanie Jean (1 August 2018). "Empathy and the Hostile Media Phenomenon". Journal of Communication. 68 (4): 809–829. doi:10.1093/joc/jqy031.

Further reading[edit]