Belief bias

Belief bias is the tendency to judge the strength of arguments based on the plausibility of their conclusion rather than how strongly they support that conclusion.[1] A person is more likely to accept arguments that supports a conclusion that aligns with our values, beliefs and prior knowledge, while rejecting counter arguments to the conclusion. [2]  Belief bias is an extremely common and therefore significant form of error; we can easily be blinded by our beliefs and reach the wrong conclusion. Belief bias has been found to influence various reasoning tasks, including conditional reasoning,[3] relation reasoning[4] and transitive reasoning.[5]

Syllogisms

A syllogism is a kind of logical argument in which one proposition (the conclusion) is inferred from two or more others (the premises) of a specific form. The classical example of a valid syllogism is:

All humans are mortal. (major premise)
Socrates is human. (minor premise)
Therefore, Socrates is mortal. (conclusion)

An example of an invalid syllogism is:

All teenage girls are ambitious.
Teenage girls study hard.
Therefore, girls study hard because they are ambitious.

Typically, a majority of test subjects in studies incorrectly identify this syllogism as one in which the conclusion follows from the premises.[1] It might be true in the real world that a) girls study and b) this is because they are ambitious. However, this argument is a fallacy, because the conclusion is not supported by its premises. The validity of an argument is different from the truth of its conclusion: there are valid arguments for false conclusions and invalid arguments for true conclusions. Hence, it is an error to judge the validity of an argument from the plausibility of its conclusion. This is the reasoning error known as belief bias.[1]

When a person gives a response that is determined by the believability of the conclusion rather than logical validity, this is referred to as belief bias only when syllogism are used. This phenomenon is so closely related to syllogistic reasoning that when it does occur, in areas such as Wason's selection task or the THOG problem, it is called "memory cueing" or the "effects of content".[2]

Dual-process theory of belief bias

Many researchers in thinking and reasoning have provided evidence for a dual-process cognitive approach to reasoning, judgement and decision making. They argue that these two systems (system 1 and system 2) engage in a constant battle for control over our brain to reason and make decisions. System 1 can be described as an automated system[6] characterised by “unconscious” ,[7] "intuitive" [8] and "rapid"[6]”; whereas system 2 is said to be a controlled system,[6] characterised by “conscious”,[7] “Analytic”[8] and “Slow”;[6] some researchers (Reber 1993, Stanovich 1999) even claimed to have found a link between general intelligence and the effectiveness of decision making. It is important that note that the dual-process cognitive theory is different from the two minds hypothesis. A research done by Johnathan Evans in 2007 provided evidence for the view that System 1 which serves as a quick heuristic processor fights for control over System 2’s slower analytical approach. In the experiment, participants were asked to evaluate syllogisms that have valid arguments with unconvincing conclusions; valid arguments with convincing conclusions; invalid arguments with unconvincing conclusions; invalid arguments with convincing conclusions. The results show that when the conclusion is believable, people blindly accept invalid conclusions more so than invalid arguments are accepted.

Influencing factors of belief bias

Time

Various studies have proved that the time period of which a subject is allowed to think when evaluating arguments is related to the tendency for belief bias to take place. In a study done by Evans and Holmes in 2005,[9] they recruited two different groups of people to answer a series of reasoning questions. One group of people were given only two seconds to answer the questions; whereas the other group of people were allowed to use as much time as they would like to answer the questions. The result obtained was that a higher percentage of incorrect answers were found in the time pressured group than the other; they concluded that this was a result of a shift in logical to belief-biased way of thinking.

Nature of content

The nature of the content presented can also affect belief bias of an individual as shown by a study done by Goel & Vartanian in 2011.[10] In their experiment, 34 participants were presented with a syllogism upon each trial. Each trial were either neutral or carried some degree of negative content. Negative content involved in the experiment were politically incorrect social norm violations, such as the statement “Some wars are not unjustified, Some wars involve raping of women, therefore, Some raping of women is not unjustified”. For syllogisms where the content was neutral, the results were consistent with studies of belief bias; however, for syllogisms with negative emotional content, participants were more likely to reason logically on invalid syllogisms with believable conclusions instead of automatically judging them to be valid. In other words, the effect of belief bias is reduced when the content presented is of negative emotion. According to Goel and Vartanian, this is a because negative emotions prompts us to reason in more careful and in more detail. This argument was supported by the observation that for questions with negative emotions, the reaction time taken was significantly longer than that of questions that a neutral.

Instructions given

In an experiment done by Evans, Newstead, Allen & Pollard in 1994,[11] where subjects were given detailed instructions which lack specific reference to the notion of logical necessity when answering questions, it was shown that a larger proportion of answers actually rejected invalid arguments with convincing conclusions as opposed to when no further instructions were given when the subjects were asked to answer the questions. The results of the experiments reflects that when elaborated instructions were given to subjects to reason logically, the effects of belief bias is decreased. 

Research

In a series of experiments by Evans, Barston and Pollard (1983),[12] participants were presented with evaluation task paradigms, containing two premises and a conclusion. The participants were asked to make an evaluation of logical validity. The subjects, however, exhibited a belief bias, evidenced by their tendency to reject valid arguments with unbelievable conclusions, and endorse invalid arguments with believable conclusions. Instead of following directions and assessing logical validity, the subjects based their assessments on personal beliefs.

Consequently, these results demonstrated a greater acceptance of more believable (80%), than unbelievable (33%) conclusions. Participants also illustrated evidence of logical competences and the results determined an increase in acceptance of valid (73%) than invalid (41%). Additionally, there’s a small difference between believable and valid (89%) in comparison to unbelievable and invalid (56%) (Evans, Barston & Pollard, 1983; Morley, Evans & Handley, 2004).[12][13]

It has been argued that using more realistic content in syllogisms can facilitate more normative performance from participants. It has been suggested that the use of more abstract, artificial content will also have a biasing effect on performance. Therefore, more research is required to understand fully how and why belief bias occurs and if there are certain mechanisms that are responsible for such things. There is also evidence of clear individual differences in normative responding that are predicted by the response times of participants.[14]

A 1989 study by Markovits and Nantel gave participants four reasoning tasks. The results indicated “a significant belief-bias effect” that existed “independently of the subjects' abstract reasoning ability.”[15]

A 2010 study by Donna Torrens examined differences in belief bias among individuals. Torrens found that “the extent of an individual's belief bias effect was unrelated to a number of measures of reasoning competence” but was, instead, related to that person's ability “to generate alternative representations of premises: the more alternatives a person generated, the less likely they were to show a belief bias effect."[16]

In a 2010 study, Chad Dube and Caren M. Rotello of the University of Massachusetts and Evan Heit of the University of California, Merced, showed that “the belief bias effect is simply a response bias effect.”[17]

In a 2012 study, Adrian P. Banks of the University of Surrey explained that “belief bias is caused by the believability of a conclusion in working memory which influences its activation level, determining its likelihood of retrieval and therefore its effect on the reasoning process.”[18]

Michelle Colleen and Elizabeth Hilscher of the University of Toronto showed in 2014 that belief bias can be affected by the difficulty level and placement of the syllogism in question.[19]

See also

References

  1. 1 2 3 Robert J. Sternberg; Jacqueline P. Leighton (2004). The Nature of Reasoning. Cambridge University Press. p. 300. ISBN 978-0-521-00928-7. Retrieved 3 September 2013.
  2. 1 2 Evans, Jonathan; Newstead, Stephen; Byrne, Ruth (1993). Human Reasoning: The Psychology of Deduction. Lawrence Erlbaum Associates, Publishers. p. 243. Retrieved 26 January 2017.
  3. Evans, Jonathan St. B. T.; Handley, Simon J.; Bacon, Alison M. (2009-01-01). "Reasoning Under Time Pressure". Experimental Psychology. 56 (2): 77–83. ISSN 1618-3169. doi:10.1027/1618-3169.56.2.77.
  4. Andrews, Glenda (2010-10-01). "Belief-based and analytic processing in transitive inference depends on premise integration difficulty". Memory & Cognition. 38 (7): 928–940. ISSN 0090-502X. doi:10.3758/MC.38.7.928.
  5. Roberts, Maxwell J.; Sykes, Elizabeth D. A. (2003-01-01). "Belief bias and relational reasoning". The Quarterly Journal of Experimental Psychology Section A. 56 (1): 131–154. ISSN 0272-4987. PMID 12587899. doi:10.1080/02724980244000233.
  6. 1 2 3 4 Schneider, Walter; Shiffrin, Richard M. (1977-01-01). "Controlled and automatic human information processing: I. Detection, search, and attention.". Psychological Review. 84 (1): 1–66. ISSN 1939-1471. doi:10.1037/0033-295x.84.1.1.
  7. 1 2 Wilson, Barbara J.; Smith, Stacy L.; Potter, W. James; Kunkel, Dale; Linz, Daniel; Colvin, Carolyn M.; Donnerstein, Edward (2002-01-01). "Violence in Children's Television Programming: Assessing the Risks". Journal of Communication. 52 (1): 5–35. ISSN 1460-2466. doi:10.1111/j.1460-2466.2002.tb02531.x.
  8. 1 2 Hammond, Thomas H. (1996-04-01). "Formal Theory and the Institutions of Governance". Governance. 9 (2): 107–185. ISSN 1468-0491. doi:10.1111/j.1468-0491.1996.tb00237.x.
  9. Evans, Jonathan St B. T.; Curtis-Holmes, Jodie (2005-09-01). "Rapid responding increases belief bias: Evidence for the dual-process theory of reasoning". Thinking & Reasoning. 11 (4): 382–389. ISSN 1354-6783. doi:10.1080/13546780542000005.
  10. Goel, Vinod; Vartanian, Oshin (2011-01-01). "Negative emotions can attenuate the influence of beliefs on logical reasoning". Cognition and Emotion. 25 (1): 121–131. ISSN 0269-9931. doi:10.1080/02699931003593942.
  11. Newstead, Stephen E.; Pollard, Paul; Evans, Jonathan St. B. T.; Allen, Julie L. (1992-12-01). "The source of belief bias effects in syllogistic reasoning". Cognition. 45 (3): 257–284. doi:10.1016/0010-0277(92)90019-E.
  12. 1 2 Evans, J. St. B.T.; Barston, J.L.; Pollard, P. (1983). "On the conflict between logic and belief in syllogistic reasoning". Memory and Cognition. 11: 295–306. doi:10.3758/bf03196976.
  13. Morely, N. J.; Evans, J. St. B. T.; Handley, S. J. (2004). "Belief bias & figural bias in syllogistic reasoning". The Quartely Journal of Experimental Psychology. 57A (4): 666–692. doi:10.1080/02724980343000440.
  14. Stupple, E.J.N.; L. J. Ball; J. St. B. T. Evans; E. Kamal-Smith (2011). "When logic and belief collide: Individual differences in reasoning times support a selective processing model". Journal of Cognitive Psychology. 23 (8): 931–941. doi:10.1080/20445911.2011.589381.
  15. Markovits, H.; Nantel, G. (January 1989). "The belief-bias effect in the production and evaluation of logical conclusions.". Mem Cognit. 17: 11–7. PMID 2913452. doi:10.3758/bf03199552.
  16. Torrens, Donna (September 24, 2010). "Individual Differences and the Belief Bias Effect: Mental Models, Logical Necessity, and Abstract Reasoning". Thinking and Reasoning. 5 (1): 1–28. doi:10.1080/135467899394066. Retrieved 6 December 2016.
  17. Dube, Chad; Rotello, Caren; Heit, Evan. "The Belief Bias Effect Is Aptly Named: A Reply to Klauer and Kellen (2011)" (PDF). Psychological Review. 118 (1): 155–163. doi:10.1037/a0021774. Retrieved 6 December 2016.
  18. Banks, Adrian (September 4, 2009). "The Influence of Activation Level on Belief Bias in Relational Reasoning" (PDF). Cognitive Science. 37: 544–577. doi:10.1111/cogs.12017. Retrieved 6 December 2016.
  19. Hilscher, Michelle. "Attenuating Belief Bias Effects in Syllogistic Reasoning: The Role of Belief-Content Conflict" (PDF). University of Toronto. Retrieved 6 December 2016.

Further reading

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.