Controversy

Controversy is a state of prolonged public dispute or debate, usually concerning a matter of opinion. The word was coined from the Latin controversia, as a composite of controversus – "turned in an opposite direction," from contra – "against" – and vertere – to turn, or versus (see verse), hence, "to turn against."

Perennial areas of controversy include history, religion, philosophy and politics. Other minor areas of controversy may include economics, science, finances, organisation, age, gender, and race. Controversy in matters of theology has traditionally been particularly heated, giving rise to the phrase odium theologicum. Controversial issues are held as potentially divisive in a given society, because they can lead to tension and ill will, as a result they are often taboo to be discussed in the light of company in many cultures.

Contents

Legal controversy

In the theory of law, a controversy differs from a legal case; while legal cases include all suits, criminal as well as civil, a controversy is a purely civil proceeding.

For example, the Case or Controversy Clause of Article Three of the United States Constitution (Section 2, Clause 1) states that "the judicial Power shall extend ... to Controversies to which the United States shall be a Party". This clause has been deemed to impose a requirement that United States federal courts are not permitted to hear cases that do not pose an actual controversy—that is, an actual dispute between adverse parties which is capable of being resolved by the court. In addition to setting out the scope of the jurisdiction of the federal judiciary, it also prohibits courts from issuing advisory opinions, or from hearing cases that are either unripe, meaning that the controversy has not arisen yet, or moot, meaning that the controversy has already been resolved.

Benford's law of controversy

Benford's law of controversy, as expressed by science-fiction author Gregory Benford in 1980, states: Passion is inversely proportional to the amount of real (true) information available.[1][2] In other words, the fewer facts are known to and agreed on by the participants, the more controversy there is, and the more is known the less controversy there is. Thus, for example, controversies in physics are limited to subject-areas where experiments cannot be carried out yet, whereas Benford's Law implies that controversy is inherent to politics, where communities must frequently decide on courses of action based on insufficient information.

Psychological bases of controversy

Controversies are frequently thought to be a result of a lack of information on the part of the disputants - as in Benford's Law of Controversy. For example, in the political controversy over anthropogenic climate change that is prevalent in the United States - it has been thought that those who are opposed to the scientific consensus did so because of a lack of evidence.[3][4] A study of 1540 US adults[5] found instead that levels of scientific literacy were correlated with the strength of opinion on climate change, but not on which side of the debate that they stood.

The puzzling phenomenon of two individuals being exposed to the same evidence and being able to reach different conclusions, has been frequently explained (particularly by Daniel Kahneman) by reference to a 'bounded rationality' - that is most judgments are made by fast acting heuristics (system 1)[6][7] that work well in every day situations, but are not amenable to decision making about complex subjects such as climate change. Anchoring has been particularly identified as relevant in climate change controversies [8] as individuals are found to be more positively inclined to believe in climate change if the outside temperature is higher, if they have been primed to think about heat, and if they are primed with higher temperatures when thinking about the future temperature increases from climate change.

In other controversies - such as that around the HPV vaccine, the same evidence seemed to license inference to radically different conclusions.[9] Kahan et al.[10] explained this by the cognitive biases of Biased Assimilation[11] and a Credibility Heuristic.[12]

Similar effects on reasoning are also seen in non-scientific controversies, for example in the gun control debate in the United States.[13] As with other controversies, it has been suggested that exposure to empirical facts would be sufficient to resolve the gun control debate once and for all.[14][15] In computer simulations of cultural communities, beliefs were found to polarize within isolated sub-groups, based on the mistaken belief of the community's unhindered access to ground truth.[13] Such confidence in the group to find the ground truth is explicable through the success of wisdom of the crowd based inferences,[16] however, if there is no access to the ground truth, as there was not in this model, the method will fail.

Bayesian decision theory allows these failures of rationality to be described as part of a statistically optimized system for decision making. Experiments and computational models in Multimodal integration have shown that sensory input from different senses is integrated in a statistically optimal way,[17] in addition, it appears that the kind of inferences used to infer single sources for multiple sensory inputs uses a Bayesian inference about the causal origin of the sensory stimuli.[18] As such, it appears neurobiologically plausible that the brain implements decision-making procedures that are close to optimal for Bayesian inference.

Brocas and Carrillo propose a model to make decisions based on noisy sensory inputs,[19] beliefs about the state of the world are modified by Bayesian updating, and then decisions are made based on beliefs passing a threshold. They show that this model, when optimized for single-step decision making, produces Belief Anchoring and Polarization of opinions - exactly as described in the global warming controversy context - in spite of identical evidence presented, the pre-existing beliefs (or evidence presented first) has an overwhelming effect on the beliefs formed. In addition, the preferences of the agent (the particular rewards that they value) also cause the beliefs formed to change - this explains the Biased Assimilation (also known as Confirmation Bias) shown above. This model allows the production of controversy to be seen as a consequence of a decision maker optimized for single-step decision making, rather than as a result of limited reasoning in the Bounded rationality of Kahneman.

See also

References

  1. ^ "EFF Quotes Collection 19.6". Electronic Frontier Foundation. 2001-04-09. http://www.eff.org/Misc/EFF/?f=quotes.eff.txt. 
  2. ^ "Quotations: Computer Laws". SysProg. Archived from the original on 2008-08-22. http://web.archive.org/web/20080822143815/http://www.sysprog.net/quotlaws.html. Retrieved 2007-03-10. 
  3. ^ Ungar, S. (2000). "Knowledge, ignorance and the popular culture: climate change versus the ozone hole". Public Understanding of Science 9 (3): 297–312. 
  4. ^ Pidgeon, N.; B. Fischhoff (2011). "The role of social and decision sciences in communicating uncertain climate risks". Nature Climate Change 1 (1): 35–41. 
  5. ^ Kahan, Dan M.; Maggie Wittlin, Ellen Peters, Paul Slovic, Lisa Larrimore Ouellette, Donald Braman, Gregory N. Mandel (2011). "The Tragedy of the Risk-Perception Commons: Culture Conflict, Rationality Conflict, and Climate Change". SSRN eLibrary. http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1871503&http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1871503. Retrieved 2011-11-26. 
  6. ^ Kahneman, Daniel (2003-12-01). "Maps of Bounded Rationality: Psychology for Behavioral Economics". The American Economic Review 93 (5): 1449–1475. ISSN 0002-8282. http://www.jstor.org/stable/3132137. Retrieved 2011-11-26. 
  7. ^ Tversky, A.; D. Kahneman (1974). "Judgment under uncertainty: Heuristics and biases". Science 185 (4157): 1124. 
  8. ^ Joireman, Jeff; Heather Barnes Truelove, Blythe Duell (2010-12). "Effect of outdoor temperature, heat primes and anchoring on belief in global warming". Journal of Environmental Psychology 30 (4): 358–367. doi:10.1016/j.jenvp.2010.03.004. ISSN 0272-4944. http://www.sciencedirect.com/science/article/pii/S0272494410000319. Retrieved 2011-11-26. 
  9. ^ Saul, Stephanie; Andrew Pollack (2007-02-17). "Furor on Rush to Require Cervical Cancer Vaccine". The New York Times. ISSN 0362-4331. http://www.nytimes.com/2007/02/17/health/17vaccine.html. Retrieved 2011-11-26. 
  10. ^ Kahan, Dan M.; Donald Braman, Geoffrey L. Cohen, Paul Slovic, John Gastil (2008-07-15). "Who Fears the HPV Vaccine, Who Doesn't, and Why? An Experimental Study of the Mechanisms of Cultural Cognition". SSRN eLibrary. http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1160654. Retrieved 2011-11-26. 
  11. ^ Lord, Charles G.; Lee Ross, Mark R. Lepper (1979). "Biased assimilation and attitude polarization: The effects of prior theories on subsequently considered evidence.". Journal of Personality and Social Psychology 37: 2098–2109. doi:10.1037/0022-3514.37.11.2098. ISSN 0022-3514. http://psycnet.apa.org/psycinfo/1981-05421-001. Retrieved 2011-11-26. 
  12. ^ HOVLAND, CARL I.; WALTER WEISS (1951-12-21). "The Influence of Source Credibility on Communication Effectiveness". Public Opinion Quarterly 15 (4): 635–650. doi:10.1086/266350. http://poq.oxfordjournals.org/content/15/4/635.abstract. Retrieved 2011-11-27. 
  13. ^ a b Braman, Donald; James Grimmelmann, Dan M. Kahan. "Modeling Cultural Cognition". SSRN eLibrary. http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1000449. Retrieved 2011-12-06. 
  14. ^ Fremling, G.M.; J.R. Lott Jr (2002). "Surprising Finding That Cultural Worldviews Don't Explain People's Views on Gun Control, The". U. Pa. L. Rev. 151: 1341. 
  15. ^ Ayres, I.; J.J. Donohue III (2002). "Shooting down the more guns, less crime hypothesis". National Bureau of Economic Research. 
  16. ^ Lee, M.D.; M. Steyvers, M. de Young, B.J. Miller. A Model-Based Approach to Measuring Expertise in Ranking Tasks. 
  17. ^ Ernst, Marc O.; Martin S. Banks (2002-01-24). "Humans integrate visual and haptic information in a statistically optimal fashion". Nature 415 (6870): 429–433. doi:10.1038/415429a. ISSN 0028-0836. http://dx.doi.org/10.1038/415429a. Retrieved 2011-11-27. 
  18. ^ Wozny, D.R.; U.R. Beierholm, L. Shams (2008). "Human trimodal perception follows optimal statistical inference". Journal of vision 8 (3). 
  19. ^ Brocas, Isabelle; Juan D. Carrillo. "From perception to action: An economic model of brain processes". Games and Economic Behavior (0). doi:10.1016/j.geb.2011.10.001. ISSN 0899-8256. http://www.sciencedirect.com/science/article/pii/S0899825611001758. Retrieved 2011-11-27.