Jump to content

Cultural cognition of risk

From Wikipedia, the free encyclopedia
(Redirected from Cultural cognition)

The cultural cognition of risk, sometimes called simply cultural cognition, is the hypothesized tendency to perceive risks and related facts in relation to personal values. Research examining this phenomenon draws on a variety of social science disciplines including psychology, anthropology, political science, sociology, and communications. The stated objectives of this research are both to understand how values shape political conflict over facts (like whether climate change exists, whether vaccination of school girls for HPV threatens their health) and to promote effective deliberative strategies for resolving such conflicts consistent with sound empirical data.

Theory and evidence

[edit]

The cultural cognition hypothesis holds that individuals are motivated by a variety of psychological processes to form beliefs about putatively dangerous activities that match their cultural evaluations of them. Persons who subscribe to relatively individualistic values, for example, tend to value commerce and industry and are inclined to disbelieve that such activities pose serious environmental risks. Persons who subscribe to relatively egalitarian and communitarian values, in contrast, readily credit claims of environmental risks, which is consistent with their moral suspicion of commerce and industry as sources of inequality and symbols of excessive self-seeking.[1][2]

Scholars have furnished two types of evidence to support the cultural cognition hypothesis. The first consists of general survey data that suggest that individuals’ values more strongly predict their risk perceptions than do other characteristics such as race, gender, economic status, and political orientations.[3][4]

The second type of evidence consists in experiments that identify discrete psychological processes that connect individuals’ values to their beliefs about risk and related facts.[5] Such experiments suggest, for example, that individuals selectively credit or dismiss information in a manner that reinforces beliefs congenial to their values.[6] They also show that individuals tend to be more persuaded by policy experts perceived to hold values similar to their own rather than by ones perceived to hold values different from them.[7] Such processes, the experiments suggest, often result in divisive forms of cultural conflict over facts, but can also be managed in fashions that reduce such disagreement.[8]

Cultural cognition project at Yale Law School

[edit]

Funded by governmental and private foundation grants,[when?] much of the work on cultural cognition has been performed by an interdisciplinary group of scholars affiliated with the Cultural Cognition Project.[9] There are currently[when?] over a dozen project members from a variety of universities. Two members of the project—Dan Kahan and Douglas Kysar—are Yale Law School faculty, although other members (such as Donald Braman of George Washington University Law School and Geoffrey Cohen of Stanford University) were previously affiliated with Yale Law School or Yale University. Students from Yale University also contribute to Project research.

Significant findings

[edit]

Science comprehension and cultural polarization

[edit]

A study conducted by Cultural Cognition Project researchers (using a nationally representative U.S. sample) found that ordinary members of the public do not become more concerned about climate change as their science comprehension increases.[10] However, the degree of polarization among cultural groups with opposing predispositions increases.

Nanotechnology

[edit]

The Cultural Cognition Project has conducted a series of studies on public perceptions of nanotechnology risks and benefits. Combining survey and experimental methods, the studies present evidence that individuals culturally predisposed to be skeptical of environmental risks are both more likely to seek out information on nanotechnology and more likely to infer from that information that nanotechnology’s benefits will outweigh its risks. Individuals culturally predisposed to credit environmental risks construe that same information, when exposed to it in the lab, as implying that nanotechnology’s risks will predominate.[6] The studies also present evidence that individuals tend to credit expert information on nanotechnology—regardless of its content—based on whether they share the perceived cultural values of the expert communicator.[11] The studies were issued by the Project on Emerging Nanotechnologies at the Woodrow Wilson International Center for Scholars, one of the research sponsors.

"Scientific consensus"

[edit]

The same dynamics that motivate individuals of diverse cultural outlooks to form competing perceptions of risks are likely to cause them to form opposing perceptions of "scientific consensus", cultural cognition researchers have concluded.[12] In an experimental study, the researchers found that subjects were substantially more likely to count a scientist (of elite credentials) as an "expert" in his field of study when the scientist was depicted as taking a position consistent with the one associated with the subjects' cultural predispositions than when that scientist took a contrary position. A related survey revealed that members of different cultural groups have significantly divergent views on what most scientific experts believe on various issues, highlighting the common occurrence of culturally biased recognition of who qualifies as an "expert." The study found that across a variety of risks (such as climate change, nuclear waste disposal, and private handgun possession), no single cultural group was more likely than any other to hold perceptions of scientific consensus that consistently aligned with those presented in "expert consensus reports" issued by the U.S. National Academy of Sciences.

Law

[edit]

Scholars have also applied the cultural cognition of risk to legal issues. One such study examined how individuals reacted to a videotape of a high-speed police chase. In Scott v. Harris,[13] the U.S. Supreme Court (by a vote of 8-1) had held that no reasonable jury could view the tape and fail to find that the driver posed a lethal risk to the public large enough to justify deadly force by the police (namely, ramming the fleeing driver's vehicle, causing it to crash). The majority of study subjects agreed with the Court, but there were significant divisions along cultural lines.[14] Other studies have found that individuals' cultural worldviews influence their perceptions of consent in an acquaintance or date rape scenario,[15] and of the imminence of violence and other facts in self-defense cases involving either battered women or interracial confrontations.[16]

Relationship to other risk perception theories

[edit]

Cultural cognition is a descendant of two other theories of risk perception. The first is the cultural theory of risk associated with anthropologist Mary Douglas and political scientist Aaron Wildavsky.[17] The cultural cognition hypothesis is derived from Douglas and Wildavsky's claim, advanced most notably in their controversial book Risk and Culture: An Essay on the Selection of Technical and Environmental Dangers (1982), that individuals selectively attend to risks in a manner that expresses and reinforces their preferred way of life.[citation needed]

Cultural cognition researchers, along with other scholars who have investigated Douglas and Wildavsky's theory empirically,[18] use attitudinal scales that reflect Douglas's worldview typology. That typology characterizes worldviews, or preferences about how society should be organized, along two cross-cutting dimensions: "group", which refers to how individualistic or group-oriented a society should be; and "grid", which refers to how hierarchical or egalitarian a society should be.[19]

The second theory is the "psychometric paradigm", to which Paul Slovic, a member of the Cultural Cognition Project, has made significant contributions. The psychometric paradigm links risk perceptions to various cognitive and social mechanisms that generally evade simpler, rational choice models associated with economics.[20][21] Cultural cognition theory posits that these mechanisms mediate between, or connect, individuals' cultural values to their perceptions of risk and other policy-relevant beliefs.

Combining the cultural theory of risk and the psychometric paradigm, cultural cognition, its exponents claim, remedies difficulties with each.[22] The mechanisms featured in the psychometric paradigm (and in social psychology generally) furnish a cogent explanation of why individuals adopt states of mind that fit and promote the aims of groups, including ones featured in Douglas’s culture theory. They do so, moreover, in a manner that avoids "functionalism," a criticized form of analysis that identifies group interests, rather than individual ones, as a cause for human action.[23][24] At the same time cultural theory, by asserting the orienting role of values, explains how the mechanisms featured in the psychometric paradigm can result in differences in risk perception among persons who hold different values. The interrelationship between individual values and perceptions of risk also calls into doubt the depiction of risk perceptions deriving from these mechanisms as products of irrationality or cognitive defect.[25]

Criticisms

[edit]

Cultural cognition has been subjected to criticisms from a variety of sources. The rational choice economists Fremling & Lott (2003), as well as the psychologist Sjöberg (1998) have suggested that the theory (and others based on the cultural theory of risk generally) explain only a small fraction of the variation in popular risk perceptions. Mary Douglas herself has criticized cultural cognition for a conception of values that is too tightly modeled on American political disputes and that implicitly disparages the "hierarchical" worldview.[26] Finally, some scholars who emphasize elements of the psychometric paradigm suggest that the influence of cultural values on risk perceptions is best understood as simply an additional source of interference with the rational processing of information.[27]

See also

[edit]

Notes

[edit]
  1. ^ Kahan (2010a), p. 296.
  2. ^ Kahan et al. (2006), pp. 1083–84.
  3. ^ Kahan & Braman (2006), pp. 155–158.
  4. ^ Kahan et al. (2006), pp. 1086–87.
  5. ^ Kahan et al. (2007), p. ?.
  6. ^ a b Kahan et al. (2009), p. ?.
  7. ^ Kahan et al. (2010), p. ?.
  8. ^ Kahan (2007), p. 138-139.
  9. ^ "Home". culturalcognition.net.
  10. ^ Kahan et al. (2012), p. ?.
  11. ^ Kahan et al. (2008), pp. 5, 14.
  12. ^ Kahan, Jenkins-Smith & Braman (2010), p. 27.
  13. ^ 550 U.S. 372 (2007).
  14. ^ Kahan, Hoffman & Braman (2009), p. 837.
  15. ^ Kahan (2010b).
  16. ^ Kahan & Braman (2008).
  17. ^ Rayner (1992), p. ?.
  18. ^ For example, Dake (1991), Jenkins-Smith (2001), and Peters & Slovic (1996)
  19. ^ Rayner (1992), pp. 87–91.
  20. ^ Slovic (2000), p. ?.
  21. ^ Kahneman, Slovic & Tversky (1982), p. ?.
  22. ^ Kahan (2008), p. ?.
  23. ^ Boholm (1996), pp. 68, 79–80.
  24. ^ Kahan & Braman (2006), p. 252.
  25. ^ Kahan et al. (2006), pp. 1088–1106.
  26. ^ Douglas (2003).
  27. ^ Kahan et al. (2006).

References

[edit]

Further reading

[edit]
[edit]