Why does the presentation of persuasive evidence — even evidence of a scientific consensus — so often fail to resolve political debates? How is it, for example, that so much of the American public on the right refuses to accept the scientific consensus regarding the causes and risks of climate change?
For a while now, I’ve thought that Dan Kahan’s theory of “cultural cognition” offered the most persuasive answer to these questions. Kahan rejects the idea that the problem lies in Republicans’ lack of information about climate science. Offering more evidence isn’t going to resolve the issue at this point. It might even aggravate the problem.
Rather, Kahan offers empirical evidence that the Republican resistance to climate science is an example of a more general phenomenon: the human tendency to arrive at conclusions that are congenial to our cultural values, and to resist, dismiss, or attack conclusions that threaten our values and identities.
But the more I’ve learned about the specifics of the cultural cognition theory, the more I’ve felt like it leaves something out.
In this post, I’d like to propose a hypothesis that complements cultural cognition’s explanation for the frequent failures of evidence-based discussion to lead to increased agreement on politically charged issues. When I first heard about Kahan’s work, I thought that the theory I’m about to present was what he meant by “cultural cognition.” But as I’ve read more about his work, it’s become clear to me that the idea I have in mind is a distinct one.
I’ll call the hypothesis “tribal cognition.”