Programmed to believe: Why people buy into falsehoods
Loading...
If you鈥檝e ever held a political conviction, you鈥檝e probably found yourself asking, 鈥淗ow on Earth can those people be so gullible?鈥
Not your people, of course, the reasonable and well-informed ones who share your politics. No, we mean the millions on the other side, the ones who have chosen to stick their tinfoil-covered heads in the sand to swallow the tripe-flavored Kool-Aid. Those people.聽
Except it turns out that, under the right conditions, any of us can be played. Behavioral scientists say that many of our beliefs about the world, true as they may be, are more likely to be determined by something other than our power to discriminate truth from fiction. And that鈥檚 not necessarily a bad thing.
Why We Wrote This
In a 鈥減ost-truth鈥 world, the media has struggled with how to address false political statements. Our reporter took a step back and examined why falsehoods persist.
鈥淭here's significant evidence that we evolved as a cooperative species, and trust is central to that,鈥 says Pamela Paxton, a sociologist at the University of Texas at Austin who studies social capital. 鈥淭here are a lot of benefits that we get, even in terms of efficiency, from high levels of trust.鈥
A bias toward belief
It all has to do with the way we process information. Until relatively recently, most psychologists implicitly adopted a model of thought developed in the 17th century by French philosopher Ren茅 Descartes, who, in his 鈥,鈥 divided the process of believing into two distinct faculties. First, what he called the 鈥渋ntellect鈥 comprehends an idea. Next, the intellect presents the idea to the 鈥渨ill,鈥 which then attempts to determine whether the idea is true or not.
Descartes鈥檚 model makes intuitive sense, but beginning in the late 1980s and early 1990s, psychologists began to see flaws with it. A series of experiments, led by psychologist Daniel Gilbert, that distracted participants while asking them to identify sets of true and false sentences found that the distraction interfered with the participants鈥 ability to identify the false statements, .
Dr. Gilbert proposed a different model of believing, one consistent with that of Descartes鈥檚 rival, the Dutch philosopher Baruch (Benedict) Spinoza, who proposed that comprehending an idea and believing it to be true are . Rejecting an idea, Spinoza said, requires an additional step that uses up additional mental resources.
Put another way, a belief is like an automatic email newsletter: you have to go out of your way to figure out how to unsubscribe.
From an efficiency standpoint, this arrangement makes sense. 鈥淚f you're trying to determine whether something you heard is true or false,鈥 says Lisa Fazio, a psychologist at Vanderbilt University in Nashville, Tenn., 鈥測ou can do this kind of labor-intensive search through your long-term memory to see what you know about the topic and whether or not it seems reasonable, or you can just go with this kind of gut-level feeling of, 鈥楾hat seems true.鈥 鈥
And your gut works just fine, in most cases. 鈥淢ost of the time, most humans are trying to tell each other things that are true and useful to them,鈥 says Gus Cooney a researcher at Harvard University in Cambridge, Mass., who studies the psychology of conversations. Dr. Cooney says he suspects that this was even more true in our distant evolutionary past, when humans lived in small, stable groups and there was no advertising.
In the modern world, however, our natural credulity can be hacked. 鈥淭he problem comes when we when we take that rule that's useful in one domain and then we misapply in a domain where people are trying to intentionally mislead us,鈥 says Cooney.
Say it again
The top three most common hacks, exploited by politicians for centuries, are repetition, repetition, and repetition. The more times we hear an assertion, say psychologists, the less effort it takes to wrap our minds around it, and humans, for some reason, confuse that ease with truthfulness.
鈥淚f you just repeat things,鈥 says Professor Fazio, 鈥測ou can make them gain this processing fluency while not actually changing the truth of the statement.鈥
So does this mean that the solution to political misinformation is for people to become less trusting of each other? Probably not. 鈥淭rust eases interactions between people,鈥 says Professor Paxton, 鈥渟o you don't have to rely on third parties to ensure or watch or regulate.鈥
Indeed, in 2010 economists found that places that rank high in interpersonal trust, such as the Nordic countries, generally have than low-trusting countries, such as Brazil and Uganda. More importantly, people living in more trusting societies generally say they are happier.
In the United States, trust has been steadily declining over the past 40 years, a phenomenon that Paxton attributes in part to headline-grabbing scandals. 鈥淚 suspect it may just be that, with more and more news,鈥 she says, 鈥渨e're just more aware constantly of negative information about institutions.鈥
Instead of harboring mistrust, Fazio suggests that, when you鈥檙e presented with a statement, you can 鈥減ause and think about it.鈥
鈥淭he one thing that we found that helps people to notice these errors, is to really slow down and read the article, like a fact checker where they have to circle the errors in the story,鈥 she says. 鈥淎nd those things work.鈥
But even this method, says Fazio, won鈥檛 catch every factual error. There鈥檚 no perfect solution, which means that we鈥檒l have to learn to live with some of us being wrong sometimes.
鈥淥f course I'd love people to be more skeptical, and approach the world like empiricists and scientists,鈥 says Cooney. 鈥淏ut I think being a little more humble and a little more tolerant might be the right answer.鈥
Spinoza, who lived through the end of a bloody religious war that consumed most of Europe, likely would have agreed. In his 鈥淓thics,鈥 written in 1664 or 1665, he wrote that his doctrine that seeing and believing are one and the same 鈥漷eaches us to hate no man, neither to despise, to deride, to envy, or to be angry with any.鈥