Even in the era of fake news, facts can shape our opinions – Grist

In a world wherepost-truth was 2016s word of the year, many people are starting to doubt the efficacy of facts. Can science make sense of anti-science and post-truthism? More generally, how can we understand what drives peoples beliefs, decisions, and behaviors?

Scientists have developed many theories to describehow people process and think about information. Unfortunately, theres an increasing tendency to see people as creatures whose reasoning mechanisms are largely dependent on a narrow set of processes. For example, one popular theory suggests that if we just communicate more accurate information to people, their behavior will change accordingly. Another suggests that people will reject evidence if it threatens their deeply held cultural worldviews and associated feelings.

Its more important than ever that our approach to communication is evidence-based and built on a strong, theoretical foundation. Many of these models contribute valuable insights and can help us design better communication, but each on its own is incomplete. And science communicators have a tendency to oversimplify, focusing on a single model and disregarding other theories.

We suggest that this is a dangerous practice and less effective thana more nuanced and holistic view. The apparent choice between fact and feeling, or between cognition and culture, is a false dilemma. In reality, both are related and address different pieces of the decision-making puzzle.

One well-known theory about how people absorb new facts is the information deficit model. The main idea here is straightforward: If you throw more facts at people, theyll eventually come around on an issue.

Most behavioral science scholars agree that this model of human thinking and behavioris clearly incomplete people rely on a range of other cues besides facts in guiding their attitudes and behavior. For example, sometimes we simply act based on how we feel about an issue. Unfortunately, the facts dont always convince.

But the term information deficit is problematic, too. People tend to have limited information in most areas of life. For example, we often dont know the thoughts and feelings of other people we trust and value. Similarly, we might have limited knowledge about appropriate cultural norms when traveling to a new country, and so on. Information deficit isnt a very meaningful term to use to theorize about human thinking.

Another theory about human thinking is called cultural cognition. In brief, it suggests that our cultural values and worldviews shapehow we think about science and society.

Its easy to be duped into thinking of the human brain as a sponge that soaks up only the information it wants to believe. For example, the theory suggests that peoples position on divisive issues such as climate change is not informed by scientific evidence, but rather by the strong commitment people have to their political groups and ideologies. The idea is that to protect our cultural worldviews, we actively reject evidence that threatens them think of someone who fears that government action on climate change undermines the free market.

In short, this narrative sounds appealing on the surface, as humans organize themselves in groups, and much psychological research has shown that we derive part of oursocial identities from the group affiliations we maintain.

Yet, its focus is overly narrow, and theres a logical fallacy in this conception of human thinking. We belong to many groups at any given time and we juggle many different public and private identities. The real question is about nuance; when and under what conditions is someone motivated to reject scientific facts in favor of their cultural worldview?

To throw all our fact-disseminating eggs into one or the other theoretical basket is oversimplistic and deprives us of important insights.

A more nuanced perspective recognizes that facts and information are embedded in social and cultural contexts. For example, peoples perception of expert consensus matters a great deal, especially on contested issues, and is often described as agateway belief that influences a range of other attitudes about an issue. The near-unanimous consensus that vaccines do not cause autism or that climate change is human-caused are all scientific facts. At the same time, consensus information is also inherently social: It describes the extent of agreement within an influential group of experts.

People often want to beaccurate in their views, and, in an uncertain world bounded by limited time and effort, we make strategic bets on what information to take into account. Consensus acts as a natural heuristic, or mental shortcut, for complicated scientific issues.Numerousstudies have found that highlighting scientific agreement on human-caused global warming can help neutralize and reduce conflicting views about climate change.

Similarly, while some studies have found a limited effect of knowledge on judgment, when you dig deeper into the data, a more nuanced and insightful picture emerges. For example, some studies claim that adeficit in scientific knowledge does not explain why people are divided on contested issues such as climate change. But whats being measured in these experiments matters. Indeed, indicators such as how well people understand numbers or their scientific literacy which is what some of these studies actually quantify are categorically different from measuring specific knowledge people have about a topic such as climate change. In fact, a survey across six countries found that when peopleunderstand the causes of climate change, their concern increases accordingly, irrespective of their values. Similarly,otherstudies show that explanations about the mechanisms of climate change can reduce biased evaluations of evidence as well as political polarization.

In short, facts do matter.

Indeed, there is no need to throw outthe baby with the bathwater. Instead, we need to dispel false dichotomies and folk psychology about human thinking that currently dominate the media. Repeating the story that people dont care about facts runs the risk of becoming a self-fulfilling prophecy. A holistic view acknowledges that people rely on cognitive shortcuts and emotions, care about social norms and group identities and are sometimes motivated in their reasoning, but it also recognizes the research showing that most people want to fundamentally hold accurate perceptions about the world.

This is particularly important as the public is currently hampered by misinformation and fake news. In two separatestudies, we each found that misinformation about climate change has a disproportionate influence on public attitudes and opinion. However, we also found that inoculating people against the false arguments neutralized misinformations influence, across the political spectrum. In essence, teaching people what false arguments might be deployed helped them overcome their cultural biases. Other work similarlyshows that the politicization of science can be counteracted with inoculation.

People are complex, social, and affected by a diverse range of influences depending on the situation. We want to hold accurate views, but emotion, group identities, and conflicting goals can get in the way. Incorporating these different insights into human thinking enriches our understanding of how people form opinions and make decisions.

Effective science communication requires an inclusive, holistic approach that integrates different social science perspectives. To simplistically focus on a single perspective paints a limited and increasingly inaccurate view of how humans form judgments about social and scientific issues.

Link:
Even in the era of fake news, facts can shape our opinions - Grist

Related Posts