Distraction, not partisanship, drives sharing of misinformation

The words

Enlarge (credit: Lewis Ogden / Flickr)

We don’t need a study to know that misinformation is rampant on social media; we just need to do a search for “vaccines” or “climate change” to confirm that. A more compelling question is why. It’s clear that, at a minimum, there are contributions from organized disinformation campaigns, rampant political partisans, and questionable algorithms. But beyond those, there are still a lot of people who choose to share stuff that even a cursory examination would show was garbage. What’s driving them?

That was the question that motivated a small international team of researchers who decided to take a look at how a group of US residents decided on which news to share. Their results suggest that some of the standard factors that people point to when explaining the tsunami of misinformation—inability to evaluate information and partisan biases—aren’t having as much influence as most of us think. Instead, a lot of the blame gets directed at people just not paying careful attention.

You shared that?

The researchers ran a number of fairly similar experiments to get at the details of misinformation sharing. This involved panels of US-based participants recruited either through Mechanical Turk or via a survey population that provided a more representative sample of the US. Each panel had several hundred to over 1,000 individuals, and the results were consistent across different experiments, so there was a degree of reproducibility to the data.

Read 14 remaining paragraphs | Comments

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top