In the run-up to the 2020 election, people appear to have become better at spotting misinformation online, according to a study that found a decline in clicks on unreliable websites.
According to prior research, some 44.3% of Americans visited websites during the 2016 US election that repeatedly made false or misleading information. During the 2020 election, that number dropped by nearly half to 26.2%.
While these findings are promising, the researchers are cautious in interpreting the results. Exposure even among fewer people can still have serious consequences, they note in the new paper in .
Extrapolating their results, they estimated that nearly 68 million Americans made a total of 1.5 billion visits to untrustworthy websites during the 2020 election.
“Although we saw a serious reduction in the overall number of people exposed to misinformation on the web, misinformation remains a serious problem in the information ecosystem for some populations, especially older adults and diverse communities,” says senior author Jeff Hancock, a professor of communication at Stanford University.
The researchers found that those who did visit websites touting false claims tended to be older and lean more to the right of the political spectrum, a finding consistent with 2016 data. They did however visit fewer untrustworthy websites and spend less time on them than they did in 2016.
Misinformation purveyors target older adults
The study builds on previous research conducted by Andrew Guess at Princeton University. In 2016, Guess compiled a list of some 490 untrustworthy websites that included pages that prominent researchers in the disinformation research community had previously identified, including Stanford economist Matthew Gentzkow.
For the current study, Hancock, along with Stanford PhD students Ryan Moore and Ross Dahlke, augmented the list with an additional 1,240 unreliable domains from , an organization that rates the credibility of news and information websites.
Their rankings are done manually by experienced journalists and editors who rigorously review and rank websites on a variety of criteria, including whether they repeatedly publish false content, issue corrections on errors in their reporting, and distinguish between news and opinion.
The researchers then recruited a representative sample of 1,151 American adults through the polling firm, YouGov. Participants completed an online survey and installed a browser plugin that allowed the researchers to passively track web activity between October 2, 2020, and November 9, 2020. In sum, they gathered some 7.5 million website visits on users’ desktop and mobile devices.
The researchers found that in 2020, 5.6% of visits to untrustworthy websites were referred to by Facebook—in 2016, it was 15.1%. They credit this decrease to efforts the social media platform took to mitigate the issue of false news on the website.
“The drop in visits referred by Facebook may reflect investment in trust and safety efforts to decrease the prevalence of misinformation on their platform, such as flagging, content moderation, and user education, which they and other platforms weren’t doing as much of in 2016,” Moore says.
While the researchers found that the average number of times a person visited a misinformation website decreased from an average of 32 visits in 2016 to 23 visits in 2020, there are a few individuals who still consumed misinformation online at extremely high clicks.
“There are some people still consuming hundreds of misinformation websites,” says Dahlke. “We need more research to understand the effect of this type of exposure on people’s beliefs and actions.”
The researchers also found that older adults were twice as likely to visit a misinformation website compared to those aged 18-29 years old. While a smaller percentage of Americans 65 and older were exposed in 2020 (56.2%) than in 2016 (37.4%), they continue to consume misinformation at much higher rates than younger adults.
“Older adults continue to be targeted by misinformation purveyors because that generation tends to be wealthier and more civically engaged than other generations, making them prime targets for bad actors trying to make money or change election outcomes,” says Hancock.
What’s ahead for 2024 election?
Misinformation is pernicious, it morphs and mutates quickly, the researchers say.
“While one could interpret our findings as evidence that the problem of online misinformation is improving in some way, they could also be interpreted as evidence that the nature of the problem is changing,” the scholars write in the paper.
The researchers only studied web browsing activity, and misinformation could have been displaced to other social media platforms or encrypted messaging services, such as WhatsApp or Signal.
Moreover, a click is not the only metric of fake news consumption; people could have still consumed untrustworthy information passively online through a meme or even just skimming a headline when scrolling through news feeds. All of these factors make it a difficult topic to study.
Hancock, Moore, and Dahlke are already thinking about what their findings might reveal about how misinformation will spread in the next general election in 2024.
They anticipate that older adults will continue to be vulnerable to fake news, an issue that Hancock’s Social Media Lab has separately been working to address with support from the Stanford Impact Lab program. In 2020, Hancock and Moore collaborated with the nonprofit news organization Poynter to create a digital media literacy intervention to help seniors identify misinformation online.
They are also concerned about the role of misinformation in under-resourced areas, such as non-English speaking communities, as highlighted in a recent paper Hancock and Moore coauthored with Stanford PhD student Angela Y. Lee on the topic.
Source: