What do we study when we study misinformation? A scoping review of experimental research (2016-2022)
Gillian Murphy, Constance de Saint Laurent, Megan Reynolds, Omar Aftab, Karen Hegarty, Yuning Sun and Ciara M. Greene
We reviewed 555 papers published from 2016–2022 that presented misinformation to participants. We identified several trends in the literature—increasing frequency of misinformation studies over time, a wide variety of topics covered, and a significant focus on COVID-19 misinformation since 2020. We also identified several important shortcomings, including overrepresentation of samples from the United States and Europe and excessive emphasis on short-term consequences of brief, text-based misinformation.

User experiences and needs when responding to misinformation on social media
Pranav Malhotra, Ruican Zhong, Victor Kuan, Gargi Panatula, Michelle Weng, Andrea Bras, Connie Moon Sehat, Franziska Roesner and Amy Zhang
This study examines the experiences of those who participate in bottom-up user-led responses to misinformation on social media and outlines how they can be better supported via software tools. Findings show that users desire support tools designed to minimize time and effort in identifying misinformation and provide tailored suggestions for crafting responses to misinformation that account for emotional and relational context.
Increasing accuracy motivations using moral reframing does not reduce Republicans’ belief in false news
Michael Stagnaro, Sophia Pink, David G. Rand and Robb Willer
In a pre-registered survey experiment with 2,009 conservative Republicans, we evaluated an intervention that framed accurate perceptions of information as consistent with a conservative political identity and conservative values (e.g., patriotism, respect for tradition, and religious purity). The intervention caused participants to report placing greater value on accuracy, and placing greater value on accuracy was correlated with successfully rating true headlines as more accurate than false headlines.
“Fact-checking” fact checkers: A data-driven approach
Sian Lee, Aiping Xiong, Haeseung Seo and Dongwon Lee
This study examined four fact checkers (Snopes, PolitiFact, Logically, and the Australian Associated Press FactCheck) using a data-driven approach. First, we scraped 22,349 fact-checking articles from Snopes and PolitiFact and compared their results and agreement on verdicts. Generally, the two fact checkers agreed with each other, with only one conflicting verdict among 749 matching claims after adjusting minor rating differences.

Misinformation reloaded? Fears about the impact of generative AI on misinformation are overblown
Felix M. Simon, Sacha Altay and Hugo Mercier
Many observers of the current explosion of generative AI worry about its impact on our information environment, with concerns being raised about the increased quantity, quality, and personalization of misinformation. We assess these arguments with evidence from communication studies, cognitive science, and political science.
Exploring partisans’ biased and unreliable media consumption and their misinformed health-related beliefs
Natasha Strydhorst, Javier Morales-Riech and Asheley R. Landrum
This study explores U.S. adults’ media consumption—in terms of the average bias and reliability of the media outlets participants report referencing—and the extent to which those participants hold inaccurate beliefs about COVID-19 and vaccination. Notably, we used a novel means of capturing the (left-right) bias and reliability of audiences’ media consumption, leveraging the Ad Fontes Media ratings of 129 news sources along each dimension.

A focus shift in the evaluation of misinformation interventions
Li Qian Tay, Stephan Lewandowsky, Mark J. Hurlstone, Tim Kurz and Ullrich K. H. Ecker
The proliferation of misinformation has prompted significant research efforts, leading to the development of a wide range of interventions. There is, however, insufficient guidance on how to evaluate these interventions. Here, we argue that researchers should consider not just the interventions’ primary effectiveness but also ancillary outcomes and implementation challenges.
Assessing misinformation recall and accuracy perceptions: Evidence from the COVID-19 pandemic
Sarah E. Kreps and Douglas L. Kriner
Misinformation is ubiquitous; however, the extent and heterogeneity in public uptake of it remains a matter of debate. We address these questions by exploring Americans’ ability to recall prominent misinformation during the COVID-19 pandemic and the factors associated with accuracy perceptions of these claims.

Did the Musk takeover boost contentious actors on Twitter?
Christopher Barrie
After his acquisition of Twitter, Elon Musk pledged to overhaul verification and moderation policies. These events sparked fears of a rise in influence of contentious actors—notably from the political right. I investigated whether these actors did receive increased engagement over this period by gathering tweet data for accounts that purchased blue-tick verification before and after the Musk takeover.
Who knowingly shares false political information online?
Shane Littrell, Casey Klofstad, Amanda Diekman, John Funchion, Manohar Murthi, Kamal Premaratne, Michelle Seelig, Daniel Verdear, Stefan Wuchty and Joseph E. Uscinski
Some people share misinformation accidentally, but others do so knowingly. To fully understand the spread of misinformation online, it is important to analyze those who purposely share it. Using a 2022 U.S. survey, we found that 14 percent of respondents reported knowingly sharing misinformation, and that these respondents were more likely to also report support for political violence, a desire to run for office, and warm feelings toward extremists.