… 7 studies you should know about …
What better way to start the new year than by learning new things about how best to battle fake news and other forms of online misinformation?
Below is a sampling of the research published in 2019 — seven journal articles that examine fake news from multiple angles, including what makes fact-checking
most effective and the potential use of crowdsourcing to help detect false content on social media.
Because getting good news is also a great way to start 2020, I included a study that suggests President Donald Trump’s “fake news” tweets aimed at discrediting news
coverage could actually help journalists.
The authors of that paper recommend journalists “engage in a sort of news jujitsu, turning the negative energy of Trump’s tweets into a force for creating
additional interest in news.”
This article was first published at Nieman Lab.
This study provides several new insights about the most effective ways to counter fake news on social media.
Researchers found that when fake news headlines were flagged with a tag that says “Rated false,” people were less likely to accept the headline as accurate than when
headlines carried a “Disputed” tag.
They also found that posting a general warning telling readers to beware of misleading content could backfire.
After seeing a general warning, study participants were less likely to believe true headlines and false ones.
The authors note that while their sample of 2,994 U.S. adults isn’t nationally representative, the feedback they got demonstrates that online fake news can be countered
“with some degree of success.” “
The findings suggest that the specific warnings were more effective because they reduced belief solely for false headlines and did not create spillover effects
on perceived accuracy of true news,” they write.
It would be time-consuming and expensive to hire crowds of professional fact-checkers to find and flag all the false content on social media.
But what if the laypeople who use those platforms pitched in?
Could they accurately assess the trustworthiness of news websites, even if prior research indicates they don’t do a good job judging the reliability
of individual news articles?
This research article, which examines the results of two related experiments with almost 2,000 participants, finds the idea has promise.
“We find remarkably high agreement between fact-checkers and laypeople,” the authors write.
“This agreement is largely driven by both laypeople and fact-checkers giving very low ratings to hyper-partisan and fake news sites.”
The authors note that in order to accurately assess sites, however, people need to be familiar with them.
When news sites are new or unfamiliar, they’re likely to be rated as unreliable, the authors explain.
Their analysis also finds that Democrats were better at gauging the trustworthiness of media organizations than Republicans
— their ratings were more similar to those of professional fact checkers.
Republicans were more distrusting of mainstream news organizations.
When Trump turns to Twitter to accuse legitimate news outlets of being “fake news,” does the public’s view of journalists change?
Are people who read his tweets less likely to believe news coverage?
To investigate such questions, researchers conducted two studies, during which they showed some participants a sampling of the president’s “fake news” tweets
and asked them to read a news story.
Here’s what the researchers learned:
The more tweets people chose to read, the greater their intent to read more news in the future.
As participants read more tweets, their assessments of news stories’ and journalists’ credibility also rose.
“If anything, we can conclude that Trump’s tweets about fake news drive greater interest in news more generally,” the authors write.
The authors’ findings, however, cannot be generalized beyond the individuals who participated in the two studies
— 331 people for the first study and then 1,588 for the second, more than half of whom were undergraduate students.
Based on their findings, the researchers offer a few suggestions for journalists.
“In the short term,” they write, “if journalists can push out stories to social media feeds immediately after Trump or others tweet about legitimate news as being
‘fake news,’ then practitioners may disarm Trump’s toxic rhetoric and even enhance the perceived credibility of and demand for their own work.
Using hashtags, quickly posting stories in response to Trump, and replying directly to him may also tether news accounts to the tweets in social media feeds.”
This study looks at whether news outlets or public figures have a greater influence on people’s perception of a news article’s trustworthiness.
The findings suggest that when a public figure such as Oprah Winfrey or Dr. Oz shares a news article on social media, people’s attitude toward the article is linked to how
much they trust the public figure.
A news outlet’s reputation appears to have far less impact.
In fact, researchers found mixed evidence that audiences will be more likely to trust and engage with news if it comes from a reputable news outlet
than if it comes from a fake news website.
The authors write that “if people do not know a [news outlet] source, they approach its information similarly to how they would a [news outlet] source
they know and trust.”
The authors note that the conditions under which they conducted the study were somewhat different from those that participants would likely encounter in real life.
Researchers asked a nationally representative sample of 1,489 adults to read and answer questions about a simulated Facebook post that focused on a news article,
which appeared to have been shared by one of eight public figures.
In real life, these adults might have responded differently had they spotted such a post on their personal Facebook feeds, the authors explain.
Still, the findings provide new insights on how people interpret and engage with news.
“For news organizations who often rely on the strength of their brands to maintain trust in their audience, this study suggests that how people perceive their reporting
on social media may have little to do with that brand,” the authors write.
“A greater presence or role for individual journalists on social networks may help them boost trust in the content they create and share.”
This paper looks at changes in the volume of misinformation circulating on social media.
The gist: Since 2016, interactions with false content on Facebook have dropped dramatically but have risen on Twitter.
Still, lots of people continue to click on, comment on, like and share misinformation.
The researchers looked at how often the public interacted with stories from 569 fake news websites that appeared on Facebook and Twitter between
January 2015 and July 2018.
They found that Facebook engagements fell from about 160 million a month in late 2016 to about 60 million a month in mid-2018.
On Twitter, material from fake news sites was shared about 4 million times a month in late 2016 and grew to about 5 million shares a month in mid-2018.
The authors write that the evidence is “consistent with the view that the overall magnitude of the misinformation problem may have declined,
possibly due to changes to the Facebook platform following the 2016 election.”
This study looks at the cognitive mechanisms behind belief in fake news by investigating whether fake news has gained traction because of political partisanship
or because some people lack strong reasoning skills.
A key finding: Adults who performed better on a cognitive test were better able to detect fake news, regardless of their political affiliation or education levels and whether
the headlines they read were pro-Democrat, pro-Republican or politically neutral.
Across two studies conducted with 3,446 participants, the evidence suggests that “susceptibility to fake news is driven more by lazy thinking than
it is by partisan bias per se,” the authors write.
The authors also discovered that study participants who supported Trump had a weaker capacity for differentiating between real and fake news than did those
who supported 2016 presidential candidate Hillary Clinton.
The authors write that they are not sure why that is, but it might explain why fake news that benefited Republicans or harmed Democrats
seemed more common before the 2016 national election.
Even as the number of fact-checking outlets continues to grow globally, individual studies of their impact on misinformation have provided contradictory results.
To better understand whether fact-checking is an effective means of correcting political misinformation, scholars from three universities teamed up to synthesize
the findings of 30 studies published or released between 2013 and 2018.
Their analysis reveals that the success of fact-checking efforts varies according to a number of factors.
The resulting paper offers numerous insights on when and how fact-checking succeeds or fails.
Some of the big takeaways:
- Fact-checking messages that feature graphical elements such as so-called “truth scales” tended to be less effective in correcting misinformation than those that did not. The authors point out that “the inclusion of graphical elements appears to backfire and attenuate correction of misinformation.”
- Fact-checkers were more effective when they tried to correct an entire statement rather than parts of one. Also, according to the analysis, “fact-checking effects were significantly weaker for campaign-related statements.”
- Fact-checking that refutes ideas that contradict someone’s personal ideology was more effective than fact-checking aimed at debunking ideas that match someone’s personal ideology.
- Simple messages were more effective. “As a whole, lexical complexity appears to detract from fact-checking efforts,” the authors explain.
By Denise-Marie Ordway
January 13, 2020
Interested in research on fake news and digital media from previous years? Please check out our research roundups from 2018 and 2017.
This image was obtained from the Flickr account of Alan Levine and is being used under a Creative Commons license.