Computer model seeks to explain the spread of misinformation and suggest countermeasures

A graphical illustration of one time step of the POD model. In the left panel, (A) depicts the initial setup of a small network with institutional agent i1 with subscribers s1, s2, s3. All agents in the network are labeled with their belief strength. The right panel, (B) depicts one time step t = 0 of agent i1 sending messages M1(t = 0) = (m0, m1). (i) shows the initial sending of m0 = 4 to subscribers, and (ii) shows s1 and s3 believing the message and propagating it to their neighbors. (iii) and (iv) show the same for m1 = 3, but only s3 believes m1. Credit: DOI: 10.1371/journal.pone.0261811

It starts with a superspreader, and winds its way through a network of interactions, eventually leaving no one untouched. Those who have been exposed previously may experience little effect when exposed to a different variant.

No, it’s not a virus. It’s the contagious spread of misinformation and disinformation— misinformation that’s fully intended to deceive.

Now Tufts University researchers have come up with a computer model that remarkably mirrors the way misinformation spreads in real life. The work might provide insight on how to protect people from the current contagion of misinformation that threatens public health and the health of democracy, the researchers say.

“Our society has been grappling with widespread beliefs in conspiracies, increasing political polarization, and distrust in scientific findings,” said Nicholas Rabb, a Ph.D. computer science student at Tufts School of Engineering and lead author of the study, which came out January 7 in the journal Public Library of Science ONE. “This model could help us get a handle on how misinformation and conspiracy theories are spread, to help come up with strategies to counter them.”

Scientists who study the dissemination of information often take a page from epidemiologists, modeling the spread of false beliefs on how a disease spreads through a social network. Most of those models, however, treat the people in the networks as all equally taking in any new belief passed on to them by contacts.

The Tufts researchers instead based their model on the notion that our pre-existing beliefs can strongly influence whether we accept new information. Many people reject factual information supported by evidence if it takes them too far from what they already believe. Health-care workers have commented on the strength of this effect, observing that some patients dying from COVID cling to the belief that COVID does not exist.

To account for this in their model, the researchers assigned a “belief” to each individual in the artificial social network. To do this, the researchers represented beliefs of the individuals in the computer model by a number from 0 to 6, with 0 representing strong disbelief and 6 representing strong belief. The numbers could represent the spectrum of beliefs on any issue.

For example, one might think of the number 0 representing the strong disbelief that COVID vaccines help and are safe, while the number 6 might be the strong belief that COVID vaccines are in fact safe and effective.

The model then creates an extensive network of virtual individuals, as well as virtual institutional sources that originate much of the information that cascades through the network. In real life those could be news media, churches, governments, and social media influencers—basically the super-spreaders of information.

The model starts with an institutional source injecting the information into the network. If an individual receives information that is close to their beliefs—for example, a 5 compared to their current 6—they have a higher probability of updating that belief to a 5. If the incoming information differs greatly from their current beliefs—say a 2 compared to a 6—they will likely reject it completely and hold on to their 6 level belief.

Other factors, such as the proportion of their contacts that send them the information (basically, peer pressure) or the level of trust in the source, can influence how individuals update their beliefs. A population-wide network model of these interactions then provides an active view of the propagation and staying power of misinformation.

Future improvements to the model will take into account new knowledge from both network science and psychology, as well as a comparison of the results from the model with real world opinion surveys and network structures over time.

While the current model suggests that beliefs can change only incrementally, other scenarios could be modeled that cause a larger shift in beliefs—for example, a jump from 3 to 6 that could occur when a dramatic event happens to an influencer and they plead with their followers to change their minds.

Over time, the computer model can become more complex to accurately reflect what is happening on the ground, say the researchers, who in addition to Rabb include his faculty advisor Lenore Cowen, a professor of computer science; computer scientist Matthias Scheutz; and J.P deRuiter, a professor of both psychology and computer science.

“It’s becoming all too clear that simply broadcasting factual information may not be enough to make an impact on public mindset, particularly among those who are locked into a belief system that is not fact-based.” said Cowen. “Our initial effort to incorporate that insight into our models of the mechanics of misinformation spread in society may teach us how to bring the public conversation back to facts and evidence.”


Social media use increases belief in COVID-19 misinformation


More information:
Nicholas Rabb et al, Cognitive cascades: How to model (and potentially counter) the spread of fake news, PLOS ONE (2022). DOI: 10.1371/journal.pone.0261811

Provided by
Tufts University


Citation:
Computer model seeks to explain the spread of misinformation and suggest countermeasures (2022, January 11)
retrieved 11 January 2022
from https://techxplore.com/news/2022-01-misinformation-countermeasures.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.

For all the latest Technology News Click Here 

 For the latest news and updates, follow us on Google News

Read original article here

Denial of responsibility! TechNewsBoy.com is an automatic aggregator around the global media. All the content are available free on Internet. We have just arranged it in one platform for educational purpose only. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials on our website, please contact us by email – [email protected]. The content will be deleted within 24 hours.