ΑΙhub.org
 

Algorithms can be useful in detecting fake news, stopping its spread and countering misinformation


by
23 June 2023



share this:
stack of newspapers

Laks V.S. Lakshmanan, University of British Columbia

Fake news is a complex problem and can span text, images and video.

For written articles in particular, there are several ways of generating fake news. A fake news article could be produced by selectively editing facts, including people’s names, dates or statistics. An article could also be completely fabricated with made-up events or people.

Fake news articles can also be machine-generated as advances in artificial intelligence make it particularly easy to generate misinformation.

Damaging effects

Questions like: “Was there voter fraud during the 2020 U.S. elections?” or “Is climate change a hoax?” can be fact-checked by analyzing available data. These questions can be answered with true or false, but there is potential for misinformation surrounding questions like these.

Misinformation and disinformation — or fake news — can have damaging effects on a large number of people in a short time. Although the notion of fake news has existed well before technological advances, social media have exacerbated the problem.

A 2018 Twitter study showed that false news stories were more commonly retweeted by humans than bots, and 70 per cent more likely to be retweeted than true stories. The same study found that it took true stories approximately six times longer to reach a group of 1,500 people and, while true stories rarely reached more than 1,000 people, popular false news could spread up to 100,000.

The 2020 U.S. presidential election, COVID-19 vaccines and climate change have all been the subject of misinformation campaigns with grave consequences. It is estimated that misinformation surrounding COVID-19 costs between US$50-300 million daily. The cost of political misinformation could be civil disorder, violence or even erosion of public trust in democratic institutions.

Detecting misinformation

Detecting misinformation can be done by a combination of algorithms, machine-learning models and humans. An important question is who is responsible for controlling, if not stopping, the spread of misinformation once it’s detected. Only social media companies are really in the position to exercise control over the spread of information through their networks.

A particularly simple but effective means of generating misinformation is to selectively edit news articles. For example, consider “Ukrainian director and playwright arrested and accused of ‘justifying terrorism.’” This was achieved by replacing “Russian” with “Ukrainian” in the original sentence in a real news article.

A multi-faceted approach is needed to detect misinformation online in order to control its growth and spread.

Communications in social media can be modelled as networks, with the users forming points in the network model and the communications forming links between them; a retweet or like of a post reflects a connection between two points. In this network model, spreaders of misinformation tend to form much more densely connected core-periphery structures than users spreading truth.

My research group has developed efficient algorithms for detecting dense structures from communication networks. This information can be analyzed further for detecting instances of misinformation campaigns.

Since these algorithms rely on communication structure alone, content analysis conducted by algorithms and humans is needed to confirm instances of misinformation.

Detecting manipulated articles takes careful analysis. Our research used a neural network-based approach that combines textual information with an external knowledge base to detect such tampering.

Stopping the spread

Detecting misinformation is just half the battle — decisive action is required to stop its spread. Strategies for combating the spread of misinformation in social networks include both intervention by internet platforms and launching counter-campaigns to neutralize fake news campaigns.

Intervention can take hard forms, like suspending a user’s account, or softer measures like labelling a post as suspicious.

Algorithms and AI-powered networks are not 100 per cent reliable. There is a cost to intervening on a true item by mistake as well as not intervening on a fake item.

To that end, we designed a smart intervention policy that automatically decides whether to intervene on an item based on its predicted truthiness and predicted popularity.

Countering fake news

Launching counter-campaigns to minimize if not neutralize the effects of misinformation campaigns needs to factor in the major differences between truth and fake news in terms of how quickly and extensively each of them spreads.

Besides these differences, reactions to stories can vary depending on the user, topic and length of the post. Our approach takes all these factors into account and devises an efficient counter campaign strategy that effectively mitigates the propagation of misinformation.

Recent advances in generative AI, particularly those powered by large language models such ChatGPT, make it easier than ever to create articles at great speed and significant volume, raising the challenge of detecting misinformation and countering its spread at scale and in real time. Our current research continues to address this ongoing challenge which has enormous societal impact.The Conversation

Laks V.S. Lakshmanan, Professor of Computer Science, University of British Columbia

This article is republished from The Conversation under a Creative Commons license. Read the original article.




The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.
The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.




            AIhub is supported by:


Related posts :



monthly digest

AIhub monthly digest: May 2025 – materials design, object state classification, and real-time monitoring for healthcare data

  30 May 2025
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

Congratulations to the #AAMAS2025 best paper, best demo, and distinguished dissertation award winners

  29 May 2025
Find out who won the awards presented at the International Conference on Autonomous Agents and Multiagent Systems last week.

The Good Robot podcast: Transhumanist fantasies with Alexander Thomas

  28 May 2025
In this episode, Eleanor talks to Alexander Thomas, a filmmaker and academic, about the transhumanist narrative.

Congratulations to the #ICRA2025 best paper award winners

  27 May 2025
The winners and finalists in the different categories have been announced.

#ICRA2025 social media round-up

  23 May 2025
Find out what the participants got up to at the International Conference on Robotics & Automation.

Interview with Gillian Hadfield: Normative infrastructure for AI alignment

  22 May 2025
Kumar Kshitij Patel spoke to Gillian Hadfield about her interdisciplinary research, career trajectory, path into AI alignment, law, and general thoughts on AI systems.

PitcherNet helps researchers throw strikes with AI analysis

  21 May 2025
Baltimore Orioles tasks Waterloo Engineering researchers to develop AI tech that can monitor pitchers using low-resolution video captured by smartphones

Interview with Filippos Gouidis: Object state classification

  20 May 2025
Read the latest interview in our series featuring the AAAI/SIGAI Doctoral Consortium participants.



 

AIhub is supported by:






©2025.05 - Association for the Understanding of Artificial Intelligence


 












©2025.05 - Association for the Understanding of Artificial Intelligence