Close Menu
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Trending

Medical misinformation and sensationalism in the age of social media

July 31, 2025

Another disinformation: Khortytsia OSGT spokesperson on Russia’s statement regarding the capture of Chasiv Yar

July 31, 2025

Online abuse against migrants doubled after Southport attack

July 31, 2025
Facebook X (Twitter) Instagram
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Subscribe
Web StatWeb Stat
Home»Disinformation
Disinformation

Meta Study Reveals GenAI Accounts for Under 1% of Election-Related Misinformation in 2024 – Firstpost

News RoomBy News RoomDecember 5, 20243 Mins Read
Facebook Twitter Pinterest WhatsApp Telegram Email LinkedIn Tumblr

In a recent analysis conducted by Meta, the role of generative AI in spreading misinformation during the major elections of 2024 was examined across 40 countries, including significant regions like India, the US, and the EU. Contrary to earlier concerns that AI might exacerbate disinformation campaigns, the study revealed that AI-generated content accounted for less than one percent of flagged posts on Meta’s platforms. This finding suggests that current safeguards implemented by the company have been effective in mitigating the potential misuse of AI technology, thereby reducing the overall risk associated with misinformation during critical electoral periods.

Nick Clegg, Meta’s president of global affairs, addressed the findings, indicating that although there were some instances of malicious AI usage, the overall volume was minimal. He emphasized the sufficiency of the company’s existing policies and tools in managing the risks linked to AI content across various platforms such as Facebook, Instagram, WhatsApp, and Threads. The findings are particularly reassuring, as they highlight the effectiveness of preventative measures already in place, designed to combat disinformation and maintain the integrity of electoral processes in multiple regions.

In addition to addressing AI-related misinformation, Meta reported significant progress in countering election interference more broadly. The company successfully dismantled over 20 covert influence campaigns, classified as Coordinated Inauthentic Behavior (CIB) networks. While these operations did utilize generative AI for some content generation, Meta concluded that the technology did not notably amplify the scale or effectiveness of these campaigns, demonstrating the company’s proactive stance in preventing such disruptive activities.

Meta’s monitoring also extended to user activity, as nearly 600,000 attempts to create deepfake images of political figures—using their AI image generator known as Imagine—were blocked. This includes fabricated images involving prominent leaders such as President-elect Trump and President Biden. These numbers underscore a significant demand for stricter regulation of AI tools during critical events, affirming the need for ongoing vigilance against attempts to manipulate public opinion through deceptive imagery.

Reflecting on the experiences of content moderation during the COVID-19 pandemic, Clegg acknowledged that Meta may have initially adopted an excessively strict approach, resulting in the removal of many harmless posts. He noted that the uncertainty of the period contributed to the company’s high error rate in moderation, which unfortunately impacted user expression. This recognition underscores the challenges that Meta faces in balancing effective content moderation while safeguarding the free expression that it aims to promote.

The overall conclusions of this study indicate that the anticipated threat of AI-generated disinformation, particularly in the context of elections, may have been overstated for the moment. Through robust monitoring and strategic policy enforcement, Meta has managed to maintain a relatively controlled environment regarding AI misuse. However, the company recognizes the ongoing challenges posed by increasingly sophisticated AI tools, underscoring the importance of refining their approaches to uphold user trust and platform integrity in the future.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
News Room
  • Website

Keep Reading

Another disinformation: Khortytsia OSGT spokesperson on Russia’s statement regarding the capture of Chasiv Yar

Fighting Fiction with Force: Tools for the Post-truth Era

Conference in Estonia will explore role of churches in countering disinformation

End Wokeness: Riot-fuelling X account still targeting Muslims

Reckless gamble: Meta’s decision on disinformation endangers its 4 billion users

Adani claims its export program helps contribute to sustainable energy – but experts say that’s ‘wilful disinformation’ | Australia news

Editors Picks

Another disinformation: Khortytsia OSGT spokesperson on Russia’s statement regarding the capture of Chasiv Yar

July 31, 2025

Online abuse against migrants doubled after Southport attack

July 31, 2025

Debunked: How misinformation spread within minutes of a garda being stabbed on Capel Street – The Journal

July 31, 2025

Fighting Fiction with Force: Tools for the Post-truth Era

July 31, 2025

Misinformation, Science and Media | Oxford Martin School

July 31, 2025

Latest Articles

Kerala Rapper Vedan Booked For Rape After Doctor Alleges False Promise Of Marriage | People News

July 31, 2025

Washington Post spreads misinformation on Operation Sindoor, quietly deletes false claims targeting Indian media

July 31, 2025

GT Investigates: Online hysteria over giant pandas fuels misinformation, exposing emotional manipulation, profit-driven agendas behind conservation claims

July 31, 2025

Subscribe to News

Get the latest news and updates directly to your inbox.

Facebook X (Twitter) Pinterest TikTok Instagram
Copyright © 2025 Web Stat. All Rights Reserved.
  • Privacy Policy
  • Terms
  • Contact

Type above and press Enter to search. Press Esc to cancel.