Close Menu
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Trending

Webinar | Knowing the facts: How communicators can identify and respond to vaccine misinformation – PAHO/WHO

June 7, 2025

Opinion: Donlin Gold deserves a fair hearing based on facts, not misinformation

June 7, 2025

BRS faults Congress for misinformation campaign on Kaleshwaram project

June 7, 2025
Facebook X (Twitter) Instagram
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Subscribe
Web StatWeb Stat
Home»Disinformation
Disinformation

Meta Study Reveals GenAI Accounts for Under 1% of Election-Related Misinformation in 2024 – Firstpost

News RoomBy News RoomDecember 5, 20243 Mins Read
Facebook Twitter Pinterest WhatsApp Telegram Email LinkedIn Tumblr

In a recent analysis conducted by Meta, the role of generative AI in spreading misinformation during the major elections of 2024 was examined across 40 countries, including significant regions like India, the US, and the EU. Contrary to earlier concerns that AI might exacerbate disinformation campaigns, the study revealed that AI-generated content accounted for less than one percent of flagged posts on Meta’s platforms. This finding suggests that current safeguards implemented by the company have been effective in mitigating the potential misuse of AI technology, thereby reducing the overall risk associated with misinformation during critical electoral periods.

Nick Clegg, Meta’s president of global affairs, addressed the findings, indicating that although there were some instances of malicious AI usage, the overall volume was minimal. He emphasized the sufficiency of the company’s existing policies and tools in managing the risks linked to AI content across various platforms such as Facebook, Instagram, WhatsApp, and Threads. The findings are particularly reassuring, as they highlight the effectiveness of preventative measures already in place, designed to combat disinformation and maintain the integrity of electoral processes in multiple regions.

In addition to addressing AI-related misinformation, Meta reported significant progress in countering election interference more broadly. The company successfully dismantled over 20 covert influence campaigns, classified as Coordinated Inauthentic Behavior (CIB) networks. While these operations did utilize generative AI for some content generation, Meta concluded that the technology did not notably amplify the scale or effectiveness of these campaigns, demonstrating the company’s proactive stance in preventing such disruptive activities.

Meta’s monitoring also extended to user activity, as nearly 600,000 attempts to create deepfake images of political figures—using their AI image generator known as Imagine—were blocked. This includes fabricated images involving prominent leaders such as President-elect Trump and President Biden. These numbers underscore a significant demand for stricter regulation of AI tools during critical events, affirming the need for ongoing vigilance against attempts to manipulate public opinion through deceptive imagery.

Reflecting on the experiences of content moderation during the COVID-19 pandemic, Clegg acknowledged that Meta may have initially adopted an excessively strict approach, resulting in the removal of many harmless posts. He noted that the uncertainty of the period contributed to the company’s high error rate in moderation, which unfortunately impacted user expression. This recognition underscores the challenges that Meta faces in balancing effective content moderation while safeguarding the free expression that it aims to promote.

The overall conclusions of this study indicate that the anticipated threat of AI-generated disinformation, particularly in the context of elections, may have been overstated for the moment. Through robust monitoring and strategic policy enforcement, Meta has managed to maintain a relatively controlled environment regarding AI misuse. However, the company recognizes the ongoing challenges posed by increasingly sophisticated AI tools, underscoring the importance of refining their approaches to uphold user trust and platform integrity in the future.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
News Room
  • Website

Keep Reading

Ukraine rejects claims of delaying exchange of soldiers’ bodies, calls out Russian disinformation

Disinformation & Democracy – Center for Informed Democracy & Social – cybersecurity (IDeaS)

The anatomy of a lie: Ways the public can predict and defend against Trump’s disinformation tactics

US needs to ‘stop spreading disinformation,’ correct ‘wrongful actions’

Pride is Here and So is the Misinformation – by Carly Winchell

The New Axis of Disinformation

Editors Picks

Opinion: Donlin Gold deserves a fair hearing based on facts, not misinformation

June 7, 2025

BRS faults Congress for misinformation campaign on Kaleshwaram project

June 7, 2025

The Truth About Sun Exposure: Doctor Sets the Record Straight amid Influencer Misinformation – People.com

June 7, 2025

BRS MLA Harish Rao defends Kaleshwaram Lift Irrigation Scheme, slams Congress for ‘misinformation campaign’ | Hyderabad News

June 7, 2025

Westfield Health Bulletin: Health and vaccine misinformation puts people at risk

June 7, 2025

Latest Articles

Ukraine rejects claims of delaying exchange of soldiers’ bodies, calls out Russian disinformation

June 7, 2025

Doctor Sets the Record Straight amid Influencer Misinformation

June 7, 2025

Misinformation On RCB’s IPL Win, Russia-Ukraine Conflict & More

June 7, 2025

Subscribe to News

Get the latest news and updates directly to your inbox.

Facebook X (Twitter) Pinterest TikTok Instagram
Copyright © 2025 Web Stat. All Rights Reserved.
  • Privacy Policy
  • Terms
  • Contact

Type above and press Enter to search. Press Esc to cancel.