Close Menu
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Trending

AI-Generated Red Deer Weather Incident Hoax Goes Viral – A New Age of Fake News?

July 5, 2025

UN climate expert urges criminalization of fossil fuel disinformation to protect basic human rights

July 5, 2025

Can AI chatbots easily be misused to spread credible health misinformation?

July 5, 2025
Facebook X (Twitter) Instagram
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Subscribe
Web StatWeb Stat
Home»AI Fake News
AI Fake News

Apple Discontinues AI System Responsible for Fabricating and Disseminating False News.

News RoomBy News RoomJanuary 17, 2025Updated:January 28, 20253 Mins Read
Facebook Twitter Pinterest WhatsApp Telegram Email LinkedIn Tumblr

Apple Halts ‘Apple Intelligence’ Feature After String of Factual Errors in News Summaries

Cupertino, CA – Apple has announced a temporary suspension of its "Apple Intelligence" feature, following weeks of widespread criticism over its inaccurate and misleading summarizations of breaking news alerts. The feature, designed to provide concise summaries of news articles, has instead become notorious for misrepresenting facts, fabricating information, and even attributing actions to individuals falsely. The upcoming iOS 18.3 update will disable these summaries for news and entertainment apps, marking a rare admission of failure for the tech giant.

The decision comes after mounting evidence of the feature’s shortcomings, culminating in a recent incident involving a grossly inaccurate summary of a Washington Post article about Donald Trump’s defense secretary nominee, Pete Hegseth. The AI-generated summary, as reported by Geoffrey Fowler of the Washington Post, “got every fact wrong,” highlighting the severity of the issue and prompting calls for Apple to disable the feature until its accuracy could be significantly improved.

Apple’s "Apple Intelligence" feature, leveraging large language model (LLM) technology, aimed to provide users with quick and digestible summaries of news articles. However, like other LLM-based products from companies like Google and OpenAI, the feature has been plagued by "hallucinations," a phenomenon where the AI generates false or misleading information, often presented with the same confidence as factual data. This issue has raised concerns about the reliability and trustworthiness of AI-generated content, particularly in the context of news dissemination.

The problems with Apple’s summarization tool extend beyond simple inaccuracies. In December, the BBC lodged a formal complaint with Apple after the feature consistently misrepresented their news reporting. One particularly egregious instance involved the feature falsely claiming that Luigi Mangione, a suspect in the killing of UnitedHealthcare CEO Brian Thompson, had shot himself. This incident underscores the potential for significant harm caused by the dissemination of false information through AI-generated summaries.

Initially, Apple seemed reluctant to disable the feature entirely, suggesting instead that a software update would provide more clarity regarding the source of the summaries. However, the escalating criticism and the growing number of documented inaccuracies appear to have forced the company to reconsider its approach. The decision to temporarily halt the feature signifies a recognition of the seriousness of the problem and a commitment to addressing the underlying issues before reintroducing the functionality.

While Apple has promised to reinstate the feature after implementing improvements, the timeline remains unclear. The company faces the challenge of mitigating the "hallucination" problem, a persistent issue that continues to plague LLM-based technologies. The "Apple Intelligence" debacle serves as a cautionary tale about the potential pitfalls of deploying AI-powered tools without adequate safeguards and rigorous testing, especially in sensitive domains like news reporting. It highlights the need for greater transparency and accountability in the development and deployment of AI systems, as well as the importance of critical evaluation of AI-generated content. The incident also raises broader questions about the readiness of LLM technology for widespread application in information dissemination and the potential consequences of relying on such systems for factual information.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
News Room
  • Website

Keep Reading

Viral band success spawns AI claims and hoaxes

How to spot AI-generated newscasts – DW – 07/02/2025

Fake news in the age of AI

AI chatbots could spread ‘fake news’ with serious health consequences

Fake, AI-generated videos about the Diddy trial are raking in millions of views on YouTube | Artificial intelligence (AI)

Meta Denies $100M Signing Bonus Claims as OpenAI Researcher Calls It ‘Fake News’

Editors Picks

UN climate expert urges criminalization of fossil fuel disinformation to protect basic human rights

July 5, 2025

Can AI chatbots easily be misused to spread credible health misinformation?

July 5, 2025

False Reports About Mosque Conversion of Ani Cathedral

July 5, 2025

‘Blatant misinformation’: Social Security Administration email praising Trump’s tax bill blasted as a ‘lie’ | US social security

July 5, 2025

Udupi: Man Arrested for Allegedly Raping Woman Under False Promise of Marriage

July 5, 2025

Latest Articles

Misinformation On Operation Sindoor, 2025 Bihar Elections & More

July 5, 2025

Young mother-of-two shares one piece of misinformation everyone needs to know about killer disease – after ‘piles’ turned out to be stage 3 bowel cancer

July 5, 2025

X brings AI into Community Notes to fight misinformation at scale humans can’t match

July 5, 2025

Subscribe to News

Get the latest news and updates directly to your inbox.

Facebook X (Twitter) Pinterest TikTok Instagram
Copyright © 2025 Web Stat. All Rights Reserved.
  • Privacy Policy
  • Terms
  • Contact

Type above and press Enter to search. Press Esc to cancel.