Close Menu
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Trending

Ayob Khan reiterates no police links to Ketereh murder suspect’s family, warns against misinformation

May 10, 2026

When lifelines are labeled as conspiracy: The toll on small and non-profit newsrooms

May 10, 2026

Disinformation about heat pumps on the Ministry of Climate and Environment’s radar – Ministry of Climate and Environment

May 10, 2026
Facebook X (Twitter) Instagram
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Subscribe
Web StatWeb Stat
Home»AI Fake News
AI Fake News

New AI could stop fake news in Urdu

News RoomBy News RoomMarch 18, 2026Updated:March 22, 20266 Mins Read
Facebook Twitter Pinterest WhatsApp Telegram Email LinkedIn Tumblr

Unmasking Deception: A New Dawn for Urdu Speakers in the Fight Against Fake News

In a world increasingly awash with information, both true and false, the ability to discern fact from fiction has become a crucial skill. For speakers of Urdu, the tenth most widely spoken language globally, this challenge has historically been exacerbated by a lack of robust tools to combat misinformation. However, a groundbreaking development offers a beacon of hope: a deep learning model, meticulously trained on over 14,000 Pakistani news articles, boasts an astounding 96% accuracy in identifying fake news. This isn’t just a technical achievement; it’s a profound step towards safeguarding public health, preserving democratic processes, and restoring trust in institutions for over 170 million Urdu speakers worldwide. It represents a human endeavor to empower individuals with the truth, providing them with a sophisticated weapon against the insidious spread of viral falsehoods that can wreak havoc on societies.

The human cost of misinformation is immeasurable. Imagine a public health crisis where false remedies spread like wildfire, undermining legitimate medical advice and endangering lives. Picture an election swayed by expertly crafted lies, eroding the very foundation of democracy. Envision public trust in essential services like the police and government crumbling under a barrage of fabricated narratives. These aren’t hypothetical scenarios; they are grim realities played out across countless communities globally. For Urdu-speaking populations, particularly those in Pakistan and the diaspora, the vulnerability to such digital contamination has been particularly acute. Previous attempts to create an AI system for Urdu fact-checking have always fallen short, leaving a critical gap in the information landscape. This new system, however, transcends those limitations. It’s not just about flagging outright fabrications; it can also pinpoint misleading content and even stories that are only partially true, offering a nuanced approach to an increasingly complex problem. This signifies a compassionate effort to protect individuals from the insidious nature of half-truths, which can be even more damaging than outright lies.

Dr. Muhammad Zeeshan Babar, an instrumental figure from Heriot-Watt University’s School of Engineering and Physical Sciences, eloquently articulates the historical disparity in AI development. “Most automated fake news detection systems are trained on English language datasets,” he explains, highlighting the inherent bias in the existing technological landscape. This oversight is particularly glaring considering Urdu’s global prominence. “Urdu is the 10th most spoken language in the world and the national language of Pakistan,” Dr. Babar emphasizes, yet it has been consistently marginalized in the realm of AI research. The core issue, as he points out, lies in the scarcity of comprehensive datasets. Urdu has long been considered a “low-resource language” in the AI community, meaning there simply hasn’t been enough structured data available to effectively train sophisticated AI systems. This lack of resources translates into a genuine human disadvantage, leaving Urdu speakers without the same digital safeguards enjoyed by those who speak more “resourced” languages. The team’s work is a direct challenge to this imbalance, a determined effort to level the playing field and ensure equitable access to reliable information for all.

The journey to building this robust system began with a critical examination of existing Urdu datasets. What Dr. Babar and his colleagues discovered was a significant and concerning void. “We found real weaknesses in the available Urdu datasets,” he reveals. A striking omission was the lack of news articles pertaining to “politics, religion, and other societal issues.” This absence wasn’t accidental; it largely stemmed from the sensitive and often controversial nature of these topics. However, as Dr. Babar rightly points out, this constituted a “critical gap” because “misinformation in Pakistani news, which is read by the diaspora around the world, touches on all of those subjects.” This realization underscores the deep understanding the team had of the specific challenges faced by Urdu speakers. They understood that to truly address misinformation, they couldn’t shy away from the very topics where it often thrives. Their commitment to tackling these sensitive areas demonstrates a profound sense of responsibility, recognizing the critical role these subjects play in the lives and opinions of millions.

Driven by this understanding, the team embarked on the monumental task of creating their own comprehensive dataset, aptly named the “Urdu Fake News Detection dataset.” This wasn’t merely a technical exercise; it was a tireless effort fueled by a human desire to build a more resilient information ecosystem. Between 2017 and 2023, they meticulously collected 14,178 Urdu language news articles, spanning 15 diverse subject areas including the previously overlooked politics and religion, as well as health, business, education, sports, and technology. Each article was then carefully labeled – 8,283 as real and 5,895 as fake. This painstaking process, involving countless hours of human review and analysis, taught the AI system to discern the subtle nuances that separate authentic reporting from fabricated content. The model learned to identify patterns in vocabulary, phrasing, sentiment, and linguistic structures that, to the human eye, might escape immediate detection. It’s a testament to the power of human-curated data, empowering a machine to mimic and even surpass human capabilities in this critical domain.

The decision to make this invaluable dataset open access, as articulated by Dr. Waseem Abbasi, Head of Computer Science at the University of Lahore, speaks volumes about the team’s collaborative spirit and their commitment to the broader scientific community. “We’ve made the dataset open access so that we can continually improve its performance,” he states, acknowledging that even at 96% accuracy, there’s always room for refinement. This open-source approach fosters further research and development, inviting other experts to build upon their foundational work. Dr. Abbasi also offers a grounded perspective on the project’s limitations. While 96% accuracy is remarkable, he recognizes that the remaining “significant margin of error” could have serious implications for content moderation, advertising, and even legal enforcement. He further cautions that “algorithms trained on past data may struggle with emerging narratives; they could misclassify satire or political dissent.” These are important considerations, demonstrating a responsible and ethical approach to AI development. Yet, despite these caveats, the potential impact of this system is undeniable. For “millions of Urdu news consumers trying to navigate a polluted information ecosystem,” this technological breakthrough offers a ray of hope, a powerful tool to empower them with verified information. The team’s foresight in extending their research to other low-resource languages further solidifies their commitment to a more equitable and truth-filled digital world, a truly humane ambition.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
News Room
  • Website

Keep Reading

AI tools to help centre catch fake Ayushman claims | India News

AI Fakes the Founder and Keeps the Money

Hackers Using Fake Claude AI Installer Pages to Trick Users Into Running Malware on Their Systems

Fake Claude AI website delivers new ‘Beagle’ Windows malware

Italian PM Giorgia Meloni Denounces AI-Generated Deepfakes as a Threat, ETEnterpriseai

The AI fitness instructors selling unreal gains

Editors Picks

When lifelines are labeled as conspiracy: The toll on small and non-profit newsrooms

May 10, 2026

Disinformation about heat pumps on the Ministry of Climate and Environment’s radar – Ministry of Climate and Environment

May 10, 2026

Chief Justice Surya Kant denies viral casteist remarks and false quotes

May 10, 2026

Challenging disinformation is a duty we must not avoid

May 10, 2026

Recommendations of the Advisory Council for Resilience to International Disinformation to the Minister of Foreign Affairs on countering disinformation in the information environment – Poland in South Africa

May 10, 2026

Latest Articles

Ayob Khan: Claims linking suspect’s father to police are false

May 10, 2026

New German book exposes how EU outsources censorship to NGOs

May 10, 2026

Police arrested a man in his 30s who synthesized false broadcast subtitles with photos of the presid..

May 10, 2026

Subscribe to News

Get the latest news and updates directly to your inbox.

Facebook X (Twitter) Pinterest TikTok Instagram
Copyright © 2026 Web Stat. All Rights Reserved.
  • Privacy Policy
  • Terms
  • Contact

Type above and press Enter to search. Press Esc to cancel.