Close Menu
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Trending

Govt calls ’35-day LPG refill rule’ misinformation, urges no panic booking

March 25, 2026

Australia has a major climate disinformation problem – and fixing it will take a huge effort

March 25, 2026

LPG Refill Booking Timelines Clarification Explained

March 25, 2026
Facebook X (Twitter) Instagram
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Subscribe
Web StatWeb Stat
Home»AI Fake News
AI Fake News

New AI could stop fake news in Urdu

News RoomBy News RoomMarch 18, 2026Updated:March 22, 20266 Mins Read
Facebook Twitter Pinterest WhatsApp Telegram Email LinkedIn Tumblr

Unmasking Deception: A New Dawn for Urdu Speakers in the Fight Against Fake News

In a world increasingly awash with information, both true and false, the ability to discern fact from fiction has become a crucial skill. For speakers of Urdu, the tenth most widely spoken language globally, this challenge has historically been exacerbated by a lack of robust tools to combat misinformation. However, a groundbreaking development offers a beacon of hope: a deep learning model, meticulously trained on over 14,000 Pakistani news articles, boasts an astounding 96% accuracy in identifying fake news. This isn’t just a technical achievement; it’s a profound step towards safeguarding public health, preserving democratic processes, and restoring trust in institutions for over 170 million Urdu speakers worldwide. It represents a human endeavor to empower individuals with the truth, providing them with a sophisticated weapon against the insidious spread of viral falsehoods that can wreak havoc on societies.

The human cost of misinformation is immeasurable. Imagine a public health crisis where false remedies spread like wildfire, undermining legitimate medical advice and endangering lives. Picture an election swayed by expertly crafted lies, eroding the very foundation of democracy. Envision public trust in essential services like the police and government crumbling under a barrage of fabricated narratives. These aren’t hypothetical scenarios; they are grim realities played out across countless communities globally. For Urdu-speaking populations, particularly those in Pakistan and the diaspora, the vulnerability to such digital contamination has been particularly acute. Previous attempts to create an AI system for Urdu fact-checking have always fallen short, leaving a critical gap in the information landscape. This new system, however, transcends those limitations. It’s not just about flagging outright fabrications; it can also pinpoint misleading content and even stories that are only partially true, offering a nuanced approach to an increasingly complex problem. This signifies a compassionate effort to protect individuals from the insidious nature of half-truths, which can be even more damaging than outright lies.

Dr. Muhammad Zeeshan Babar, an instrumental figure from Heriot-Watt University’s School of Engineering and Physical Sciences, eloquently articulates the historical disparity in AI development. “Most automated fake news detection systems are trained on English language datasets,” he explains, highlighting the inherent bias in the existing technological landscape. This oversight is particularly glaring considering Urdu’s global prominence. “Urdu is the 10th most spoken language in the world and the national language of Pakistan,” Dr. Babar emphasizes, yet it has been consistently marginalized in the realm of AI research. The core issue, as he points out, lies in the scarcity of comprehensive datasets. Urdu has long been considered a “low-resource language” in the AI community, meaning there simply hasn’t been enough structured data available to effectively train sophisticated AI systems. This lack of resources translates into a genuine human disadvantage, leaving Urdu speakers without the same digital safeguards enjoyed by those who speak more “resourced” languages. The team’s work is a direct challenge to this imbalance, a determined effort to level the playing field and ensure equitable access to reliable information for all.

The journey to building this robust system began with a critical examination of existing Urdu datasets. What Dr. Babar and his colleagues discovered was a significant and concerning void. “We found real weaknesses in the available Urdu datasets,” he reveals. A striking omission was the lack of news articles pertaining to “politics, religion, and other societal issues.” This absence wasn’t accidental; it largely stemmed from the sensitive and often controversial nature of these topics. However, as Dr. Babar rightly points out, this constituted a “critical gap” because “misinformation in Pakistani news, which is read by the diaspora around the world, touches on all of those subjects.” This realization underscores the deep understanding the team had of the specific challenges faced by Urdu speakers. They understood that to truly address misinformation, they couldn’t shy away from the very topics where it often thrives. Their commitment to tackling these sensitive areas demonstrates a profound sense of responsibility, recognizing the critical role these subjects play in the lives and opinions of millions.

Driven by this understanding, the team embarked on the monumental task of creating their own comprehensive dataset, aptly named the “Urdu Fake News Detection dataset.” This wasn’t merely a technical exercise; it was a tireless effort fueled by a human desire to build a more resilient information ecosystem. Between 2017 and 2023, they meticulously collected 14,178 Urdu language news articles, spanning 15 diverse subject areas including the previously overlooked politics and religion, as well as health, business, education, sports, and technology. Each article was then carefully labeled – 8,283 as real and 5,895 as fake. This painstaking process, involving countless hours of human review and analysis, taught the AI system to discern the subtle nuances that separate authentic reporting from fabricated content. The model learned to identify patterns in vocabulary, phrasing, sentiment, and linguistic structures that, to the human eye, might escape immediate detection. It’s a testament to the power of human-curated data, empowering a machine to mimic and even surpass human capabilities in this critical domain.

The decision to make this invaluable dataset open access, as articulated by Dr. Waseem Abbasi, Head of Computer Science at the University of Lahore, speaks volumes about the team’s collaborative spirit and their commitment to the broader scientific community. “We’ve made the dataset open access so that we can continually improve its performance,” he states, acknowledging that even at 96% accuracy, there’s always room for refinement. This open-source approach fosters further research and development, inviting other experts to build upon their foundational work. Dr. Abbasi also offers a grounded perspective on the project’s limitations. While 96% accuracy is remarkable, he recognizes that the remaining “significant margin of error” could have serious implications for content moderation, advertising, and even legal enforcement. He further cautions that “algorithms trained on past data may struggle with emerging narratives; they could misclassify satire or political dissent.” These are important considerations, demonstrating a responsible and ethical approach to AI development. Yet, despite these caveats, the potential impact of this system is undeniable. For “millions of Urdu news consumers trying to navigate a polluted information ecosystem,” this technological breakthrough offers a ray of hope, a powerful tool to empower them with verified information. The team’s foresight in extending their research to other low-resource languages further solidifies their commitment to a more equitable and truth-filled digital world, a truly humane ambition.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
News Room
  • Website

Keep Reading

AI: Fake influencers are sexualising black women – BBC

Two brothers charged with spying for Iran, using AI to fake military intel

The last 48 hours proves reality is broken

RSS plaint over fake AI-generated video | Guwahati News

How AI fakes erode trust and how to stay safe

Delve accused of misleading customers with ‘fake compliance’

Editors Picks

Australia has a major climate disinformation problem – and fixing it will take a huge effort

March 25, 2026

LPG Refill Booking Timelines Clarification Explained

March 25, 2026

Read first, then react: Academic expert on combating disinformation

March 25, 2026

Government Clarifies No Change in LPG Booking Rules, Debunks False Reports

March 25, 2026

IEBC calls for high-tech war room to tackle misinformation in upcoming polls

March 25, 2026

Latest Articles

Ukraine accuses Russia of weaponizing migration, warns of European destabilization | Ukraine news

March 25, 2026

Dad brutally murders wife in front of children after false affair claim

March 25, 2026

Rae blames ‘misinformation’ for complaints about integrated assessment tool

March 25, 2026

Subscribe to News

Get the latest news and updates directly to your inbox.

Facebook X (Twitter) Pinterest TikTok Instagram
Copyright © 2026 Web Stat. All Rights Reserved.
  • Privacy Policy
  • Terms
  • Contact

Type above and press Enter to search. Press Esc to cancel.