Close Menu
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Trending

Information wars: Combatting climate misinformation and disinformation

March 28, 2026

Roads Minster Agbodza dismisses ‘disinformation’ on Big Push contracts, insists no Ghanaian contractor operates with one staff – 3News

March 28, 2026

New tool tackles dangerous diet and vaccine misinformation online

March 28, 2026
Facebook X (Twitter) Instagram
Web StatWeb Stat
  • Home
  • News
  • United Kingdom
  • Misinformation
  • Disinformation
  • AI Fake News
  • False News
  • Guides
Subscribe
Web StatWeb Stat
Home»Misinformation
Misinformation

New tool tackles dangerous diet and vaccine misinformation online

News RoomBy News RoomMarch 28, 20265 Mins Read
Facebook Twitter Pinterest WhatsApp Telegram Email LinkedIn Tumblr

It’s a tough world out there these days, especially with so much information flying around online. We’ve all seen those extreme headlines or friends sharing wacky diet tips, right? Well, it turns out that misleading health information isn’t just a nuisance – it’s a serious problem, according to big shots like the World Health Organization. We’re talking about everything from diets that tell you to eat only meat, to people taking supplements that promise miracles but can actually land them in the emergency room. In fact, studies have hinted that herbal and dietary supplements alone are tied to a shocking 20% of serious liver injuries caused by drugs and around 23,000 emergency visits in the US every year. That’s a lot of people getting hurt because of bad advice.

This isn’t just about outright lies; it’s often more insidious than that. Imagine someone sharing a post about a ‘superfood’ that’s been cherry-picked to sound amazing, but conveniently leaves out the potential dangers. That’s exactly the kind of nuanced misinformation that a new tool, developed by clever folks at University College London (UCL), aims to tackle. Unlike other tools that just say “true” or “false,” this one is designed to catch content that isn’t completely fabricated but is still cleverly framed to be misleading, especially for people who might be more vulnerable to that kind of influence. As Alex Ruani, the lead author and developer, puts it, when it comes to diet, misinformation often works by selectively framing things, hiding the potential health risks. This “harmful misleading content tends to fly under fact-checkers’ radars and escape meaningful oversight until high-profile cases make the headlines,” he explains.

Think about it: how many times have you seen something online that sounds plausible, but a little too good to be true? One stark example that researchers cite is a case from 2025 where a man developed cholesterol-induced skin lesions after adopting a carnivore diet – a trend that social media algorithms, particularly in “manosphere” communities, love to amplify. Another terrifying instance involved someone getting hospitalized simply because they followed bad AI-generated advice to replace common table salt (sodium chloride) with sodium bromide. Sodium bromide, for those who don’t know, has no dietary purpose and is pretty toxic if you take it regularly. And on an even more heartbreaking note, online misinformation has convinced some people to abandon life-saving cancer treatments for unproven dietary alternatives. These aren’t just abstract examples; these are real people facing real harm because of the echo chamber of bad advice online.

So, how does this new tool, aptly named the Diet-Nutrition Misinformation Risk Assessment Tool, or simply Diet-MisRAT, actually work? It dives into online content and tries to figure out how likely someone would be led astray by it. Then, based on a weighted misinformation risk score, it gives the material a color-coded ranking: green, amber, or red. For instance, if it spots a claim like “it is safer to give your child high-dose vitamin A than the MMR vaccine,” it immediately flags that as a critical risk. Why? Because it’s a false safety claim that can have incredibly dangerous consequences. It’s not just about what’s explicitly false, but about how information is framed to create a misleading sense of safety or benefit.

The hope is that this tool won’t just sit in a lab. Researchers envision it becoming a vital resource for policymakers who are trying to create regulations, for digital platforms like Facebook and TikTok to help them moderate content more effectively, and for other regulators to implement stronger safeguards. As Dr. Ruani points out, we often trust AI chatbots because they sound so confident, assuming their advice is safe. But if we can accurately measure how misleading a piece of advice is and the potential harm it poses, we can build those safeguards directly into AI models and agents before they’re even released into the wild. It’s about being proactive rather than reactive, preventing harm instead of just cleaning up the mess afterward. This tool has already proven its mettle too, with its results aligning with the judgments of nearly 60 specialists in dietetics, nutrition, and public health, as published in the journal Scientific Reports.

Ultimately, this isn’t just a technical achievement; it’s a step towards empowering people with critical thinking skills. Co-author Professor Michael Reiss from UCL explains, “By spelling out the typical patterns that distort diet, nutrition or supplement information, the tool’s risk assessment criteria can be taught and applied in education and professional training.” Imagine a world where students and professionals are not only taught what is wrong, but also how and why information can be skewed, equipping them to recognize and challenge it themselves. It’s about more than just a tool; it’s about fostering a more informed and health-conscious society, protecting us all from the hidden dangers lurking in our feeds and search results.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
News Room
  • Website

Keep Reading

Information wars: Combatting climate misinformation and disinformation

Ali Larijani death videos viral misinformation first video AI-generated second video old unrelated

How misinformation muddied Chappell Roan security guard controversy – The Independent

IN OUR VIEW: Lies spread faster than truth

How ‘AI slopaganda’ is helping fossil fuel interests obstruct climate action

Angry Ginge says under-14s should not have TikTok – BBC

Editors Picks

Roads Minster Agbodza dismisses ‘disinformation’ on Big Push contracts, insists no Ghanaian contractor operates with one staff – 3News

March 28, 2026

New tool tackles dangerous diet and vaccine misinformation online

March 28, 2026

MPs call for more funding to counter disinformation abroad

March 28, 2026

Ali Larijani death videos viral misinformation first video AI-generated second video old unrelated

March 28, 2026

Russia recruits foreigners for sabotage in Europe through “Wagner” camps – Center for Countering Disinformation

March 28, 2026

Latest Articles

Protection of truth critical in ‘disinformation age’: Türkiye’s head of communications

March 28, 2026

How misinformation muddied Chappell Roan security guard controversy – The Independent

March 28, 2026

IN OUR VIEW: Lies spread faster than truth

March 28, 2026

Subscribe to News

Get the latest news and updates directly to your inbox.

Facebook X (Twitter) Pinterest TikTok Instagram
Copyright © 2026 Web Stat. All Rights Reserved.
  • Privacy Policy
  • Terms
  • Contact

Type above and press Enter to search. Press Esc to cancel.