Oh boy, where do I even begin with the tangled web of information we find ourselves in these days? It’s like trying to navigate a dense fog, only instead of a trusty compass, we’ve got a dozen people shouting conflicting directions at us. For years, I’ve heard this charming anecdote about Socrates, the football player, having played for University College Dublin’s team. It was a fun little tale, even if it turned out he actually played for a graduate club called Pegasus. Then, nope, not Pegasus either, he just studied medicine at UCD. Wait, no, he was at the Royal College of Surgeons and decided against playing after seeing their team. The sheer number of variations and corrections I’ve witnessed over three decades of following the Students in the League of Ireland is astounding. This playful, low-stakes legend perfectly illustrates how easily facts can get blurry, even without malicious intent. It’s like a game of telephone that never ends, with each retelling adding a new twist or detail, true or not.
This “fog” of information isn’t new, of course. Humans have been muddling up stories since the dawn of communication. But the advent of artificial intelligence, my friends, has taken this natural human tendency and supercharged it into something truly potent and, frankly, a bit terrifying. We saw this in stark relief during recent fuel protests in Ireland. As tensions mounted on the roads, it became practically impossible to discern what was real and what was fabricated. The sheer volume of misleading narratives was overwhelming. There were fake documents flying around, to the point where the national police force, An Garda Síochána, had to issue warnings about bogus memos. And, naturally, we were bombarded with the usual suspects: recycled videos, out-of-context images, all designed to fan the flames of discontent. It was a digital wildfire, and AI, unfortunately, was pouring gasoline on it.
Think about it: the Defence Forces had to publicly clarify that their activities in Limerick were routine U.N. peacekeeping exercises, not some military takeover. Anyone who lives near an army base can tell you that military movements are a common sight, but to someone absorbing information through a highly filtered, algorithm-driven lens, a single tweet from an X (formerly Twitter) user can suddenly transform routine training into a conspiracy. It’s not just anonymous accounts either. Ciarán Mullooly, an MEP and former journalist, initially referenced some of this questionable imagery in a European Union meeting, only to walk it back later when the truth became clear. We even had TV Hercules, Kevin Sorbo, sharing an old anti-immigration protest video, claiming it was from the recent fuel protests. This isn’t just about human error anymore; AI’s ability to seamlessly alter and distribute content means a single lie can mutate and spread in countless plausible forms, targeting specific audiences with pinpoint precision.
And this is where it gets truly insidious. While AI allows misinformation to spread faster than ever, it also adds a chilling new dimension: mutation. If you ask an AI chatbot like Grok for a summary of an event, it can unwittingly (or intentionally) draw upon the very false narratives being circulated, presenting them as a simple, objective report. In the pre-AI era, even with social media, the spread of misinformation, while bad, was somewhat manageable. There was a limit to its reach, and the ways stories twisted and changed took time, often burning out before the embellished versions could fully take root. Remember the pandemic, and those endless messages about imminent total lockdowns? At least then, if you asked five people what a “total lockdown” meant, you’d likely get six different answers, quickly exposing the rumor for the bunk it was. But AI has not only accelerated the pace of rumor-spreading; it has weaponized the very “fog” that comes with tension, making it incredibly difficult to see clearly. The same rumor can now be endlessly repackaged with compelling pictures and videos, tailored to resonate with different groups, making confirmation bias far more likely to take hold. We’ve all been raised to believe “where there’s smoke, there’s fire,” a notion that, while sometimes true, can be disastrous when dealing with digital smoke machines.
The ultimate goal behind this deluge of misinformation isn’t necessarily a sudden, dramatic shift in how we consume media. It’s far more subtle and, perhaps, more dangerous: to erode trust in professional media outlets and sow deep distrust in institutions. We’ve witnessed the devastating effects of this erosion in other democracies. It’s a brutal battle because those who spread misinformation are selling alluringly simple messages, often using incredibly sophisticated technology to do so. They don’t need to win every argument; they just need to introduce enough doubt to make people question everything, creating a paralysis of belief.
So, what do we do in this bewildering landscape? We can’t entirely eliminate the spread of misleading information; that much is clear. But we can certainly work to reduce its impact. And it begins, quite fundamentally, with common sense. The “Socrates to UCD” story, while untrue, endured for so long because it was harmless and fun. Nothing of consequence hinged on its veracity. Similarly, tales like Andre the Giant and Samuel Beckett, which have a kernel of truth but are wildly exaggerated, are amusing precisely because they don’t carry serious weight. When something doesn’t truly matter, we shouldn’t waste our precious energy dissecting its truth. Leave that to people like me who enjoy a good factual unraveling. However, when it comes to issues that directly affect us, like those fuel protests and the government’s response, that’s when we need to be fiercely skeptical. We must scrutinize everything, questioning how and why information is being shared. Remember that a lot of what we see online, especially content that neatly aligns with our existing beliefs, could be AI-generated or manipulated. The true danger isn’t just in passively consuming misinformation, but in becoming an active participant, endlessly re-sharing and allowing these toxic narratives to fester and grow. We have a responsibility to ourselves and to a functioning society to demand clarity in this digital fog.

