The digital world is currently experiencing a seismic shift, one that’s blurring the lines between reality and fabrication. The culprit? Tools like OpenAI’s Sora, a revolutionary application capable of churning out incredibly realistic videos from simple text prompts. In the mere two months since Sora’s debut, social media platforms – TikTok, X, YouTube, Facebook, and Instagram – have been inundated with a surge of deceptive videos. This digital deluge has triggered alarm bells among experts, signaling the dawn of a new, highly effective era of disinformation and sophisticated fakes. While many of these AI-generated videos might seem harmless, like funny memes or adorable (but fake) images of babies and pets, a significant number are weaponized to fuel the very divisiveness that plagues online political discourse. We’ve already seen them pop up in foreign influence operations, with Russia, for instance, employing them in its ongoing campaign to undermine Ukraine. This onslaught raises a critical question: how can we, as humans navigating this increasingly complex digital landscape, discern what’s true from what’s a meticulously crafted illusion?
The challenge lies in the inadequacy of existing safeguards. Major social media companies do have policies demanding disclosure of AI use and broadly banning deceptive content. However, these rules have proven woefully insufficient against the technological leaps exemplified by tools like Sora. Researchers, who are diligently tracking these deceptive uses, are now pointing fingers at these companies, urging them to step up and ensure that users can differentiate between reality and AI-generated fiction. Sam Gregory, executive director of Witness, a human rights organization focused on technology’s threats, is unequivocal: “Could they do better in content moderation for mis- and disinformation? Yes, they’re clearly not doing that.” He further stresses that these platforms “could do better in proactively looking for A.I.-generated information and labeling it themselves? The answer is yes, as well.” This isn’t just about general fakery; we’ve seen these fakes have real-world consequences. Take, for example, a fake video about food stamps that circulated during a U.S. government shutdown, preying on the anxieties of real families struggling to feed themselves. Even a major news outlet, Fox News, fell for a similar fake, presenting it as genuine public outrage, an article that has since been quietly removed from their website. This incident highlights the profound impact these deceptive videos can have, not just on individuals, but on the broader informational ecosystem.
Beyond mocking vulnerable populations, these fakes have also been used to target prominent figures. One particularly unsettling video on TikTok showed a fabricated White House scene with what sounded like an AI-generated voice of former President Trump berating his cabinet over the release of documents concerning Jeffrey Epstein. This video, despite lacking an AI label, amassed over three million views in a matter of days, according to NewsGuard, a disinformation tracking company. This demonstrates how easily misinformation, propelled by these sophisticated AI tools, can spread like wildfire, influencing public perception without any clear indication of its artificial origin. The current system primarily relies on content creators to disclose their use of AI, a reliance that has proven fragile, as creators often choose not to. And while platforms like YouTube and TikTok possess the technological capabilities to detect AI-generated videos, they don’t always flag them for viewers immediately. Nabiha Syed, executive director of the Mozilla Foundation, a tech-safety nonprofit, expresses her dismay, stating that social media companies “should have been prepared.” The rapid advancement of AI video generation has clearly outpaced the development and implementation of robust detection and labeling mechanisms, leaving users vulnerable.
In response to growing pressure, the companies behind these AI tools are attempting to make their generated content identifiable. Sora and Google’s rival tool, Veo, both embed visible watermarks into the videos they produce. Sora, for instance, adds a “Sora” label to each video. Furthermore, both companies include invisible metadata, a digital fingerprint that can be read by a computer, establishing the origin of each fake. The intention is clear: to inform viewers that what they are seeing is not genuine and to provide platforms with the digital signals needed for automatic detection. Some platforms are indeed beginning to leverage this technology. TikTok, recognizing the alarming persuasiveness of these fake videos, recently announced stricter rules regarding the disclosure of AI use and promised new tools to empower users to control the amount of synthetic content they encounter. YouTube, for its part, utilizes Sora’s invisible watermark to append a small label indicating that AI videos have been “altered or synthetic.” As Jack Malon, a YouTube spokesman, noted, “Viewers increasingly want more transparency about whether the content they’re seeing is altered or synthetic.” This move, while welcome, often comes with a significant delay, with labels sometimes appearing only after millions of people have already viewed the deceptive content, highlighting the ongoing challenge of real-time detection and user protection.
However, the effectiveness of these labeling efforts is already being undermined by those with malicious intent. It’s becoming increasingly clear that circumventing disclosure rules is surprisingly easy. Some simply disregard them entirely. Others manipulate the videos to remove the identifying watermarks. The Times, for example, unearthed dozens of instances of Sora videos circulating on YouTube without the automated AI label. A cottage industry has even sprung up, with companies offering to remove logos and watermarks for a fee. Furthermore, the very acts of editing or sharing videos can inadvertently strip away the embedded metadata, further obscuring their AI origin. Even when watermarks remain visible, the rapid-fire scrolling habits of social media users often mean these subtle indicators are easily missed. A New York Times analysis of comments on a TikTok video about food stamps revealed that nearly two-thirds of over 3,000 respondents reacted as if the video were real, despite its artificial nature. This underscores the profound impact of these fakes and the difficulty of relying solely on individual vigilance. Sam Gregory aptly points out, “There’s kind of this individual vigilance model. That doesn’t work if your whole timeline is stuff that you have to apply closer vigilance to. It bears no resemblance to how we interact with our things.” The onus, therefore, remains on the platforms and AI developers to implement more robust, user-friendly safeguards.
The problem is further exacerbated by the motivations of the social media platforms themselves. While OpenAI states it prohibits deceptive uses of Sora and acts against violators, acknowledging that its app is just one of many similar tools, the ecosystem-wide nature of the problem means a unified approach is necessary. A Meta spokesman, representing Facebook and Instagram, admitted the difficulty of labeling every AI-generated video, given the rapid evolution of the technology, but asserted efforts to improve their detection systems. X and TikTok, interestingly, remained silent when asked about the flood of AI fakes. Alon Yamin, CEO of Copyleaks, a company specializing in AI content detection, offers a stark, financially driven explanation: social media platforms, he argues, lack a short-term financial incentive to restrict these videos as long as users continue to engage with them. “In the long term, once 90 percent of the traffic for the content in your platform becomes A.I., it begs some questions about the quality of the platform and the content,” Yamin concedes, suggesting that a future where AI content dominates might eventually force a strategic shift. However, in the immediate present, the allure of engagement often trumps the imperative for authenticity. This lack of immediate financial motivation creates a fertile ground for disinformation, fraud, and foreign influence operations to thrive. Sora videos have already been observed in recent Russian disinformation campaigns on platforms like TikTok and X, with crudely obscured watermarks, exploiting sensitive political issues and fabricating emotional narratives like weeping soldiers. The chilling conclusion from former State Department officials James P. Rubin and Darjan Vujica, who fought foreign influence operations, is that AI advancements are intensifying efforts to destabilize democratic societies. They cite examples like AI videos in India designed to stoke religious tensions, further underscoring the severe and widespread threat these technologies pose to global stability and truth itself.

