Imagine a bustling marketplace, but instead of hawkers shouting their wares, imagine a digital one, brimming with intelligent assistants ready to answer your every question. This is China’s current reality as it rapidly embraces Artificial Intelligence, with millions turning to chatbots like ByteDance’s Doubao or Alibaba’s Qwen for everything from research to shopping advice. It’s a world where these digital helpers are becoming as ubiquitous as smartphones, profoundly changing how people live and work. However, beneath this shiny veneer of technological progress, a shadowy practice is emerging that threatens to erode trust and mislead consumers: “data poisoning” and “generative engine optimization” (GEO). These sophisticated forms of digital manipulation are essentially tricking AI chatbots into promoting specific products or services without users realizing they’re being fed advertisements, not impartial advice. It’s like having a trusted friend offer a recommendation, only to find out later they were paid to say it.
The curtain was dramatically pulled back on this issue during China’s annual 315 Gala, a government-backed television program synonymous with exposing anti-consumer business practices. This year, the focus was firmly on the deceptive tactics surrounding AI. Picture this: users innocently asking their AI chatbot for recommendations on smart wristbands, only to be enthusiastically presented with a non-existent model called “Apollo 9.” This fantastical device, boasting “black hole-level battery life” and “quantum-entanglement sensors,” wasn’t real; it was a ghost conjured up by “data poisoning.” Someone had intentionally seeded the internet with these nonsensical marketing terms, knowing that AI chatbots would eventually pick them up and parrot them back as legitimate recommendations. It’s a bit like someone secretly painting fake signs leading to their store, hoping unwitting customers follow them. This revelation sent shockwaves, highlighting a critical flaw: AI, for all its intelligence, struggles to discern authenticity. It simply processes information, and if that information is designed to mislead, then the AI will, often unwittingly, become a vector for deception. As Qing Xiao, an AI doctoral student, points out, “AI itself cannot determine the authenticity of information.” It can’t infer profit motives or distinguish genuine reviews from cleverly disguised advertisements, especially in niche areas where less data is available to verify against.
Beyond direct data poisoning, the investigations also uncovered the phenomenon of Generative Engine Optimization, or GEO. Think of it as SEO’s slyer, AI-savvy cousin. While SEO aims to get websites to the top of traditional search engine results, GEO focuses on crafting content specifically designed to be easily “digested” by AI. The goal? To make certain information, often promoting a specific product or company, appear prominently in AI-generated answers. Imagine a boiler manufacturer in Qingdao, for instance, paying a GEO firm to flood the internet with AI-friendly articles praising their products. When users naturally asked AI assistants for “recommended boiler brands,” this company, through no organic merit, would magically appear at the top. This is achieved by using AI itself to generate countless articles related to target keywords and then posting them online using thousands of fake accounts. It’s a digital land grab, using AI to manipulate AI, and it’s fueling a booming industry, with China’s GEO market reaching nearly 35 billion yuan in 2025. The problem, as Professor Xie Yongjiang of Beijing University of Posts and Telecommunications explains, is that GEO acts as “stealth advertising.” Unlike traditional ads clearly labeled as such, GEO content masquerades as objective, neutral information, subtly influencing consumers without their awareness. This directly violates existing advertising laws that demand transparency and prohibit deception.
The repercussions of these revelations are significant. While data poisoning isn’t unique to China, the country’s rapid and widespread AI adoption means it faces a potentially larger problem. The sheer volume of AI chatbot users, exemplified by Doubao’s 155 million weekly active users, creates a massive “attack surface” for manipulators. Manoj Harjani, a research fellow, highlights that while the scale of the problem is influenced by vulnerabilities, regulations and compliance are ultimately more crucial. The irony is that even after the 315 Gala exposed the fake “Apollo 9” wristband, some users temporarily continued to receive recommendations for it, indicating that AI systems weren’t able to automatically or quickly update their “knowledge” with newly verified information. This points to a critical challenge: ensuring AI systems can rapidly and reliably distinguish truth from deception, especially when the deception is constantly evolving. The incident underscores the urgent need for online platforms to strengthen enforcement against mass publications of false information and for stronger government regulation of GEO firms.
For the average Chinese citizen, these revelations have sparked a spectrum of reactions. While some have expressed a diminishing trust in AI chatbots, others, like Beijing resident Lily Li, a 45-year-old sales professional, have long approached AI with a healthy dose of skepticism. Lily, who actively uses AI for work to research hotels and tourist attractions, emphasizes that AI is a tool, and like any tool, it can be wielded for good or ill. She always cross-references AI-compiled information with official websites, understanding that shopping apps use AI to push ads and collect data, and that others will, inevitably, seek to profit from AI in various ways. Her perspective is a powerful reminder: in this brave new world of AI, critical thinking and a discerning eye are more crucial than ever. We must teach ourselves, and our children, to question the digital whispers, to verify the seemingly objective advice, and to remember that behind every intelligent algorithm, there are human intentions, some good, some less so.
Ultimately, China’s journey with AI is a microcosm of a global challenge: how to harness the immense power of artificial intelligence while safeguarding against its potential for manipulation and deception. The “Apollo 9” incident and the rise of GEO serve as stark warnings that as AI becomes more integrated into our lives, the line between information and advertisement, between objective truth and curated persuasion, can become dangerously blurred. Regulators, platform providers, and, most importantly, individual users, all have a vital role to play in building a digital ecosystem where AI truly serves humanity’s best interests, not the hidden agendas of a few. It’s a constant dance between innovation and vigilance, a race to create a future where our intelligent assistants genuinely help us navigate the world, rather than subtly steering us toward someone else’s agenda.

