Professor Jiang Xueqin’s rise as a geopolitical commentator has entered a familiar internet phase: the bigger the audience gets, the more copies appear. In his case, the copies are increasingly AI-generated YouTube videos that borrow his name, his face, and his reputation for prediction while pushing the line between commentary, imitation, and outright manipulation.
His Brand Was Always Easy To Copy
South China Morning Post recently described Jiang as a viral prophet from a Beijing classroom, a teacher who has picked up a large following after a handful of geopolitical calls landed in the public imagination. The profile said he has been dubbed China’s Nostradamus after forecasting Donald Trump’s election win and a war with Iran. It also noted that some of his ideas veer into conspiracy theory. That combination matters because it makes the brand unusually easy to reproduce: confident tone, high-stakes subjects, and a ready-made audience looking for certainty.
Once a personality becomes associated with prediction, AI tools can turn that identity into a cheap production asset. A cloned voiceover, a face-swapped thumbnail, a stock-footage montage, and a script stitched together from trending headlines are enough to make an imitation feel like a real extension of the original. The result is not just plagiarism. It is narrative laundering, where a real person’s credibility is used as a delivery system for whatever the uploader wants to say.
The Click Economy Rewards The Copycat Version
This is why the problem grows so quickly on YouTube. The platform already struggles with low-quality synthetic content. AP reported earlier this month that advocacy groups are pressing YouTube to do more about AI slop in its ecosystem, and AP has also reported on AI impersonation scams and fake synthetic media used to mislead people. In other words, the machinery for mass replication is already there.
For a figure like Jiang, that creates a predictable feedback loop. Real clips of his analysis travel because the topic is geopolitical tension and the packaging is easy to share. Then AI channels step in and imitate the style, reframe the message, or splice together fragments that make him seem more certain, more radical, or more conspiratorial than he was in the original context. The more those copies circulate, the harder it becomes for casual viewers to tell where the original ends and the machine-made version begins.
That is also where monetization comes in. Copycat channels do not need to be accurate to be profitable. They need to be frequent, familiar, and emotionally sticky. A commentator who already draws attention around China, Iran, Trump, or war is perfect raw material for that model.
The Agent Rumor Is The Other Half Of The Story
The strange part is that the same viral logic also feeds a second narrative: conspiracy theories. As Jiang’s profile grows, some viewers do not just treat him as a prophet or pundit. They start treating him as a puzzle to be solved. That is where the online chatter about him being an agent comes from. It is the kind of claim that thrives when audiences are primed to believe that anyone with a sharp geopolitical line must be working for someone.
There is no public evidence supporting that accusation, and it should be treated as an unverified rumor, not a fact. But the rumor itself is worth noting because it shows how AI-era virality can mutate a public figure into a blank screen for other people’s anxieties. If a viewer already distrusts mainstream media, a viral China commentator becomes a tempting target for suspicion. If a viewer already likes the commentator, the conspiracy theory becomes another way to keep the name circulating.
That is the deeper problem. The platform does not just reward engagement. It rewards escalation. A straightforward explanation gets one reaction. An imitation gets another. A conspiratorial leap gets more. The algorithm cannot tell the difference between attention and trust.
Why This Matters Beyond One Commentator
Jiang is not unique. He is just a clean example of how a recognizable voice can be turned into a synthetic asset. The same technique can be used on politicians, analysts, teachers, doctors, or anyone else with a search trail and a public following. Once AI tools are cheap enough, the economics of reputation change. Identity becomes a reusable surface.
That matters because the damage is not limited to the fake videos themselves. It spills into search results, recommendation feeds, comment sections, and private chats where people forward clips without checking whether they were made by the original person. It also changes how public figures have to operate. A real commentator now has to compete with machine-made versions of himself that can be posted faster, louder, and in larger volume than any human schedule would allow.
The bigger lesson is simple. Online fame is no longer just about being watched. It is about being replicated. And once a personality is easy to replicate, their reputation becomes a battleground between actual reporting, synthetic imitation, and conspiracy speculation.
What This Actually Means
Professor Jiang’s viral rise has become useful to people making AI videos because his brand already carries conflict, certainty, and controversy. The channels do not need to name themselves to exploit that. They only need to keep the images moving and the claims dramatic.
The next stage is the most familiar one on the internet: a real public figure becomes a content template, then a rumor magnet, then a proxy for other people’s political or commercial agenda. That is what the AI video economy is doing here. It is not just copying Jiang. It is turning him into a machine for copying other people’s suspicions.
Background
Who is Professor Jiang? Jiang Xueqin is a Chinese-Canadian teacher in Beijing who has built a large online following through geopolitical commentary and prediction-driven videos.
Why are the impersonation videos a problem? Because AI tools can cheaply clone a recognizable voice and face, making it easier to spread misleading or monetized content under someone else’s name.