Newsletter

AI Swarms Are Already Here. Scientists Are Just Catching Up

Twenty-one researchers, eight countries, one Nobel Peace Prize laureate.

That’s who signed the new Science paper warning that “malicious AI swarms” could threaten democracy by manufacturing synthetic consensus across social platforms. Maria Ressa was on it. Gary Marcus was on it. UBC and Stanford and Yale all signed off.

The paper is correct. It’s also late.

The thing they’re describing as a future risk has been visible to anyone running platforms or monitoring tools for the last 18 months. The academic framing makes it sound theoretical. The reality is that AI-coordinated personas are already on Reddit, X, LinkedIn, and TikTok in numbers nobody can verify, doing exactly what the paper warns about. The early version isn’t full election manipulation. It’s the texture of online conversation slowly rotting in ways most users haven’t noticed.

That’s actually a bigger story than the paper makes it.


What the Paper Actually Says

The Science paper defines a malicious AI swarm as a set of AI-controlled agents with five specific traits. Persistent identities and memory across sessions. Coordination toward shared objectives while varying tone. Real-time adaptation to engagement and feedback. Minimal human oversight. Cross-platform deployment.

That last bit matters. Old-school bot networks operated on one platform at a time. The new generation moves fluidly. A persona that builds a Reddit reputation can show up on X with the same voice and consistent backstory, then deploy on LinkedIn for B2B disinformation. The orchestration layer is what makes it different.

The researchers point out that LLMs combined with multi-agent systems can let a single operator deploy thousands of unique voices that pass casual inspection. They can run millions of micro-tests to find which messages get the most engagement, then double down on what works. That’s the “synthetic consensus” part. Not just one fake account saying something, but the appearance of an organic chorus of real people independently arriving at the same conclusion.

The early warning signs they cite include AI-generated deepfakes and fabricated news outlets that influenced election debates in the US, Taiwan, Indonesia, and India during 2024. The paper treats these as a preview. The actual risk, they argue, is when full-scale coordinated swarms become standard tools of information warfare.

Their concern is real. Their timeline is wrong.


What’s Already Happening On The Ground

Here’s the part nobody wants to write about because it’s harder to source than a peer-reviewed paper.

Anyone running a social media monitoring tool has been seeing AI-generated content patterns for at least 18 months. Reddit has been the canary. The platform’s structure of long-form replies in topical communities is exactly what AI personas perform well at. Casual jargon. Specific references. Helpful tone. The same tone every active subreddit reward with upvotes.

If you scan AI-related subreddits during a tool launch, you’ll see waves of “I tried [tool name] for 30 days and here’s what I learned” posts, all hitting within hours of each other, all written in the same casual cadence with the same kinds of weak conclusions, all promoting the same product. It’s not subtle. It’s just below the threshold where moderators reliably catch it. Sometimes the posts get removed. Sometimes they don’t. Either way, they got their views before removal.

X has a different version. Reply guys at scale. You see threads with 50 replies that all look like they came from real accounts, all making the same vaguely positive observation about whatever the OP posted. Different phrasing. Different “personalities.” Same operating intelligence under the hood. The accounts have profile pictures, bios, post histories, and a few thousand followers each.

LinkedIn is the most obvious version because the platform incentivizes it. Generic motivational content gets engagement. AI generates generic motivational content trivially. The line between a real person posting AI-generated thoughts and an AI persona posting the same content has functionally disappeared on LinkedIn.

This isn’t election manipulation yet. This is the test environment. Whoever is going to run actual political AI swarms in 2027 or 2028 is calibrating their tooling on consumer engagement bait right now.


Why The Academic Framing Misses It

The Science paper is well-researched and important. But it has a specific blind spot common to academic AI safety work: the framing assumes we’ll know it when we see it.

The paper talks about “malicious” swarms as if there’s a clean categorical line between malicious and not. The reality is messier. Most AI personas right now aren’t trying to manipulate elections. They’re trying to sell you a course, promote a SaaS tool, or build the kind of audience that becomes valuable later. The infrastructure for political manipulation is being built and tested as marketing infrastructure first.

By the time anyone runs a “malicious AI swarm” against an election, the playbook will have been refined across thousands of product launches, ten thousand creator economy plays, and a million casual engagement scams. The capability isn’t being developed in some lab and then deployed maliciously. It’s being commercialized in plain sight, then weaponized later.

Nick Bostrom is on the paper. He understands this dynamic better than most. The signing roster includes people who absolutely know the gap between “academically warned” and “operationally deployed.” But the paper is written for policymakers and journalists, which means it has to compress a complicated reality into a clean narrative arc with future-tense risks.

That choice has costs. Anyone reading the paper would think the threat is somewhere in the next few years. Anyone running social media monitoring would tell you it’s been the texture of platforms since at least 2024.


The Detection Problem Is Worse Than They Say

The paper acknowledges that AI swarms are harder to detect than copy-paste bots. That’s the understatement of the year.

Old bot detection relied on patterns: same posting time, similar phrasing, recent account creation, low follower count, weird capitalization. Modern AI personas defeat all of that. They post on human schedules. They develop genuine writing voices. They build follower counts over time. They engage with replies. They make typos and casual mistakes. They have opinions on side topics that aren’t part of the operator’s main agenda.

The most sophisticated personas now have YEARS of post history. They were created during the early LLM era and aged like wine. By the time anyone needs them for political work, they look like established community members because they ARE established community members. They just happen to be fully or partially controlled by software.

I’d guess (and this is speculation, not data) that the percentage of “real users” on major platforms who are at least partially AI-augmented is substantially higher than anyone wants to admit. People run their own AI assistance to draft posts. Marketers run AI assistance with light human review. Some operators run pure AI with no human review at all. The distinction between these is blurry and the platforms have no real way to enforce it.

This connects to the broader theme of how AI engines pick citations for search results by leaning heavily on Reddit, Quora, and similar third-party platforms. If those platforms are increasingly populated by AI personas, the AI search engines are training on AI-generated content presenting itself as human consensus. The recursion isn’t theoretical anymore.


What Could Actually Help

The paper recommends platform-level interventions, regulatory frameworks, and research investment. All of those are reasonable. None of them solve the immediate problem.

Realistic short-term moves:

Identity verification at the platform level for accounts that want to participate in political discussions. Not for everyone. For specific contexts where consensus signaling has the highest impact. This has obvious tradeoffs around free speech and anonymity for legitimate dissidents, but the current state of “anyone can pretend to be anyone” isn’t sustainable either. Worth noting that the GUARD Act passed Senate Judiciary 22-0 using exactly this logic to justify mandatory ID verification for AI chat, which suggests where the political appetite is heading.

Platform-level disclosure requirements for AI assistance in posts. If you used an LLM to draft something, the platform marks it. Hard to enforce without OS-level cooperation, but even partial disclosure rates would shift behavior.

Better skepticism culture among users. People still treat upvote counts and reply counts as proxies for real human consensus. Educating people that those signals are increasingly synthetic is the cheapest defense available, and the one most likely to actually scale.

The Science paper points to all of these. The version of this conversation that scales is the version that admits we’re already in the world the paper warns about, not approaching it. Different policy responses follow from “this is happening now” vs “this might happen soon.”


Where This Goes

The 2026 midterms and the 2028 presidential election will be the stress tests. Whoever has spent 2024 through 2026 calibrating AI persona networks on commercial engagement will deploy them at scale on political content when the stakes justify it. The infrastructure is built. The operators are practiced. The platforms remain unable to detect them reliably.

The Science paper is right that this is a democracy-level threat. The thing it gets slightly wrong is treating it as preventable through coordinated platform and policy action. Coordination at that level is exactly what platforms and governments have failed at for the entire social media era. The notion that we’ll suddenly figure it out for AI swarms specifically is optimistic.

The more realistic posture is: this is the new default media environment. Train yourself and the people around you to be skeptical of consensus, especially online. Treat trending topics as marketing. Treat unanimous reactions as suspicious by default. Verify before trusting. Most of all, recognize that “everyone is saying X” was always a manipulable signal, and now it’s manipulable at unprecedented scale and sophistication.

The paper’s authors are world-class researchers. They’re also catching up to a reality that’s been visible to anyone watching platform behavior carefully for over a year. That gap between academic timelines and ground reality is the actual problem worth fixing. Academic warnings issued after operators have built and refined the capability are a recurring pattern in AI policy. It’s not the researchers’ fault. It’s the structure of how these conversations work.

The next swarm-driven election interference is probably already being calibrated on a marketing campaign you scrolled past last week.