When Anthropic published a technical blog post on February 23, 2026, accusing three Chinese AI laboratories of stealing Claude's most valuable capabilities through coordinated distillation attacks, it triggered something unprecedented: the three most fiercely competitive AI companies in the world , OpenAI, Anthropic, and Google , began voluntarily sharing their most sensitive security intelligence. The fact that these companies had to form an emergency intelligence-sharing pact says everything about how severely the AI industry had underestimated the distillation threat. The fact that it took 16 million stolen conversations before anyone went public says something even more uncomfortable about how long Silicon Valley looked the other way.
What Actually Happened
To understand what distillation means in practice, consider the economics. Training a frontier AI model like Claude costs hundreds of millions of dollars , in compute, data, human feedback, and years of iterative development. Distillation attacks let a competitor skip most of that. The technique works by flooding a target AI system with carefully crafted prompts designed to elicit its most valuable reasoning behaviors, then using those AI-generated outputs as training data for a cheaper model. Done at sufficient scale, distillation can transfer the implicit knowledge of a $400 million training run into a competitor's system for a few thousand dollars in API fees , before the target company even realizes what's happening.
According to Anthropic's February 2026 disclosure, DeepSeek, Moonshot AI, and MiniMax collectively executed more than 16 million unauthorized exchanges with Claude, generated from approximately 24,000 fraudulently created accounts. The three campaigns shared a strikingly similar playbook: shared payment methods, coordinated timing that Anthropic's security team described as "load balancing," and highly repetitive prompt structures specifically targeting the capability domains most valuable to replicate. This was not opportunistic hacking. It was coordinated, systematic industrial espionage at a scale that required significant organizational resources and deliberate strategic planning.
The breakdown by company reveals calculated strategic intent. MiniMax drove the largest volume, generating over 13 million exchanges targeting Claude's general capabilities. Moonshot AI executed more than 3.4 million exchanges with surgical focus on agentic reasoning, tool use, coding, data analysis, computer-use agent development, and computer vision , precisely the capabilities that define the next generation of AI products. DeepSeek's campaign was smallest in raw volume at 150,000+ exchanges, but arguably most disturbing in intent. Rather than targeting performance benchmarks, DeepSeek specifically probed Claude's safety-tuning behaviors, censorship-safe responses, and policy-sensitive queries. DeepSeek was not trying to copy Claude's intelligence , it was trying to reverse-engineer how Anthropic had constrained it.
Why This Matters More Than People Think
The immediate financial damage is measurable. US officials estimate that AI model distillation theft costs Silicon Valley billions of dollars annually , a figure encompassing not just compute costs from unauthorized API abuse, but the far larger losses from having years of training investment and safety research replicated for pennies on the dollar. Anthropic spent hundreds of millions building Claude's safety architecture, its agentic reasoning stack, and its Constitutional AI framework. DeepSeek, Moonshot, and MiniMax did not have to spend any of that. They needed 16 million conversations and a few thousand fake accounts. The return on that investment , assuming even a fraction of Claude's capabilities transferred , likely exceeds the R&D budget of many mid-sized technology companies.
The deeper threat is the compounding competitive advantage that successful distillation creates. Every capability that Chinese labs extract from frontier US models is a capability they can iterate on, combine with proprietary research, and deploy in products without the safety constraints that cost US companies so much to build in. Google's own 2026 security report, released alongside the Frontier Model Forum announcement, noted a significant increase in distillation attacks targeting Gemini in the fourth quarter of 2025 , suggesting the problem was accelerating well before Anthropic went public. If industrial-scale distillation continues unchecked, the gap between US frontier AI and Chinese AI narrows not because Chinese labs are innovating faster, but because they are systematically extracting the innovations of their competitors at zero R&D cost.
The national security implications extend beyond commercial losses. The Trump administration's AI Action Plan, released in early 2026, explicitly calls for an information-sharing center dedicated to combating adversarial distillation, framing it as a matter of national security rather than industry compliance. For Washington, the nightmare scenario is not a Chinese company beating OpenAI at a benchmark , it is a Chinese AI system with Claude-level capabilities but without Claude's safety constraints, deployed at scale in contexts where those constraints would have mattered.
The Competitive Landscape
The three companies named by Anthropic are not bit players. DeepSeek became globally known in January 2026 when its open-source models delivered performance rivaling GPT-4o at roughly one-tenth the compute cost , a disclosure that wiped more than $500 billion from US tech stocks in a single trading day and forced Washington to reconsider its AI export control strategy. Moonshot AI, founded in 2023 with backing from Alibaba and Tencent, has built one of China's most capable long-context models and is expanding aggressively into agentic AI. MiniMax has developed multimodal capabilities that directly compete with GPT-4o and is targeting enterprise workflow automation.
On April 6, 2026, OpenAI, Anthropic, and Google began formally sharing threat intelligence through the Frontier Model Forum, an industry nonprofit the three companies co-founded with Microsoft in 2023. The Forum is now functioning as an early-warning system: when one lab detects a novel distillation technique, it alerts the others. For context on how extraordinary this is , these companies compete for the same top AI researchers with eight-figure compensation packages, litigate against each other in multiple jurisdictions over talent and IP, and would normally never disclose details of a security breach that could undermine commercial confidence. The willingness to now openly exchange distillation attack data suggests the threat has reached a severity that overrides the normal competitive calculus entirely.
Hidden Insight: The Architecture of Vulnerability
Here is the uncomfortable truth that the Frontier Model Forum announcement papers over: for distillation attacks to have reached 16 million exchanges before anyone went public, the AI industry's standard defenses failed completely. Anthropic described campaigns with "coordinated timing" and "load balancing" , terms implying the attacks were engineered specifically to evade automated detection systems. Twenty-four thousand fraudulent accounts, coordinated with enough sophistication that months of sustained theft went undetected. Either the companies' abuse-detection systems were inadequate, or they detected the attacks and chose to gather evidence before acting publicly. Neither version is reassuring to the companies, their customers, or their regulators.
The willingness to share intelligence now raises a harder question: why were they not sharing before? The Frontier Model Forum was founded in 2023, ostensibly to coordinate on exactly this category of existential risk. For nearly three years, competitive dynamics appear to have prevented the open intelligence-sharing that would have been necessary to detect distillation attacks at scale. Each lab apparently treated its distillation data as proprietary security intelligence rather than as an industry-level threat requiring collective defense. Anthropic's disclosure changed that calculus , but only after the damage was done.
The deepest structural problem is architectural. The entire business model of API-based frontier AI depends on making the most powerful capabilities accessible to paying customers. That openness is not a bug , it is the product. But it means every new capability added to Claude, GPT-5, or Gemini immediately becomes a distillation target. Detection systems can improve, and the Forum can share attack signatures. But you cannot prevent distillation purely through better detection. The industry must rethink how frontier capabilities are exposed , through rate limiting, behavioral fingerprinting, output watermarking, or differential capability tiers that restrict the most transferable reasoning behaviors from the most open access tiers. None of those solutions exist at production scale today. The industry is playing defense in a game where the offense has a structural advantage baked into the product architecture itself.
What to Watch Next
In the next 30 days, watch whether the US Commerce Department moves to designate AI model distillation as a sanctionable activity under export control regulations. The AI Action Plan provides political cover for exactly this kind of aggressive intervention, and the Frontier Model Forum's threat intelligence would give Commerce the technical evidence to build specific cases. The most likely mechanism is an extension of chip export controls , not to physical hardware, but to trained model weights or distilled outputs. Whether that framework can be enforced against Chinese labs operating outside US jurisdiction is the central open question that no regulator has yet answered.
In the next 90 days, watch whether Moonshot AI, MiniMax, or DeepSeek acknowledge the attacks publicly. So far, none of the three has issued a substantive response. The most revealing signal will come from product launches: if their next-generation agentic AI systems demonstrate capabilities that closely mirror Claude's specific domains , computer use, multi-step tool execution, agentic reasoning under uncertainty , without any plausible organic research path to those capabilities, the distillation campaigns will be difficult to dismiss. Watch also for OpenAI and Google to disclose their own distillation statistics. Anthropic has set a transparency precedent. The industry pressure to follow will be substantial , and what those disclosures reveal will determine whether 16 million is the beginning of this story, or just the part we know about so far.
The first great threat to American AI supremacy was not a smarter model or a better chip , it was 24,000 fake accounts and 16 million stolen conversations.
Key Takeaways
- 16 million unauthorized exchanges , DeepSeek, Moonshot AI, and MiniMax collectively conducted 16M+ distillation attacks on Claude via ~24,000 fraudulent accounts, disclosed by Anthropic on February 23, 2026
- MiniMax executed 13M+ interactions , the largest attacker by volume, targeting Claude's broad general capabilities across reasoning and agentic domains
- DeepSeek targeted safety architecture , its 150,000+ exchanges focused on replicating Claude's safety-tuning and censorship-safe behaviors, not just raw performance benchmarks
- Frontier Model Forum now operational as threat-sharing hub , OpenAI, Anthropic, and Google began formally exchanging distillation attack intelligence on April 6, 2026, a historic collaboration between bitter competitors
- Billions in annual losses estimated , US officials put the annual cost of distillation theft in the billions; Google's 2026 report documented a measurable acceleration in attacks targeting Gemini in Q4 2025
Questions Worth Asking
- If API-first AI businesses are structurally vulnerable to distillation attacks, is there any version of the current business model that is defensible against well-resourced, state-backed competitors who face no legal risk in their home jurisdiction?
- DeepSeek's specific focus on copying Claude's safety-tuning behaviors , not its performance , raises a question the AI safety community has not addressed publicly: what are the consequences if the world's next dominant AI system has frontier capabilities but deliberately omits safety constraints?
- If your organization depends on API access to foundational AI models, what is your contingency plan if those APIs become significantly restricted as the industry arms its defenses , or if the underlying models are found to have been systematically compromised?