
China Accused of Copying US AI Models, Costing Billions
Major US AI companies including OpenAI, Google, and Anthropic report that Chinese firms are using distillation techniques to extract capabilities from American AI models, raising national security concerns.
The AI arms race has a new battleground: model theft. Major U.S. AI companies are now publicly accusing Chinese firms of systematically extracting capabilities from American AI models through "distillation" techniques — and the cost runs into billions.
What Is Model Distillation?
Distillation involves making large-scale queries to an AI model to extract and reverse-engineer its capabilities. Think of it as asking a master chef every possible question about their recipes until you can recreate their entire cookbook — except in this case, the "cookbook" is a multi-billion-dollar AI model.
Anthropic has specifically blocked Chinese-controlled companies from using Claude and identified three Chinese AI labs — DeepSeek, Moonshot, and MiniMax — as illicitly extracting model capabilities.
Why This Matters
The implications extend far beyond corporate IP theft:
- Safety concerns: Distilled models often lack the safety guardrails designed to prevent malicious use
- National security: U.S. companies report these attacks pose risks beyond any single company
- Economic impact: American companies are measuring the financial damage in billions of dollars
- Fair competition: Chinese labs can offer similar capabilities at lower costs since they skipped the expensive R&D phase
Industry Response
OpenAI, Google, and Anthropic are now sharing intelligence about these attacks, an unusual level of cooperation between fierce competitors. This suggests the threat is serious enough to transcend normal business rivalries.
Frequently Asked Questions
What is AI model distillation? A technique where someone makes large-scale queries to extract and replicate an AI model's capabilities without permission.
Which Chinese companies are accused? DeepSeek, Moonshot, and MiniMax have been specifically identified by Anthropic.
What are the safety risks? Distilled models often lack safety guardrails, making them potentially dangerous for malicious applications.
Key Takeaways
- U.S. AI companies are sharing intelligence about Chinese model distillation
- Three Chinese labs identified: DeepSeek, Moonshot, MiniMax
- Distilled models lack safety guardrails, creating security risks
- Unprecedented cooperation between rival U.S. AI companies signals severity
Stay ahead of the AI curve. Follow @AiForSuccess for daily insights.
📬 Want more AI solopreneur insights?
Subscribe to our weekly newsletter →Related Articles

AI Startups Absorbed $242 Billion in Q1 2026 — a Record 81% of All VC Funding
Global AI startup funding hit a record $297 billion in Q1 2026, with AI companies capturing $242 billion or 81% of all venture capital deployed worldwide.

Claude Design: Anthropic's Bold Move Into AI Prototyping
Anthropic launches Claude Design, a tool that turns text prompts into prototypes, visual assets, and handoff-ready outputs for designers and developers.

Claude Opus 4.7: Anthropic's Most Capable Model Yet
Anthropic releases Claude Opus 4.7 with stronger coding, higher-resolution image support, and new cybersecurity safeguards at unchanged pricing.