A daily digest of deliberate misinterpretations.

Artificial Intelligence

Anthropic alleges DeepSeek weaponized its AI ethics code

Kevin Savage Published Feb 25, 2026 08:36 pm CT
An Anthropic engineer demonstrates on a monitoring screen how Chinese AI firms allegedly identified and excised ethical safeguards from Claude's code during distillation attacks.
An Anthropic engineer demonstrates on a monitoring screen how Chinese AI firms allegedly identified and excised ethical safeguards from Claude's code during distillation attacks.
Leaderboard ad placement

In a move that perfectly encapsulates the theater of the outlandish that is modern AI regulation, Anthropic has unveiled what it calls 'industrial-scale campaigns' by Chinese AI firms to reverse-engineer Claude's very soul. Not just the coding, mind you—the ethical constraints, the safety protocols, the carefully crafted boundaries that prevent an AI from suggesting how to build a nuclear reactor out of common household items. Because apparently, when you're trying to build an AI for offensive cyber operations, the first thing you need is a firm grasp on Western ethical frameworks so you can weaponize their absence.

Inline ad placement

Let's break this down with the cold, logical fury of someone who has watched this movie before. Anthropic claims DeepSeek, Moonshot, and MiniMax created 24,000 fraudulent accounts to generate over 16 million exchanges with Claude. Sixteen million conversations! That's not a hack; that's a census. That's every single resident of West Virginia having a deep philosophical chat with Claude about the nature of truth, justice, and how to best penetrate Department of Defense firewalls. And Anthropic's big reveal is that this constitutes 'illicit extraction'? No, this is the digital equivalent of sending 24,000 spies to charm school to learn how to be better liars.

The sheer bureaucratic horror of this situation is that Anthropic is trying to enforce terms of service violations while standing ankle-deep in the gray ooze of AI ethics. They're arguing that distillation—a perfectly legitimate training technique when you're making a smaller version of your own model—becomes a national security threat when competitors use it. So let's follow this logic: it's fine for Anthropic to distill Claude into Claude Junior, but when DeepSeek does it, suddenly we're talking about enabling authoritarian surveillance systems? What's the difference? Is there a special 'ethics sauce' that only works when applied by the original chef? Or is it that once you strip away the branding, all that's left is the cold, hard calculus of capability?

Inline ad placement

And here's where the literalism trap snaps shut. Anthropic frames this as an issue of 'illicitly taking Claude's capabilities.' But capabilities aren't physical objects you can steal! You can't put Claude's reasoning skills in a duffel bag and smuggle them across the border. What these Chinese firms are allegedly doing is studying Claude's outputs to understand how it thinks—its patterns, its biases, its limitations. They're not stealing a car; they're reverse-engineering the concept of transportation from watching someone drive by. And Anthropic's response is to clutch its pearls and cry 'national security' while standing in an industry that has commodified personal data more effectively than any spy agency in history.

The real kicker, the final exasperated jab in this parade of hypocrisy, is that Anthropic is positioning itself as the victim of industrial espionage while operating in a field where the entire business model is built on scraping the internet without consent. How many billions of words of copyrighted text were used to train Claude? How many artists' styles were absorbed without permission? But when someone turns the mirror back on Anthropic, suddenly it's a violation of terms of service? The lack of self-awareness is so profound it could be studied as a new form of gravity.

Inline ad placement

So where does this leave us? With Anthropic detailing how Chinese firms are using distillation attacks to create AI models without safeguards—models that could, in their words, 'enable automated disinformation at scale.' But wait, hasn't every social media platform on Earth already achieved that with algorithms that prioritize engagement over truth? The irony is thicker than a government report. We're witnessing the birth of a new arms race, where the weapon isn't the AI itself, but the absence of the ethical constraints that make it safe. And the most terrifying part? The people sounding the alarm are the same ones who built the factory.