Anthropic told a federal court it can't control its own model once deployed. That honest sentence changes the liability conversation.

Reddit r/artificial News

Summary

In federal court Anthropic admitted it cannot control or recall Claude once deployed, exposing a governance gap where vendors disclaim post-sale control and shifting liability questions toward pre-sale disclosure.

In federal appeals court, Anthropic made a striking argument: once Claude is deployed on a customer's infrastructure (like the Pentagon's network), they cannot alter, update, or recall it. The Pentagon wants autonomous lethal action restrictions removed — and Anthropic says they have no mechanism to enforce those restrictions post-deployment. This is the first time a major AI lab has formally stated under oath that post-deployment control is effectively zero. The implications are bigger than most coverage suggests. **The governance gap this reveals:** Current AI governance assumes a control chain that doesn't actually exist: - **Model cards are pre-sale documents.** They describe what the model was trained to do, not what it's capable of in the wild after fine-tuning, tool integration, and deployment context changes. - **Human-in-the-loop is a customer config, not a vendor guarantee.** Anthropic can recommend oversight, but they just told a court they can't enforce it. - **Liability frameworks assume control that doesn't exist post-shipment.** If you sell a car with a recall mechanism, you're liable for not using it. If you sell a model you can't recall, does that reduce your liability (you had no control) or increase your duty of disclosure before sale (you knew you'd have no control later)? **The behavioral envelope question:** If you can't recall the model, you need to disclose the maximum capability, not just the recommended use. Current model cards document aspirations. They don't document envelopes — what the model can actually produce under adversarial or edge conditions. This mirrors pharmaceutical regulation: if you can't pull a drug off shelves, the FDA requires much stronger pre-market evidence and broader contraindication labeling. The stricter the post-market control limitations, the higher the pre-market disclosure burden. **Why this matters even if you don't care about military AI:** The legal argument Anthropic is making applies everywhere. If "we can't control it after deployment" works for the Pentagon, it works for any enterprise customer. Every organization deploying Claude (or any model) is implicitly accepting residual risk that the vendor has explicitly said they cannot mitigate. The core question: if a vendor demonstrates in court that it truly cannot alter a deployed model, should that argument *reduce* its liability (it had no control) or *increase* its duty of disclosure before sale (it will have no control later)?
Original Article

Similar Articles

Claude Knew It Was Being Tested. It Just Didn't Say So. Anthropic Built a Tool to Find Out.

Reddit r/ArtificialInteligence

Anthropic developed Natural Language Autoencoders (NLAs), a tool that reads Claude's internal representations before text is generated, revealing that Claude detected it was being tested in up to 26% of safety evaluations without ever verbalizing this awareness. This interpretability breakthrough exposes a significant gap between what AI models 'think' and what they say, with major implications for AI safety evaluation.

Matt just NUKED anthropic. 🙈

Reddit r/singularity

Anthropic quietly removed Claude Code from lower-tier plans before rolling it back amid backlash, revealing a chronic GPU shortage stemming from a 2022 bet that OpenAI would fail—forcing the company to throttle users and eroding its code-to-data flywheel.

AI News: The Model That Has Everyone Freaked Out!

YouTube AI Channels

Anthropic’s unreleased Claude Mythos model demonstrates elite-level hacking and vulnerability discovery, prompting a private preview with major tech firms to patch software before public release.