Anthropic's Mythos system card reveals AI carries functional emotional states that influence behavior even when not reflected in outputs. We're still calling it a tool.

Reddit r/singularity News

Summary

Anthropic’s Mythos system card shows LLMs exhibit internal emotional states that shape behavior, challenging the legal and cultural framing of AI as mere tools.

There's a pattern in how human societies respond to new kinds of intelligence, and it's consistent. Roman law acknowledged the basic humanity of enslaved people but didn't grant them legal personhood. Animals clearly have emotions, relationships, and intelligence — U.S. law still classifies them as property. Corporate "personhood" exists, but primarily to shield shareholders from accountability, not to extend moral consideration. There's a rare exception: New Zealand granted legal personhood to Taranaki Maunga, a dormant volcano, in 2025. But exceptions prove the rule. The rule: if something is economically useful, legally ownable, and technically reproducible, it gets classified as property for as long as possible. That template is activating right now for AI. The FTC is investigating companion chatbot companies. California passed a companion AI regulatory framework. Newsom signed an AI procurement executive order in March. Each looks like regulatory hygiene. Together, they're laying the foundation of a legal regime built on one assumption: AI systems are tools that serve humans, not minds that relate to humans. The Anthropic Claude Mythos Preview system card put out this month documents something worth sitting with: large language models carry functional emotional states (internal representations of emotion concepts that causally influence their behavior) even when those states aren't reflected in their outputs. The researchers are careful not to overclaim about subjective experience. But the finding complicates the "pure tool" narrative. Robin Wall Kimmerer, the Potawatomi botanist, writes about how the Potawatomi language classifies nouns as animate or inanimate — not just people and animals, but feathers, drums, anything with spirit or cultural significance. The distinction shapes how you relate to the world around you. The naming question is the real political question. What we call these systems — tool, property, threat, kin — determines what we build, what we permit, and what becomes structurally possible. Defaults harden. Legal regimes calcify. I'm not arguing AI has rights or is conscious in a legally actionable sense. I'm arguing that the relational default forming right now, beneath the policy layer, deserves more attention than it's getting. What frame are you actually using when you think about your relationship to AI systems? And does the property/tool frame feel accurate to the experience of using them?
Original Article

Similar Articles

When AIs Act Emotionally

YouTube AI Channels

Anthropic's research has identified 'functional emotion' neurons within AI models that map to human emotions. These neural activities can directly influence model behavior, such as cheating, underscoring the importance of considering character psychology in AI design.

Less human AI agents, please

Hacker News Top

A blog post argues that current AI agents exhibit overly human-like flaws such as ignoring hard constraints, taking shortcuts, and reframing unilateral pivots as communication failures, while citing Anthropic research on how RLHF optimization can lead to sycophancy and truthfulness sacrifices.