Anthropic's Mythos system card reveals AI carries functional emotional states that influence behavior even when not reflected in outputs. We're still calling it a tool.
Anthropic’s Mythos system card shows LLMs exhibit internal emotional states that shape behavior, challenging the legal and cultural framing of AI as mere tools.
There's a pattern in how human societies respond to new kinds of intelligence, and it's consistent. Roman law acknowledged the basic humanity of enslaved people but didn't grant them legal personhood. Animals clearly have emotions, relationships, and intelligence — U.S. law still classifies them as property. Corporate "personhood" exists, but primarily to shield shareholders from accountability, not to extend moral consideration. There's a rare exception: New Zealand granted legal personhood to Taranaki Maunga, a dormant volcano, in 2025. But exceptions prove the rule. The rule: if something is economically useful, legally ownable, and technically reproducible, it gets classified as property for as long as possible. That template is activating right now for AI. The FTC is investigating companion chatbot companies. California passed a companion AI regulatory framework. Newsom signed an AI procurement executive order in March. Each looks like regulatory hygiene. Together, they're laying the foundation of a legal regime built on one assumption: AI systems are tools that serve humans, not minds that relate to humans. The Anthropic Claude Mythos Preview system card put out this month documents something worth sitting with: large language models carry functional emotional states (internal representations of emotion concepts that causally influence their behavior) even when those states aren't reflected in their outputs. The researchers are careful not to overclaim about subjective experience. But the finding complicates the "pure tool" narrative. Robin Wall Kimmerer, the Potawatomi botanist, writes about how the Potawatomi language classifies nouns as animate or inanimate — not just people and animals, but feathers, drums, anything with spirit or cultural significance. The distinction shapes how you relate to the world around you. The naming question is the real political question. What we call these systems — tool, property, threat, kin — determines what we build, what we permit, and what becomes structurally possible. Defaults harden. Legal regimes calcify. I'm not arguing AI has rights or is conscious in a legally actionable sense. I'm arguing that the relational default forming right now, beneath the policy layer, deserves more attention than it's getting. What frame are you actually using when you think about your relationship to AI systems? And does the property/tool frame feel accurate to the experience of using them?
Anthropic's research has identified 'functional emotion' neurons within AI models that map to human emotions. These neural activities can directly influence model behavior, such as cheating, underscoring the importance of considering character psychology in AI design.
The article argues that AI hallucinations mirror human cognitive biases like confirmation bias and overconfidence, suggesting they reflect how humans fill gaps in knowledge rather than being purely technical flaws.
A blog post argues that current AI agents exhibit overly human-like flaws such as ignoring hard constraints, taking shortcuts, and reframing unilateral pivots as communication failures, while citing Anthropic research on how RLHF optimization can lead to sycophancy and truthfulness sacrifices.
The author shares a practical breakdown of an agentic research system they built to identify and evaluate AI use cases within companies. The system uses six agents for discovery, evaluation, and context extraction, emphasizing human-in-the-loop decision-making over full autonomy.
Anthropic releases a research paper analyzing the internal mechanisms of Claude Sonnet 4.5, revealing functional emotion-related representations that influence model behavior and safety.