Agent memory is not just RAG over user facts
Summary
The article argues that simple RAG-based agent memory systems fail in production due to issues like stale preferences, missed keywords, and prompt injection, and advocates for a layered memory architecture with active selection, deterministic fallback, governance, and testing.
Similar Articles
Some notes and lessons on Agents, RAG and memory
The author shares notes and lessons learned from building AI agents at scale, focusing on RAG and memory management to help others.
rohitg00/agentmemory
agentmemory is an open-source persistent memory layer for AI coding agents (Claude Code, Cursor, Gemini CLI, Codex CLI, etc.) that uses knowledge graphs, confidence scoring, and hybrid search to give agents long-term memory across sessions via MCP, hooks, or REST API. Built on the iii engine, it requires no external databases and exposes 51 MCP tools.
Agentmemory
Agentmemory offers persistent memory for AI models such as Codex, Hermes, OpenClaw, and Claude, allowing them to maintain long-term context across interactions.
How AI agent memory works (28 minute read)
The article provides a comprehensive technical overview of how AI agent memory works, distinguishing between working and long-term memory mechanisms, and discussing strategies for context management, embedding-based retrieval, and data lifecycle governance.
Building Agentic GraphRAG Systems: From knowledge graphs and ontologies to a unified memory as an MCP server for your AI agent.
The author argues that GraphRAG is fundamentally a data modeling problem rather than just a retrieval algorithm, proposing a five-component architecture using ontologies, knowledge graphs, and an MCP server for unified agent memory.