Tag
The ByteDance TRAE team has released the '2026 Enterprise AI Programming Practice Manual' and published an internally compiled list of the Top 10 recommended Agent Skills. This list highlights the importance of frontend design, code review, and automated testing, showcasing best practices from a major tech player in the field of AI-assisted programming.
Anthropic engineers share insights on making AI agents succeed in production, highlighting proven patterns from their work on Claude.
The author shares findings on the lack of portability for JSON Schema structured outputs across AI providers like OpenAI, Gemini, and Anthropic, highlighting inconsistencies in constraint enforcement and offering practical advice for robust integration.
The author shares practical insights on building client trust in AI agent systems, emphasizing the importance of narrow scope, robust error handling, and clear communication of system status.
The article argues that high autonomy in AI agents increases the cost of errors, advocating instead for constrained, reliable agents that prioritize safety and predictability over unrestricted capability.
A tweet highlighting key principles for building agent systems, emphasizing scaffolding, memory, and reusable tools, based on an article by Yohei Nakajima.
Akshay Pachaar outlines essential skills for AI engineers beyond prompt engineering, including caching strategies, observability, and cost attribution.
A practitioner shares ten critical lessons for deploying AI agents in production, emphasizing code-based constraints, context management, and security over relying solely on prompts.
The author shares notes and lessons learned from building AI agents at scale, focusing on RAG and memory management to help others.
The author consolidates a series of articles on software testing fundamentals, covering topics such as the purpose of testing, assertions, code coverage, and handling flaky tests.
The author describes improving AI agent reliability by replacing a single general-purpose agent with a four-agent workflow specializing in intake, research, action, and review. This shift prioritized system predictability and easier debugging over raw autonomy.
The author emphasizes the importance of treating AI agents as measurable systems early in development, using evaluations as the primary substrate for improvement and production readiness.
The article details an expanded 12-rule CLAUDE.md configuration template that builds upon Andrej Karpathy's original 4 rules to further reduce AI coding errors and handle complex agent orchestration issues.
The article summarizes a talk by Matt Pocock criticizing 'specs-to-code' approaches, arguing that solid software engineering fundamentals like TDD and modular design are more critical than ever for effectively using AI coding assistants like Claude Code.
The Perplexity team has published guidelines for the design, iteration, and maintenance of Agent Skills, emphasizing that writing Skills is not traditional coding but rather constructing context for the model. The article proposes a counter-intuitive methodology focused on evaluation-first approaches, progressive loading, and optimizing Agent behavior by handling edge cases (Gotchas).
The article shares key insights from a workshop by Boris on using CLAUDE.md for context injection in Claude, highlighting three usage levels, specific commands like /loop, and plan mode to improve developer workflows.
An X thread arguing that production AI agents need operational scaffolding (runbooks, permissions, logs, rollback, verification) rather than just better prompts. The author draws parallels to DevOps evolution, stating that prompts provide advice while runbooks provide control, and that agent systems require platform engineering solutions for permissions, state management, verification, observability, and rollback capabilities.
Bjarne Stroustrup answers common questions about memory leaks in C++, providing guidance on modern C++ memory management techniques.
This article explains how to build a Claude agent using Python, emphasizing the importance of handling tool failure cases effectively rather than just relying on happy-path scenarios.
The article highlights practical system-level failures in AI agent workflows, such as context bleed and hallucinated details, arguing that these are often infrastructure issues rather than model defects.