"At what point does adding another agent actually hurt your system? Asking because my 6-agent pipeline is slower and less reliable than my old 2-agent one

Reddit r/AI_Agents News

Summary

A developer shares real-world experiences with AI orchestration frameworks (LangGraph, CrewAI, AutoGen), noting trade-offs between ease of prototyping and production reliability, and asks the community about handling failures, human-in-the-loop, and token costs.

I've been evaluating orchestration frameworks for the past few months and I'm getting tired of benchmark posts and YouTube tutorials that all conveniently end right before deployment. Here's where I landed after actually shipping a few things: **LangGraph** \- solid for stateful workflows where you need explicit control over the graph. The checkpointing is genuinely useful. But the debugging story is rough. When something breaks mid-graph in production, tracing back what state you were in is painful unless you've built your own observability layer on top. **CrewAI** \- great for prototyping fast. Role-based agents feel intuitive to set up. But I hit a wall when I needed anything non-standard. The abstraction that makes it easy early on becomes a ceiling. Also had reliability issues with longer tasks - agents would go off-script in ways that were hard to reproduce. **AutoGen** \- haven't shipped this one, only used it in demos. The conversational multi-agent loop looks impressive but I genuinely don't know how you'd put guardrails around it in a real production environment. Happy to be wrong on this. What I actually use now is a lighter custom setup for anything customer-facing, and LangGraph only when I need durable state across long-running tasks. Curious what others have actually shipped - not what looked good in a notebook. Specifically interested in: 1. How you handle failures mid-workflow? 2. Whether you're using any of these with human-in-the-loop steps 3. Token costs at scale - did the framework choice affect this at all? Thanks in advance
Original Article

Similar Articles

AI agents are changing how people think about compute costs

Reddit r/AI_Agents

The article discusses how AI agent workflows are shifting optimization focus from pure inference costs to broader challenges like latency, orchestration overhead, and reliability. It highlights a trend toward hybrid architectures and dynamic model routing to address these multi-step workflow complexities.

@djfarrelly: https://x.com/djfarrelly/status/2052779234234380479

X AI KOLs Timeline

The article argues that AI agent development should rely on stable execution primitives rather than rigid frameworks, which frequently change with emerging orchestration patterns. It emphasizes durable steps, persistent state, parallel coordination, event-driven flow, and observability to prevent costly rewrites as best practices evolve.

All your agents are going async

Hacker News Top

The article argues that AI agents are shifting from synchronous chat interfaces to asynchronous background workflows, highlighting new features from Anthropic, OpenAI, and Cursor that decouple agent lifetimes from HTTP request-response cycles.

Building effective agents

Anthropic Engineering

Anthropic publishes engineering guidelines for building effective AI agents, advocating for simple, composable patterns and direct API usage over complex frameworks. The article distinguishes between workflows and autonomous agents, providing practical advice on when to use each architecture.