Tag
Anthropic's Head of Product released a free 28-minute masterclass on putting AI agents into production, covering prompt caching, tool search, programmatic tool calling, compaction, and advisor strategy.
A tutorial by Vasco Schiavo explaining the math behind the cost of AI agents, focusing on why agents can be expensive and the importance of prompt caching.
A new optimization technique for open-source RL training engines introduces prompt caching during training, achieving up to 7.5x speedup on long-prompt, short-response workloads by reducing redundant compute.
OpenClaw is a CLI tool that supports Anthropic Claude models via API key or Claude CLI reuse, with features including adaptive thinking defaults for Claude 4.6, fast mode service tier toggling, and configurable prompt caching. Anthropic has reportedly re-allowed OpenClaw-style Claude CLI usage.
OpenAI introduces Prompt Caching, an automatic feature that reduces API costs by 50% and improves latency by reusing recently cached input tokens on GPT-4o, GPT-4o mini, o1-preview, and o1-mini models. The feature automatically applies to prompts longer than 1,024 tokens without requiring developer integration changes.