Tag
The author argues that heavily relying on AI coding agents causes human developers to lose critical technical intuition and code review skills over time, proposing measures like mandatory hands-on coding days to maintain supervisory competence.
The author introduces `coding-review-agent-loop`, an open-source local CLI that orchestrates multiple coding agents (Claude Code, Codex, Gemini) to review each other's GitHub PRs using existing local authentication, avoiding additional API costs.
Introduces adamsreview, an open-source Claude Code plugin that enhances pull request reviews using a multi-agent pipeline with parallel sub-agents, validation gates, and an automated fix loop to detect more bugs with fewer false positives.
Cursor 3 introduces a new integrated PR review experience that allows users to manage pull requests from creation to merge within the editor.
The article analyzes how AI agents disrupt traditional code review processes, creating a 'principal-agent problem' where reviewers cannot effectively gauge effort or quality, leading to an increase in low-quality 'slop PRs' in open source.
Hunk is a review-first terminal diff viewer for agent-authored changesets, offering features like multi-file review streams, inline AI annotations, and Git/Jujutsu support.
Simon Willison reflects on how vibe coding and agentic engineering are converging in his own workflow, raising concerns about code review responsibilities as AI coding agents like Claude Code become increasingly reliable. He explores the ethical tension between trusting AI-generated code in production and maintaining software engineering standards.
Claude Code Ultrareview offers cloud-based code review using a fleet of parallel AI agents.
Foil AI Code Scanner is a Mac-native tool that performs AI-powered security reviews of code entirely on-device.
Datadog integrated OpenAI's Codex into their code review process and found it detected 22% of historical incidents that human reviewers missed, demonstrating superior system-level reasoning capabilities compared to traditional static analysis tools.
CodeRabbit launches enhanced code review capabilities using OpenAI's o3, o4-mini, and GPT-4.1 models, enabling developers to ship 4x faster and reduce production bugs by 50%. The tool now includes VS Code integration and uses multi-step reasoning to catch bugs, refactors, and architecture issues across codebases.
OpenAI introduced CriticGPT, a GPT-4-based model designed to catch errors in ChatGPT's code output. When human trainers use CriticGPT for code review, they outperform those without assistance 60% of the time, addressing a fundamental limitation of RLHF as models become increasingly capable.