AJ-Bench: Benchmarking Agent-as-a-Judge for Environment-Aware Evaluation
Summary
AJ-Bench introduces a benchmark to evaluate Agent-as-a-Judge systems that interact with environments to verify agent behaviors across 155 tasks in search, data systems, and GUI domains.
View Cached Full Text
Cached at: 04/22/26, 10:35 AM
Paper page - AJ-Bench: Benchmarking Agent-as-a-Judge for Environment-Aware Evaluation
Source: https://huggingface.co/papers/2604.18240 Authors:
,
,
,
,
,
,
,
,
,
Abstract
Agent-as-a-Judge benchmark evaluates automated verification capabilities across multiple domains with comprehensive task assessment.
Asreinforcement learningcontinues to scale the training of large language model-based agents, reliably verifying agent behaviors in complex environments has become increasingly challenging. Existing approaches rely on rule-based verifiers orLLM-as-a-Judgemodels, which struggle to generalize beyond narrow domains.Agent-as-a-Judgeaddresses this limitation by actively interacting with environments and tools to acquire verifiable evidence, yet its capabilities remain underexplored. We introduce a benchmarkAJ-Benchto systematically evaluateAgent-as-a-Judgeacross three domains-search, data systems, and graphical user interfaces-comprising 155 tasks and 516 annotated trajectories. The benchmark comprehensively assesses judge agents’ abilities ininformation acquisition,state verification, andprocess verification. Experiments demonstrate consistent performance gains overLLM-as-a-Judgebaselines, while also revealing substantial open challenges inagent-based verification. Our data and code are available at https://aj-bench.github.io/.
View arXiv pageView PDFProject pageGitHub0Add to collection
Get this paper in your agent:
hf papers read 2604\.18240
Don’t have the latest CLI?curl \-LsSf https://hf\.co/cli/install\.sh \| bash
Models citing this paper0
No model linking this paper
Cite arxiv.org/abs/2604.18240 in a model README.md to link it from this page.
Datasets citing this paper0
No dataset linking this paper
Cite arxiv.org/abs/2604.18240 in a dataset README.md to link it from this page.
Spaces citing this paper0
No Space linking this paper
Cite arxiv.org/abs/2604.18240 in a Space README.md to link it from this page.
Collections including this paper0
No Collection including this paper
Add this paper to acollectionto link it from this page.
Similar Articles
ProgramBench (5 minute read)
ProgramBench is a new benchmark that evaluates AI agents' ability to reconstruct complete software projects from compiled binaries and documentation without access to source code or decompilation tools.
Partial Evidence Bench: Benchmarking Authorization-Limited Evidence in Agentic Systems
This paper introduces Partial-Evidence-Bench, a deterministic benchmark for measuring 'authorization-limited evidence' failures in agentic AI systems. It evaluates how models handle tasks where access control restricts visibility, assessing their ability to recognize and report incomplete information rather than silently producing seemingly complete but incomplete answers.
MLE-bench: Evaluating Machine Learning Agents on Machine Learning Engineering
OpenAI introduces MLE-bench, a benchmark of 75 Kaggle ML competitions to evaluate AI agents on real-world ML engineering tasks. The best setup, o1-preview with AIDE scaffolding, achieves at least a Kaggle bronze medal in 16.9% of competitions.
SWE-WebDevBench: Evaluating Coding Agent Application Platforms as Virtual Software Agencies
This paper introduces SWE-WebDevBench, a comprehensive 68-metric framework for evaluating AI-powered application development platforms as virtual software agencies. The study highlights critical gaps in current platforms regarding specification understanding, backend reliability, production readiness, and security.
I built a benchmark for AI “memory” in coding agents. looking for others to beat it.
Developer created a new benchmark called continuity-benchmarks to test AI coding agents' ability to maintain consistency with project rules during active development, addressing gaps in existing memory benchmarks that focus on semantic recall rather than real-time architectural consistency and multi-session behavior.