Papers
Topics
Authors
Recent
Search
2000 character limit reached

EXP-Bench: Can AI Conduct AI Research Experiments?

Published 30 May 2025 in cs.AI | (2505.24785v2)

Abstract: Automating AI research holds immense potential for accelerating scientific progress, yet current AI agents struggle with the complexities of rigorous, end-to-end experimentation. We introduce EXP-Bench, a novel benchmark designed to systematically evaluate AI agents on complete research experiments sourced from influential AI publications. Given a research question and incomplete starter code, EXP-Bench challenges AI agents to formulate hypotheses, design and implement experimental procedures, execute them, and analyze results. To enable the creation of such intricate and authentic tasks with high-fidelity, we design a semi-autonomous pipeline to extract and structure crucial experimental details from these research papers and their associated open-source code. With the pipeline, EXP-Bench curated 461 AI research tasks from 51 top-tier AI research papers. Evaluations of leading LLM-based agents, such as OpenHands and IterativeAgent on EXP-Bench demonstrate partial capabilities: while scores on individual experimental aspects such as design or implementation correctness occasionally reach 20-35%, the success rate for complete, executable experiments was a mere 0.5%. By identifying these bottlenecks and providing realistic step-by-step experiment procedures, EXP-Bench serves as a vital tool for future AI agents to improve their ability to conduct AI research experiments. EXP-Bench is open-sourced at https://github.com/Just-Curieous/Curie/tree/main/benchmark/exp_bench.

Summary

  • The paper presents EXP-Bench, a benchmark designed to assess AI agents in executing complete research experiments from hypothesis formulation to result analysis.
  • It employs a semi-automated multi-modal extraction pipeline that curates tasks from peer-reviewed AI publications, ensuring realistic experimental challenges.
  • Evaluation metrics reveal agents score below 30% on key phases, pinpointing critical bottlenecks in design, implementation, and execution.

EXP-Bench: Evaluating AI Agents on AI Research Experiments

The paper "EXP-Bench: Can AI Conduct AI Research Experiments?" (2505.24785) introduces EXP-Bench, a benchmark designed to evaluate the ability of AI agents to conduct end-to-end research experiments. EXP-Bench comprises tasks extracted from peer-reviewed AI publications, challenging agents to formulate hypotheses, design experiments, implement procedures, execute them, and analyze results. The paper highlights the limitations of current AI agents in performing complete, executable experiments, with success rates as low as 0.5%, despite achieving higher scores on individual experimental aspects. EXP-Bench serves as a tool for identifying bottlenecks and guiding future AI agents in improving their ability to automate AI research experiments.

EXP-Bench Design and Construction

EXP-Bench addresses the challenge of automating AI research by providing a structured benchmark for evaluating AI agents on end-to-end research experiments. The benchmark is built upon influential AI research papers and their associated codebases, linking high-level ideas to executable implementations. This allows for the assessment of an agent's ability to conduct established scientific procedures grounded in real-world AI experimentation.

To enable the scalable construction of high-fidelity tasks, the authors developed a semi-automated dataset curation pipeline (Figure 1). This pipeline integrates multi-modal extraction with lightweight human verification. The process begins with source selection and filtering, targeting influential papers from top-tier AI conferences (e.g., NeurIPS, ICLR) that are accompanied by publicly available code repositories. Task extraction then proceeds in two stages: (1) a multi-modal extraction phase that identifies the core elements of the research problem from papers, supplementary materials, and code; and (2) an implementation extraction phase that locates relevant code and assembles scripts to solve the specified task. Finally, all tasks are validated and finalized with execution traces checked against expected outputs. Figure 2

Figure 2: EXP-Bench evaluates AI agents on research experiment tasks extracted semi-autonomously from peer-reviewed AI papers. Given a research question, a high-level method description, and starter code, agents are tasked with designing, implementing, and executing complete experiments. Performance is validated through ground-truth comparisons and implementation execution.

Dataset Composition and Task Specification

The EXP-Bench dataset comprises 461 research tasks derived from 51 papers published at NeurIPS and ICLR 2024 (Figure 3). These tasks span diverse AI subfields such as reinforcement learning, AI applications, and generative models. Each task entry includes a problem statement for the agent and a corresponding ground-truth solution derived from the original research artifacts.

The problem statement provides the agent with: (1) a research question derived from the source paper's experiments; (2) a high-level method description guiding the required experimental approach; and (3) access to the relevant code, potentially with specific components masked or requiring modification. The ground-truth experimental solution comprises: (1) an experimental design specifying key variables, constants, and procedures; (2) the necessary code modifications, assessed via a git diff against the provided repository; and (3) a final conclusion that directly answers the research question based on experimental results. Figure 4

Figure 4: One AI research task example from ICLR 2024 MogaNet [li2024moganetmultiordergatedaggregation].

The multi-metric evaluation pipeline assesses agent performance across all core phases of experimentation—design, implementation, execution, and conclusion. Each metric captures a distinct capability, and their conjunctive use ensures that agents correctly understand and complete the experiment. Figure 3

Figure 3: EXP-Bench's dataset comprises tasks from a diverse set of ML research categories.

Evaluation Metrics and Agent Performance

The evaluation of AI agents on EXP-Bench involves a multi-metric assessment pipeline that evaluates agent performance across design, implementation, execution, and conclusion. The key metrics used are design correctness (D), implementation correctness (I), execution (E), and conclusion correctness (C). The initial evaluations of leading agents reveal that while they often succeed at executing routine procedures, they struggle when tasked with conducting complex experiments.

The evaluation results, shown in Table 1, demonstrate that the agents often score below 30% across all metrics. Specifically, the agents exhibit failures in: (a) conceptualizing and operationalizing sound experimental designs from high-level research questions and methods (16.1% misclassified design variables); (b) translating abstract research methodologies into complete and correct code implementations (39.7% missing essential implementation components); and (c) ensuring the robust and reproducible execution of complex experimental software stacks (29.4% environment or dependency misconfigurations or 23.8% script-level errors).

\subsection{Error Analysis and Failure Patterns}

A detailed analysis of error traces categorizes failures across four key phases of experimentation: implementation, execution, design, and conclusion. The most prevalent issues emerge during the implementation phase, with 39.71% of failures stemming from missing essential components. In the execution phase, failures are most commonly due to environment or dependency misconfigurations (29.38%). Design-related failures also occur frequently, with 16.05% involving incomplete or misclassified experimental variables. Finally, conclusion-phase errors highlight limitations in agents’ interpretive reasoning. Figure 5

Figure 5: Error Analysis and Comprehensive Explanation of the agent's failure to complete the task in Example 1.

Figure 6

Figure 6

Figure 6: Cost–time trade-offs across agents.

\subsection{Implications and Future Directions}

The EXP-Bench benchmark and dataset have implications for the development of AI agents capable of automating AI research. By identifying key bottlenecks in the experimentation process, EXP-Bench helps target specific research components for improvement and advance next-generation AI agents for autonomous research. One promising direction is to apply reinforcement learning with verifiable rewards, enabling agents to autonomously navigate the research lifecycle and accelerate scientific discovery.

Conclusion

The paper introduces EXP-Bench, a benchmark for evaluating AI agents in conducting end-to-end AI research experiments. The benchmark leverages a semi-automated curation pipeline to source tasks from peer-reviewed publications and accompanying codebases. Initial evaluations reveal significant bottlenecks in conceptualizing complex experiments and ensuring robust code implementation and execution. EXP-Bench serves as a tool and dataset to guide future AI agents in automating AI research.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 0 likes about this paper.