Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper AI researchers,LLM developers,Agent developers,Benchmark creators 1 week ago

Automating Benchmark Design

large-language-models › evaluation
📄 Abstract

Abstract: The rapid progress and widespread deployment of LLMs and LLM-powered agents has outpaced our ability to evaluate them. Hand-crafted, static benchmarks are the primary tool for assessing model capabilities, but these quickly become saturated. In contrast, dynamic benchmarks evolve alongside the models they evaluate, but are expensive to create and continuously update. To address these challenges, we develop BeTaL (Benchmark Tuning with an LLM-in-the-loop), a framework that leverages environment design principles to automate the process of dynamic benchmark design. BeTaL works by parameterizing key design choices in base benchmark templates and uses LLMs to reason through the resulting parameter space to obtain target properties (such as difficulty and realism) in a cost-efficient manner. We validate this approach on its ability to create benchmarks with desired difficulty levels. Using BeTaL, we create two new benchmarks and extend a popular agentic benchmark $\tau$-bench. Extensive evaluation on these three tasks and multiple target difficulty levels shows that BeTaL produces benchmarks much closer to the desired difficulty, with average deviations ranging from 5.3% to 13.2% -- a 2-4x improvement over the baselines.
Authors (9)
Amanda Dsouza
Harit Vishwakarma
Zhengyang Qi
Justin Bauer
Derek Pham
Thomas Walshe
+3 more
Submitted
October 28, 2025
arXiv Category
cs.SE
arXiv PDF

Key Contributions

Develops BeTaL, a framework that automates dynamic benchmark design using LLMs. By parameterizing benchmark templates and leveraging LLM reasoning, BeTaL efficiently creates benchmarks with target properties like difficulty and realism, addressing the saturation of static benchmarks and the cost of manual dynamic ones.

Business Value

Accelerates the development and deployment of reliable LLMs and agents by providing continuously evolving and relevant evaluation tools, reducing the time and cost associated with benchmark creation.