Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: The rapid progress and widespread deployment of LLMs and LLM-powered agents
has outpaced our ability to evaluate them. Hand-crafted, static benchmarks are
the primary tool for assessing model capabilities, but these quickly become
saturated. In contrast, dynamic benchmarks evolve alongside the models they
evaluate, but are expensive to create and continuously update. To address these
challenges, we develop BeTaL (Benchmark Tuning with an LLM-in-the-loop), a
framework that leverages environment design principles to automate the process
of dynamic benchmark design. BeTaL works by parameterizing key design choices
in base benchmark templates and uses LLMs to reason through the resulting
parameter space to obtain target properties (such as difficulty and realism) in
a cost-efficient manner. We validate this approach on its ability to create
benchmarks with desired difficulty levels. Using BeTaL, we create two new
benchmarks and extend a popular agentic benchmark $\tau$-bench. Extensive
evaluation on these three tasks and multiple target difficulty levels shows
that BeTaL produces benchmarks much closer to the desired difficulty, with
average deviations ranging from 5.3% to 13.2% -- a 2-4x improvement over the
baselines.
Authors (9)
Amanda Dsouza
Harit Vishwakarma
Zhengyang Qi
Justin Bauer
Derek Pham
Thomas Walshe
+3 more
Submitted
October 28, 2025
Key Contributions
Develops BeTaL, a framework that automates dynamic benchmark design using LLMs. By parameterizing benchmark templates and leveraging LLM reasoning, BeTaL efficiently creates benchmarks with target properties like difficulty and realism, addressing the saturation of static benchmarks and the cost of manual dynamic ones.
Business Value
Accelerates the development and deployment of reliable LLMs and agents by providing continuously evolving and relevant evaluation tools, reducing the time and cost associated with benchmark creation.