Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Testing deep reinforcement learning (DRL) agents in safety-critical domains
requires discovering diverse failure scenarios. Existing tools such as INDAGO
rely on single-objective optimization focused solely on maximizing failure
counts, but this does not ensure discovered scenarios are diverse or reveal
distinct error types. We introduce INDAGO-Nexus, a multi-objective search
approach that jointly optimizes for failure likelihood and test scenario
diversity using multi-objective evolutionary algorithms with multiple diversity
metrics and Pareto front selection strategies. We evaluated INDAGO-Nexus on
three DRL agents: humanoid walker, self-driving car, and parking agent. On
average, INDAGO-Nexus discovers up to 83% and 40% more unique failures (test
effectiveness) than INDAGO in the SDC and Parking scenarios, respectively,
while reducing time-to-failure by up to 67% across all agents.
Authors (3)
Antony Bartlett
Cynthia Liem
Annibale Panichella
Submitted
October 16, 2025
Key Contributions
INDAGO-Nexus introduces a multi-objective search approach for testing DRL agents, jointly optimizing for failure likelihood and test scenario diversity using evolutionary algorithms. This method significantly outperforms single-objective approaches like INDAGO in discovering unique failures and reducing time-to-failure.
Business Value
Enhances the safety and reliability of AI systems, particularly in critical domains like autonomous driving and robotics, by providing more comprehensive testing and uncovering subtle failure modes.