Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 95% Match Research Paper AI researchers,Medical professionals,AI ethicists,System designers 3 days ago

Learning to Seek Evidence: A Verifiable Reasoning Agent with Causal Faithfulness Analysis

ai-safety › interpretability
📄 Abstract

Abstract: Explanations for AI models in high-stakes domains like medicine often lack verifiability, which can hinder trust. To address this, we propose an interactive agent that produces explanations through an auditable sequence of actions. The agent learns a policy to strategically seek external visual evidence to support its diagnostic reasoning. This policy is optimized using reinforcement learning, resulting in a model that is both efficient and generalizable. Our experiments show that this action-based reasoning process significantly improves calibrated accuracy, reducing the Brier score by 18\% compared to a non-interactive baseline. To validate the faithfulness of the agent's explanations, we introduce a causal intervention method. By masking the visual evidence the agent chooses to use, we observe a measurable degradation in its performance ($\Delta$Brier=+0.029), confirming that the evidence is integral to its decision-making process. Our work provides a practical framework for building AI systems with verifiable and faithful reasoning capabilities.
Authors (4)
Yuhang Huang
Zekai Lin
Fan Zhong
Lei Liu
Submitted
November 3, 2025
arXiv Category
cs.AI
arXiv PDF

Key Contributions

This paper introduces an interactive agent that generates verifiable explanations for AI models in high-stakes domains by strategically seeking external visual evidence. This approach, optimized via reinforcement learning, significantly improves calibrated accuracy and provides a causal intervention method to validate the faithfulness of the agent's reasoning, addressing the critical need for trust and auditability in AI decision-making.

Business Value

Enhances trust and reliability in AI systems used in critical applications like healthcare, leading to better decision-making and potentially reducing errors and liability.