Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper XAI researchers,ML engineers,Data scientists,Developers of interpretable AI systems 2 weeks ago

CONFEX: Uncertainty-Aware Counterfactual Explanations with Conformal Guarantees

ai-safety › interpretability
📄 Abstract

Abstract: Counterfactual explanations (CFXs) provide human-understandable justifications for model predictions, enabling actionable recourse and enhancing interpretability. To be reliable, CFXs must avoid regions of high predictive uncertainty, where explanations may be misleading or inapplicable. However, existing methods often neglect uncertainty or lack principled mechanisms for incorporating it with formal guarantees. We propose CONFEX, a novel method for generating uncertainty-aware counterfactual explanations using Conformal Prediction (CP) and Mixed-Integer Linear Programming (MILP). CONFEX explanations are designed to provide local coverage guarantees, addressing the issue that CFX generation violates exchangeability. To do so, we develop a novel localised CP procedure that enjoys an efficient MILP encoding by leveraging an offline tree-based partitioning of the input space. This way, CONFEX generates CFXs with rigorous guarantees on both predictive uncertainty and optimality. We evaluate CONFEX against state-of-the-art methods across diverse benchmarks and metrics, demonstrating that our uncertainty-aware approach yields robust and plausible explanations.
Authors (4)
Aman Bilkhoo
Mehran Hosseini
Milad Kazemi
Nicola Paoletti
Submitted
October 22, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

CONFEX is a novel method for generating uncertainty-aware counterfactual explanations (CFXs) with formal guarantees. It combines Conformal Prediction (CP) and Mixed-Integer Linear Programming (MILP) to provide local coverage guarantees, addressing the issue of CFX reliability in regions of high predictive uncertainty.

Business Value

Enhances trust and transparency in AI systems by providing reliable, actionable explanations for model predictions, enabling users to understand and influence outcomes.