Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Counterfactual explanations (CFXs) provide human-understandable
justifications for model predictions, enabling actionable recourse and
enhancing interpretability. To be reliable, CFXs must avoid regions of high
predictive uncertainty, where explanations may be misleading or inapplicable.
However, existing methods often neglect uncertainty or lack principled
mechanisms for incorporating it with formal guarantees. We propose CONFEX, a
novel method for generating uncertainty-aware counterfactual explanations using
Conformal Prediction (CP) and Mixed-Integer Linear Programming (MILP). CONFEX
explanations are designed to provide local coverage guarantees, addressing the
issue that CFX generation violates exchangeability. To do so, we develop a
novel localised CP procedure that enjoys an efficient MILP encoding by
leveraging an offline tree-based partitioning of the input space. This way,
CONFEX generates CFXs with rigorous guarantees on both predictive uncertainty
and optimality. We evaluate CONFEX against state-of-the-art methods across
diverse benchmarks and metrics, demonstrating that our uncertainty-aware
approach yields robust and plausible explanations.
Authors (4)
Aman Bilkhoo
Mehran Hosseini
Milad Kazemi
Nicola Paoletti
Submitted
October 22, 2025
Key Contributions
CONFEX is a novel method for generating uncertainty-aware counterfactual explanations (CFXs) with formal guarantees. It combines Conformal Prediction (CP) and Mixed-Integer Linear Programming (MILP) to provide local coverage guarantees, addressing the issue of CFX reliability in regions of high predictive uncertainty.
Business Value
Enhances trust and transparency in AI systems by providing reliable, actionable explanations for model predictions, enabling users to understand and influence outcomes.