Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper AI researchers,ML engineers,Data scientists,Developers of interpretable AI systems 2 weeks ago

Towards more holistic interpretability: A lightweight disentangled Concept Bottleneck Model

ai-safety › interpretability
📄 Abstract

Abstract: Concept Bottleneck Models (CBMs) enhance interpretability by predicting human-understandable concepts as intermediate representations. However, existing CBMs often suffer from input-to-concept mapping bias and limited controllability, which restricts their practical value, directly damage the responsibility of strategy from concept-based methods. We propose a lightweight Disentangled Concept Bottleneck Model (LDCBM) that automatically groups visual features into semantically meaningful components without region annotation. By introducing a filter grouping loss and joint concept supervision, our method improves the alignment between visual patterns and concepts, enabling more transparent and robust decision-making. Notably, Experiments on three diverse datasets demonstrate that LDCBM achieves higher concept and class accuracy, outperforming previous CBMs in both interpretability and classification performance. By grounding concepts in visual evidence, our method overcomes a fundamental limitation of prior models and enhances the reliability of interpretable AI.
Authors (3)
Gaoxiang Huang
Songning Lai
Yutao Yue
Submitted
October 17, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

This paper introduces the Lightweight Disentangled Concept Bottleneck Model (LDCBM) to address limitations in existing Concept Bottleneck Models (CBMs), such as input-to-concept bias and limited controllability. LDCBM automatically groups visual features into semantically meaningful components without requiring region annotations, improving the alignment between visual patterns and concepts through a filter grouping loss and joint concept supervision. This leads to more transparent, robust decision-making and superior concept and class accuracy.

Business Value

Enhances trust and accountability in AI systems by providing more transparent and robust decision-making, crucial for applications where understanding the 'why' behind a prediction is critical.