Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Thesis/Dissertation AI Safety Researchers,ML Engineers,Researchers in probabilistic ML,Developers of safety-critical AI systems 1 week ago

Epistemic Deep Learning: Enabling Machine Learning Models to Know When They Do Not Know

ai-safety › interpretability
📄 Abstract

Abstract: Machine learning has achieved remarkable successes, yet its deployment in safety-critical domains remains hindered by an inherent inability to manage uncertainty, resulting in overconfident and unreliable predictions when models encounter out-of-distribution data, adversarial perturbations, or naturally fluctuating environments. This thesis, titled Epistemic Deep Learning: Enabling Machine Learning Models to 'Know When They Do Not Know', addresses these critical challenges by advancing the paradigm of Epistemic Artificial Intelligence, which explicitly models and quantifies epistemic uncertainty: the uncertainty arising from limited, biased, or incomplete training data, as opposed to the irreducible randomness of aleatoric uncertainty, thereby empowering models to acknowledge their limitations and refrain from overconfident decisions when uncertainty is high. Central to this work is the development of the Random-Set Neural Network (RS-NN), a novel methodology that leverages random set theory to predict belief functions over sets of classes, capturing the extent of epistemic uncertainty through the width of associated credal sets, applications of RS-NN, including its adaptation to Large Language Models (LLMs) and its deployment in weather classification for autonomous racing. In addition, the thesis proposes a unified evaluation framework for uncertainty-aware classifiers. Extensive experiments validate that integrating epistemic awareness into deep learning not only mitigates the risks associated with overconfident predictions but also lays the foundation for a paradigm shift in artificial intelligence, where the ability to 'know when it does not know' becomes a hallmark of robust and dependable systems. The title encapsulates the core philosophy of this work, emphasizing that true intelligence involves recognizing and managing the limits of one's own knowledge.
Authors (1)
Shireen Kudukkil Manchingal
Submitted
October 25, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

Advances the paradigm of Epistemic Artificial Intelligence by developing the Random-Set Neural Network (RS-NN) to explicitly model and quantify epistemic uncertainty. This enables ML models to 'know when they do not know,' preventing overconfident predictions in safety-critical situations.

Business Value

Crucial for deploying AI in high-stakes applications like healthcare and autonomous driving, where understanding model confidence is paramount for safety and reliability.