Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 95% Match Research Paper AI Safety Researchers,ML Engineers,Developers of LLM applications,Researchers in Finance and Healthcare AI 2 weeks ago

Confidence-Based Response Abstinence: Improving LLM Trustworthiness via Activation-Based Uncertainty Estimation

large-language-models › alignment
📄 Abstract

Abstract: We propose a method for confidence estimation in retrieval-augmented generation (RAG) systems that aligns closely with the correctness of large language model (LLM) outputs. Confidence estimation is especially critical in high-stakes domains such as finance and healthcare, where the cost of an incorrect answer outweighs that of not answering the question. Our approach extends prior uncertainty quantification methods by leveraging raw feed-forward network (FFN) activations as auto-regressive signals, avoiding the information loss inherent in token logits and probabilities after projection and softmax normalization. We model confidence prediction as a sequence classification task, and regularize training with a Huber loss term to improve robustness against noisy supervision. Applied in a real-world financial industry customer-support setting with complex knowledge bases, our method outperforms strong baselines and maintains high accuracy under strict latency constraints. Experiments on Llama 3.1 8B model show that using activations from only the 16th layer preserves accuracy while reducing response latency. Our results demonstrate that activation-based confidence modeling offers a scalable, architecture-aware path toward trustworthy RAG deployment.

Key Contributions

Proposes a novel confidence estimation method for RAG systems using raw FFN activations, which aligns better with LLM output correctness than traditional methods. It models confidence as a sequence classification task with Huber loss regularization, enabling response abstinence and outperforming baselines in a real-world financial setting under strict latency constraints.

Business Value

Enhances the safety and reliability of LLMs in critical applications like finance and healthcare by enabling them to abstain from answering when uncertain, preventing potentially harmful incorrect responses. This builds user trust and reduces risk.