Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper AI researchers,NLP engineers,Social media platforms,Fact-checkers,Journalists 2 weeks ago

Misinformation Detection using Large Language Models with Explainability

ai-safety › interpretability
📄 Abstract

Abstract: The rapid spread of misinformation on online platforms undermines trust among individuals and hinders informed decision making. This paper shows an explainable and computationally efficient pipeline to detect misinformation using transformer-based pretrained language models (PLMs). We optimize both RoBERTa and DistilBERT using a two-step strategy: first, we freeze the backbone and train only the classification head; then, we progressively unfreeze the backbone layers while applying layer-wise learning rate decay. On two real-world benchmark datasets, COVID Fake News and FakeNewsNet GossipCop, we test the proposed approach with a unified protocol of preprocessing and stratified splits. To ensure transparency, we integrate the Local Interpretable Model-Agnostic Explanations (LIME) at the token level to present token-level rationales and SHapley Additive exPlanations (SHAP) at the global feature attribution level. It demonstrates that DistilBERT achieves accuracy comparable to RoBERTa while requiring significantly less computational resources. This work makes two key contributions: (1) it quantitatively shows that a lightweight PLM can maintain task performance while substantially reducing computational cost, and (2) it presents an explainable pipeline that retrieves faithful local and global justifications without compromising performance. The results suggest that PLMs combined with principled fine-tuning and interpretability can be an effective framework for scalable, trustworthy misinformation detection.
Authors (4)
Jainee Patel
Chintan Bhatt
Himani Trivedi
Thanh Thi Nguyen
Submitted
October 21, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

This paper presents an explainable and computationally efficient pipeline for misinformation detection using transformer-based PLMs (RoBERTa, DistilBERT). It integrates LIME and SHAP to provide token-level rationales and global feature attributions, enhancing transparency and trust in the detection process.

Business Value

Enables platforms to more effectively combat misinformation with transparent and understandable AI tools, fostering trust and informed decision-making among users.