Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
This paper presents the first comprehensive study of Chain-of-Thought (CoT) faithfulness in Large Vision-Language Models (LVLMs). It introduces a novel evaluation pipeline to categorize bias articulation patterns, revealing that image-based biases are rarely articulated compared to text-based ones, and identifying a new phenomenon in model reasoning.
Improves the reliability and trustworthiness of multimodal AI systems by providing deeper insights into their reasoning processes and biases. This is crucial for applications where accurate and unbiased visual understanding is required.