Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 95% Match Research Paper AI Safety Researchers,ML Engineers,AI Ethicists,Researchers in Multimodal AI 1 week ago

Modal Aphasia: Can Unified Multimodal Models Describe Images From Memory?

large-language-models β€Ί multimodal-llms
πŸ“„ Abstract

Abstract: We present modal aphasia, a systematic dissociation in which current unified multimodal models accurately memorize concepts visually but fail to articulate them in writing, despite being trained on images and text simultaneously. For one, we show that leading frontier models can generate near-perfect reproductions of iconic movie artwork, but confuse crucial details when asked for textual descriptions. We corroborate those findings through controlled experiments on synthetic datasets in multiple architectures. Our experiments confirm that modal aphasia reliably emerges as a fundamental property of current unified multimodal models, not just as a training artifact. In practice, modal aphasia can introduce vulnerabilities in AI safety frameworks, as safeguards applied to one modality may leave harmful concepts accessible in other modalities. We demonstrate this risk by showing how a model aligned solely on text remains capable of generating unsafe images.
Authors (4)
Michael Aerni
Joshua Swanson
Kristina Nikolić
Florian Tramèr
Submitted
October 22, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

This paper introduces 'modal aphasia,' a phenomenon where unified multimodal models can visually memorize concepts but fail to articulate them in text, despite joint training. It highlights this as a fundamental property, not a training artifact, and demonstrates how it creates vulnerabilities in AI safety by allowing harmful concepts to be accessible through one modality even if aligned on another.

Business Value

Crucial for developing more robust and safer AI systems by understanding and mitigating potential cross-modal vulnerabilities, preventing misuse of generative capabilities.