Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
π Abstract
Abstract: We present modal aphasia, a systematic dissociation in which current unified
multimodal models accurately memorize concepts visually but fail to articulate
them in writing, despite being trained on images and text simultaneously. For
one, we show that leading frontier models can generate near-perfect
reproductions of iconic movie artwork, but confuse crucial details when asked
for textual descriptions. We corroborate those findings through controlled
experiments on synthetic datasets in multiple architectures. Our experiments
confirm that modal aphasia reliably emerges as a fundamental property of
current unified multimodal models, not just as a training artifact. In
practice, modal aphasia can introduce vulnerabilities in AI safety frameworks,
as safeguards applied to one modality may leave harmful concepts accessible in
other modalities. We demonstrate this risk by showing how a model aligned
solely on text remains capable of generating unsafe images.
Authors (4)
Michael Aerni
Joshua Swanson
Kristina NikoliΔ
Florian Tramèr
Submitted
October 22, 2025
Key Contributions
This paper introduces 'modal aphasia,' a phenomenon where unified multimodal models can visually memorize concepts but fail to articulate them in text, despite joint training. It highlights this as a fundamental property, not a training artifact, and demonstrates how it creates vulnerabilities in AI safety by allowing harmful concepts to be accessible through one modality even if aligned on another.
Business Value
Crucial for developing more robust and safer AI systems by understanding and mitigating potential cross-modal vulnerabilities, preventing misuse of generative capabilities.