Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
π Abstract
Abstract: Transformer components such as non-linear activations and normalization are
inherently non-injective, suggesting that different inputs could map to the
same output and prevent exact recovery of the input from a model's
representations. In this paper, we challenge this view. First, we prove
mathematically that transformer language models mapping discrete input
sequences to their corresponding sequence of continuous representations are
injective and therefore lossless, a property established at initialization and
preserved during training. Second, we confirm this result empirically through
billions of collision tests on six state-of-the-art language models, and
observe no collisions. Third, we operationalize injectivity: we introduce
SipIt, the first algorithm that provably and efficiently reconstructs the exact
input text from hidden activations, establishing linear-time guarantees and
demonstrating exact invertibility in practice. Overall, our work establishes
injectivity as a fundamental and exploitable property of language models, with
direct implications for transparency, interpretability, and safe deployment.
Authors (6)
Giorgos Nikolaou
Tommaso Mencattini
Donato Crisostomi
Andrea Santilli
Yannis Panagakis
Emanuele RodolΓ
Submitted
October 17, 2025
Key Contributions
This paper mathematically proves that transformer language models are injective and thus lossless, a property preserved during training. It empirically validates this through extensive collision tests and introduces SipIt, an algorithm that efficiently reconstructs the exact input text from hidden activations, demonstrating practical invertibility. This work establishes injectivity as a fundamental property of transformers, challenging prior assumptions about information loss.
Business Value
Understanding the lossless nature of transformers can lead to more reliable information extraction and reconstruction from model representations, potentially improving downstream applications that rely on precise input recovery.