Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper Machine Learning Researchers,Deep Learning Theorists,NLP Engineers 2 weeks ago

Language Models are Injective and Hence Invertible

large-language-models β€Ί model-architecture
πŸ“„ Abstract

Abstract: Transformer components such as non-linear activations and normalization are inherently non-injective, suggesting that different inputs could map to the same output and prevent exact recovery of the input from a model's representations. In this paper, we challenge this view. First, we prove mathematically that transformer language models mapping discrete input sequences to their corresponding sequence of continuous representations are injective and therefore lossless, a property established at initialization and preserved during training. Second, we confirm this result empirically through billions of collision tests on six state-of-the-art language models, and observe no collisions. Third, we operationalize injectivity: we introduce SipIt, the first algorithm that provably and efficiently reconstructs the exact input text from hidden activations, establishing linear-time guarantees and demonstrating exact invertibility in practice. Overall, our work establishes injectivity as a fundamental and exploitable property of language models, with direct implications for transparency, interpretability, and safe deployment.
Authors (6)
Giorgos Nikolaou
Tommaso Mencattini
Donato Crisostomi
Andrea Santilli
Yannis Panagakis
Emanuele RodolΓ 
Submitted
October 17, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper mathematically proves that transformer language models are injective and thus lossless, a property preserved during training. It empirically validates this through extensive collision tests and introduces SipIt, an algorithm that efficiently reconstructs the exact input text from hidden activations, demonstrating practical invertibility. This work establishes injectivity as a fundamental property of transformers, challenging prior assumptions about information loss.

Business Value

Understanding the lossless nature of transformers can lead to more reliable information extraction and reconstruction from model representations, potentially improving downstream applications that rely on precise input recovery.