Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 92% Match Theoretical Research Paper Theoretical Computer Scientists,ML Theorists,Researchers in Formal Methods 2 weeks ago

Transformers are Inherently Succinct

large-language-models β€Ί model-architecture
πŸ“„ Abstract

Abstract: We propose succinctness as a measure of the expressive power of a transformer in describing a concept. To this end, we prove that transformers are highly expressive in that they can represent formal languages substantially more succinctly than standard representations of formal languages like finite automata and Linear Temporal Logic (LTL) formulas. As a by-product of this expressivity, we show that verifying properties of transformers is provably intractable (i.e. EXPSPACE-complete).
Authors (3)
Pascal BergstrÀßer
Ryan Cotterell
Anthony W. Lin
Submitted
October 22, 2025
arXiv Category
cs.FL
arXiv PDF

Key Contributions

This paper introduces 'succinctness' as a measure of a transformer's expressive power and proves that transformers can represent formal languages significantly more succinctly than standard representations like finite automata or LTL formulas. As a consequence, it shows that verifying properties of transformers is provably intractable (EXPSPACE-complete).

Business Value

Provides a foundational understanding of the theoretical capabilities and limitations of transformer models, which can inform the design of more efficient architectures and guide expectations about their performance in complex reasoning tasks.