Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: With the rapid discovery of emergent phenomena in deep learning and large
language models, understanding their cause has become an urgent need. Here, we
propose a rigorous entropic-force theory for understanding the learning
dynamics of neural networks trained with stochastic gradient descent (SGD) and
its variants. Building on the theory of parameter symmetries and an entropic
loss landscape, we show that representation learning is crucially governed by
emergent entropic forces arising from stochasticity and discrete-time updates.
These forces systematically break continuous parameter symmetries and preserve
discrete ones, leading to a series of gradient balance phenomena that resemble
the equipartition property of thermal systems. These phenomena, in turn, (a)
explain the universal alignment of neural representations between AI models and
lead to a proof of the Platonic Representation Hypothesis, and (b) reconcile
the seemingly contradictory observations of sharpness- and flatness-seeking
behavior of deep learning optimization. Our theory and experiments demonstrate
that a combination of entropic forces and symmetry breaking is key to
understanding emergent phenomena in deep learning.
Authors (3)
Liu Ziyin
Yizhou Xu
Isaac Chuang
Key Contributions
This paper proposes a rigorous entropic-force theory to explain the learning dynamics of neural networks trained with SGD. It demonstrates how emergent entropic forces, arising from stochasticity and discrete-time updates, systematically break parameter symmetries and lead to gradient balance phenomena resembling thermal systems. This theory explains universal representation alignment and proves the Platonic Representation Hypothesis, reconciling observations about sharpness and flatness.
Business Value
Provides a fundamental theoretical understanding of how deep learning models learn, which can guide the development of more stable, efficient, and generalizable AI systems. This could lead to better model design and training strategies.