Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: What features neural networks learn, and how, remains an open question. In
this paper, we introduce Alternating Gradient Flows (AGF), an algorithmic
framework that describes the dynamics of feature learning in two-layer networks
trained from small initialization. Prior works have shown that gradient flow in
this regime exhibits a staircase-like loss curve, alternating between plateaus
where neurons slowly align to useful directions and sharp drops where neurons
rapidly grow in norm. AGF approximates this behavior as an alternating two-step
process: maximizing a utility function over dormant neurons and minimizing a
cost function over active ones. AGF begins with all neurons dormant. At each
iteration, a dormant neuron activates, triggering the acquisition of a feature
and a drop in the loss. AGF quantifies the order, timing, and magnitude of
these drops, matching experiments across several commonly studied
architectures. We show that AGF unifies and extends existing saddle-to-saddle
analyses in fully connected linear networks and attention-only linear
transformers, where the learned features are singular modes and principal
components, respectively. In diagonal linear networks, we prove AGF converges
to gradient flow in the limit of vanishing initialization. Applying AGF to
quadratic networks trained to perform modular addition, we give the first
complete characterization of the training dynamics, revealing that networks
learn Fourier features in decreasing order of coefficient magnitude.
Altogether, AGF offers a promising step towards understanding feature learning
in neural networks.
Authors (8)
Daniel Kunin
Giovanni Luca Marchetti
Feng Chen
Dhruva Karkada
James B. Simon
Michael R. DeWeese
+2 more
Key Contributions
Introduces Alternating Gradient Flows (AGF), a theoretical framework that models feature learning dynamics in two-layer neural networks. AGF approximates the observed staircase-like loss curve as an alternating process of maximizing utility for dormant neurons and minimizing cost for active ones, quantifying the order, timing, and magnitude of loss drops.
Business Value
Provides a deeper theoretical understanding of neural network learning, which can lead to more interpretable models and potentially more efficient training strategies in the future.