Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Computationally resolving turbulence remains a central challenge in fluid
dynamics due to its multi-scale interactions. Fully resolving large-scale
turbulence through direct numerical simulation (DNS) is computationally
prohibitive, motivating data-driven machine learning alternatives. In this
work, we propose EddyFormer, a Transformer-based spectral-element (SEM)
architecture for large-scale turbulence simulation that combines the accuracy
of spectral methods with the scalability of the attention mechanism. We
introduce an SEM tokenization that decomposes the flow into grid-scale and
subgrid-scale components, enabling capture of both local and global features.
We create a new three-dimensional isotropic turbulence dataset and train
EddyFormer to achieves DNS-level accuracy at 256^3 resolution, providing a 30x
speedup over DNS. When applied to unseen domains up to 4x larger than in
training, EddyFormer preserves accuracy on physics-invariant metrics-energy
spectra, correlation functions, and structure functions-showing domain
generalization. On The Well benchmark suite of diverse turbulent flows,
EddyFormer resolves cases where prior ML models fail to converge, accurately
reproducing complex dynamics across a wide range of physical conditions.
Authors (2)
Yiheng Du
Aditi S. Krishnapriyan
Submitted
October 28, 2025
Key Contributions
EddyFormer is a Transformer-based spectral-element architecture that achieves DNS-level accuracy for large-scale 3D turbulence simulation at a 30x speedup over traditional methods. It uses SEM tokenization to capture both local and global flow features, enabling accurate simulations on domains up to 4x larger than training data.
Business Value
Significantly accelerates complex fluid dynamics simulations, enabling faster design cycles and more accurate predictions in industries like aerospace, automotive, and energy.