Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Representation learning for high-dimensional, complex physical systems aims
to identify a low-dimensional intrinsic latent space, which is crucial for
reduced-order modeling and modal analysis. To overcome the well-known
Kolmogorov barrier, deep autoencoders (AEs) have been introduced in recent
years, but they often suffer from poor convergence behavior as the rank of the
latent space increases. To address this issue, we propose the learnable
weighted hybrid autoencoder, a hybrid approach that combines the strengths of
singular value decomposition (SVD) with deep autoencoders through a learnable
weighted framework. We find that the introduction of learnable weighting
parameters is essential -- without them, the resulting model would either
collapse into a standard POD or fail to exhibit the desired convergence
behavior. Interestingly, we empirically find that our trained model has a
sharpness thousands of times smaller compared to other models. Our experiments
on classical chaotic PDE systems, including the 1D Kuramoto-Sivashinsky and
forced isotropic turbulence datasets, demonstrate that our approach
significantly improves generalization performance compared to several competing
methods. Additionally, when combining with time series modeling techniques
(e.g., Koopman operator, LSTM), the proposed technique offers significant
improvements for surrogate modeling of high-dimensional multi-scale PDE
systems.