Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: While many problems in machine learning focus on learning mappings between
finite-dimensional spaces, scientific applications require approximating
mappings between function spaces, i.e., operators. We study the problem of
learning collections of operators and provide both theoretical and empirical
advances. We distinguish between two regimes: (i) multiple operator learning,
where a single network represents a continuum of operators parameterized by a
parametric function, and (ii) learning several distinct single operators, where
each operator is learned independently. For the multiple operator case, we
introduce two new architectures, $\mathrm{MNO}$ and $\mathrm{MONet}$, and
establish universal approximation results in three settings: continuous,
integrable, or Lipschitz operators. For the latter, we further derive explicit
scaling laws that quantify how the network size must grow to achieve a target
approximation accuracy. For learning several single operators, we develop a
framework for balancing architectural complexity across subnetworks and show
how approximation order determines computational efficiency. Empirical
experiments on parametric PDE benchmarks confirm the strong expressive power
and efficiency of the proposed architectures. Overall, this work establishes a
unified theoretical and practical foundation for scalable neural operator
learning across multiple operators.
Authors (4)
Adrien Weihs
Jingmin Sun
Zecheng Zhang
Hayden Schaeffer
Submitted
October 29, 2025
Key Contributions
This paper introduces new neural operator architectures (MNO, MONet) for learning mappings between function spaces, providing theoretical guarantees (universal approximation) for continuous, integrable, and Lipschitz operators. It also derives explicit scaling laws for Lipschitz operators, quantifying network size requirements for a target accuracy.
Business Value
Enables faster and more accurate simulations in scientific and engineering fields by learning complex physical processes, potentially reducing the need for expensive traditional simulations and accelerating R&D.