Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 75% Match Research Paper Researchers in scientific machine learning,Applied mathematicians,Engineers using simulations,Deep learning theorists 1 week ago

A Deep Learning Framework for Multi-Operator Learning: Architectures and Approximation Theory

generative-ai › flow-models
📄 Abstract

Abstract: While many problems in machine learning focus on learning mappings between finite-dimensional spaces, scientific applications require approximating mappings between function spaces, i.e., operators. We study the problem of learning collections of operators and provide both theoretical and empirical advances. We distinguish between two regimes: (i) multiple operator learning, where a single network represents a continuum of operators parameterized by a parametric function, and (ii) learning several distinct single operators, where each operator is learned independently. For the multiple operator case, we introduce two new architectures, $\mathrm{MNO}$ and $\mathrm{MONet}$, and establish universal approximation results in three settings: continuous, integrable, or Lipschitz operators. For the latter, we further derive explicit scaling laws that quantify how the network size must grow to achieve a target approximation accuracy. For learning several single operators, we develop a framework for balancing architectural complexity across subnetworks and show how approximation order determines computational efficiency. Empirical experiments on parametric PDE benchmarks confirm the strong expressive power and efficiency of the proposed architectures. Overall, this work establishes a unified theoretical and practical foundation for scalable neural operator learning across multiple operators.
Authors (4)
Adrien Weihs
Jingmin Sun
Zecheng Zhang
Hayden Schaeffer
Submitted
October 29, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper introduces new neural operator architectures (MNO, MONet) for learning mappings between function spaces, providing theoretical guarantees (universal approximation) for continuous, integrable, and Lipschitz operators. It also derives explicit scaling laws for Lipschitz operators, quantifying network size requirements for a target accuracy.

Business Value

Enables faster and more accurate simulations in scientific and engineering fields by learning complex physical processes, potentially reducing the need for expensive traditional simulations and accelerating R&D.