Published March 2025 | Published
Journal Article

An operator learning perspective on parameter-to-observable maps

  • 1. ROR icon California Institute of Technology

Abstract

Computationally efficient surrogates for parametrized physical models play a crucial role in science and engineering. Operator learning provides data-driven surrogates that map between function spaces. However, instead of full-field measurements, often the available data are only finite-dimensional parametrizations of model inputs or finite observables of model outputs. Building on Fourier Neural Operators, this paper introduces the Fourier Neural Mappings (FNMs) framework that is able to accommodate such finite-dimensional vector inputs or outputs. The paper develops universal approximation theorems for the method. Moreover, in many applications the underlying parameter-to-observable (PtO) map is defined implicitly through an infinite-dimensional operator, such as the solution operator of a partial differential equation. A natural question is whether it is more data-efficient to learn the PtO map end-to-end or first learn the solution operator and subsequently compute the observable from the full-field solution. A theoretical analysis of Bayesian nonparametric regression of linear functionals, which is of independent interest, suggests that the end-to-end approach can actually have worse sample complexity. Extending beyond the theory, numerical results for the FNM approximation of three nonlinear PtO maps demonstrate the benefits of the operator learning perspective that this paper adopts.

Copyright and License

© 2024 American Institute of Mathematical Sciences.

Acknowledgement

The first author is supported by the high-performance computing platform of Peking University. The second author acknowledges support from the National Science Foundation Graduate Research Fellowship Program under award number DGE-1745301 and from the Amazon/Caltech AI4Science Fellowship, and partial support from the Air Force Office of Scientific Research under MURI award number FA9550-20-1-0358 (Machine Learning and Physics-Based Modeling and Simulation). The third author is supported by the Department of Energy Computational Science Graduate Fellowship under award number DE-SC00211. The second and third authors are also grateful for partial support from the Department of Defense Vannevar Bush Faculty Fellowship held by Andrew M. Stuart under Office of Naval Research award number N00014-22-1-2790.

The computations presented in this paper were partially conducted on the Resnick High Performance Computing Center, a facility supported by the Resnick Sustainability Institute at the California Institute of Technology. The authors thank Kaushik Bhattacharya for useful discussions about learning functionals, Andrew Stuart for helpful remarks about the universal approximation theory, and Zachary Morrow for providing the code for the advection–diffusion equation solver. The authors are also grateful for the helpful feedback from two anonymous referees.

Data Availability

Links to datasets and all code used to produce the numerical results and figures in this paper are available at

https://github.com/nickhnelsen/fourier-neural-mappings.

Additional details

Created:
December 17, 2024
Modified:
December 17, 2024