Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

On the influence of over-parameterization in manifold based surrogates and deep neural operators

Journal Article · · Journal of Computational Physics
 [1];  [2];  [3];  [2]
  1. Johns Hopkins Univ., Baltimore, MD (United States); OSTI
  2. Brown Univ., Providence, RI (United States)
  3. Johns Hopkins Univ., Baltimore, MD (United States)
Constructing accurate and generalizable approximators (surrogate models) for complex physico-chemical processes exhibiting highly non-smooth dynamics is challenging. The main question is what type of surrogate models we should construct and should these models be under-parameterized or over-parameterized. In this work, we propose new developments and perform comparisons for two promising approaches: manifold-based polynomial chaos expansion (m-PCE) and the deep neural operator (DeepONet), and we examine the effect of over-parameterization on generalization. While m-PCE enables the construction of a mapping by first identifying low-dimensional embeddings of the input functions, parameters, and quantities of interest (QoIs), a neural operator learns the nonlinear mapping via the use of deep neural networks. Here, we demonstrate the performance of these methods in terms of generalization accuracy by solving the 2D time-dependent Brusselator reaction-diffusion system with uncertainty sources, modeling an autocatalytic chemical reaction between two species. We first propose an extension of the m-PCE by constructing a mapping between latent spaces formed by two separate embeddings of the input functions and the output QoIs. To further enhance the accuracy of the DeepONet, we introduce weight self-adaptivity in the loss function. We demonstrate that the performance of m-PCE and DeepONet is comparable for cases of relatively smooth input-output mappings. However, when highly non-smooth dynamics is considered, DeepONet shows higher approximation accuracy. We also find that for m-PCE, modest over-parameterization leads to better generalization, both within and outside of distribution, whereas aggressive over-parameterization leads to over-fitting. In contrast, an even highly over-parameterized DeepONet leads to better generalization for both smooth and non-smooth dynamics. Furthermore, we compare the performance of the above models with another recently proposed operator learning model, the Fourier Neural Operator, and show that its over-parameterization also leads to better generalization. Taken together, our studies show that m-PCE can provide very good accuracy at very low training cost, whereas a highly over-parameterized DeepONet can provide better accuracy and robustness to noise but at higher training cost. In both methods, the inference cost is negligible.
Research Organization:
Brown Univ., Providence, RI (United States); Johns Hopkins Univ., Baltimore, MD (United States); Johns Hopkins University, Baltimore, MD (United States)
Sponsoring Organization:
US Air Force Office of Scientific Research (AFOSR); USDOE Office of Science (SC), Advanced Scientific Computing Research (ASCR)
Grant/Contract Number:
SC0019453; SC0020428
OSTI ID:
2421330
Alternate ID(s):
OSTI ID: 2563171
OSTI ID: 1959111
Journal Information:
Journal of Computational Physics, Journal Name: Journal of Computational Physics Journal Issue: C Vol. 479; ISSN 0021-9991
Publisher:
ElsevierCopyright Statement
Country of Publication:
United States
Language:
English

References (37)

Generative Deep Neural Networks for Inverse Materials Design Using Backpropagation and Active Learning journal January 2020
Robust adaptive least squares polynomial chaos expansions in high‐frequency applications
  • Loukrezis, Dimitrios; Galetzka, Armin; De Gersem, Herbert
  • International Journal of Numerical Modelling: Electronic Networks, Devices and Fields, Vol. 33, Issue 6 https://doi.org/10.1002/jnm.2725
journal February 2020
An adaptive polynomial chaos expansion for high-dimensional reliability analysis journal July 2020
Sparse Pseudo Spectral Projection Methods with Directional Adaptation for Uncertainty Quantification journal December 2015
Adaptive multi-element polynomial chaos with discrete measure: Algorithms and application to SPDEs journal April 2015
Sparse pseudospectral approximation method journal July 2012
Least squares polynomial chaos expansion: A review of sampling strategies journal April 2018
Sparse polynomial chaos expansions via compressed sensing and D-optimal design journal July 2018
Data-driven surrogates for high dimensional models using Gaussian process regression on the Grassmann manifold journal October 2020
Bayesian neural networks for uncertainty quantification in data-driven materials modeling journal December 2021
A physics-informed variational DeepONet for predicting crack path in quasi-brittle materials journal March 2022
An adaptive multi-element generalized polynomial chaos method for stochastic differential equations journal November 2005
Adaptive sparse polynomial chaos expansion based on least angle regression journal March 2011
Uncertainty propagation using infinite mixture of Gaussian processes and variational Bayesian inference journal March 2015
Gaussian processes with built-in dimensionality reduction: Applications to high-dimensional uncertainty propagation journal September 2016
Basis adaptive sample efficient polynomial chaos (BASE-PC) journal October 2018
Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations journal February 2019
Compressive sensing adaptation for polynomial chaos expansions journal March 2019
UQpy: A general purpose Python package and development environment for uncertainty quantification journal November 2020
Kernel PCA for novelty detection journal March 2007
Data-driven uncertainty quantification using the arbitrary polynomial chaos expansion journal October 2012
Reliability analysis of structures by iterative improved response surface method journal May 2016
Physics‐Informed Deep Neural Networks for Learning Parameters and Constitutive Relationships in Subsurface Flow Problems journal May 2020
Deep neural networks for the evaluation and design of photonic devices journal December 2020
Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators journal March 2021
Numerical modeling of three dimensional Brusselator reaction diffusion system journal January 2019
Operator learning for predicting multiscale bubble growth dynamics journal March 2021
Reconciling modern machine-learning practice and the classical bias–variance trade-off journal July 2019
Theoretical issues in deep networks journal June 2020
Polynomial Chaos in Stochastic Finite Elements journal March 1990
Time, Structure, and Fluctuations journal September 1978
Multi-Element Generalized Polynomial Chaos for Arbitrary Probability Measures journal January 2006
Adaptive Smolyak Pseudospectral Approximations journal January 2013
Numerical Gaussian Processes for Time-Dependent and Nonlinear Partial Differential Equations journal January 2018
The Wiener--Askey Polynomial Chaos for Stochastic Differential Equations journal January 2002
Physical Systems with Random Uncertainties: Chaos Representations with Arbitrary Probability Measure journal January 2004
Extending Classical Surrogate Modeling to high Dimensions Through Supervised Dimensionality Reduction: a Data-Driven Approach journal January 2020