Efficient hybrid explicit-implicit learning for multiscale problems
Journal Article
·
· Journal of Computational Physics
- Texas A & M Univ., College Station, TX (United States); North-Eastern Federal University, Yakutsk (Russian Federation); OSTI
- Univ. of California, Irvine, CA (United States)
- Purdue Univ., West Lafayette, IN (United States)
Splitting method is a powerful method to handle application problems by splitting physics, scales, domain, and so on. Many splitting algorithms have been designed for efficient temporal discretization. Here, in this paper, our goal is to use temporal splitting concepts in designing machine learning algorithms and, at the same time, help splitting algorithms by incorporating data and speeding them up. We propose a machine learning assisted splitting scheme which improves the efficiency of the scheme meanwhile preserves the accuracy. We consider a recently introduced multiscale splitting algorithms, where the multiscale problem is solved on a coarse grid. To approximate the dynamics, only a few degrees of freedom are solved implicitly, while others explicitly. This splitting concept allows identifying degrees of freedom that need implicit treatment. In this paper, we use this splitting concept in machine learning and propose several strategies. First, the implicit part of the solution can be learned as it is more difficult to solve, while the explicit part can be computed. This provides a speed-up and data incorporation for splitting approaches. Secondly, one can design a hybrid neural network architecture because handling explicit parts requires much fewer communications among neurons and can be done efficiently. Thirdly, one can solve the coarse grid component via PDEs or other approximation methods and construct simpler neural networks for the explicit part of the solutions. We discuss these options and implement one of them by interpreting it as a machine translation task. This interpretation of the splitting scheme successfully enables us using the Transformer since it can perform model reduction for multiple time series and learn the connection between them. We also find that the splitting scheme is a great platform to predict the coarse solution with insufficient information of the target model: the target problem is partially given and we need to solve it through a known problem which approximates the target. Our machine learning model can incorporate and encode the given information from two different problems and then solve the target problems. We conduct four numerical examples and the results show that our method is stable and accurate.
- Research Organization:
- Brookhaven National Laboratory (BNL), Upton, NY (United States); Purdue Univ., West Lafayette, IN (United States)
- Sponsoring Organization:
- National Science Foundation (NSF); USDOE Office of Science (SC), Advanced Scientific Computing Research (ASCR)
- Grant/Contract Number:
- SC0021142
- OSTI ID:
- 2421758
- Alternate ID(s):
- OSTI ID: 1961085
- Journal Information:
- Journal of Computational Physics, Journal Name: Journal of Computational Physics Journal Issue: C Vol. 467; ISSN 0021-9991
- Publisher:
- ElsevierCopyright Statement
- Country of Publication:
- United States
- Language:
- English
Similar Records
Explicit physics-informed neural networks for nonlinear closure: The case of transport in tissues
Outsmarting neural networks: an alternative paradigm for machine learning
NH-PINN: Neural homogenization-based physics-informed neural network for multiscale problems
Journal Article
·
Thu Oct 28 20:00:00 EDT 2021
· Journal of Computational Physics
·
OSTI ID:1977273
Outsmarting neural networks: an alternative paradigm for machine learning
Conference
·
Tue Oct 01 00:00:00 EDT 1996
·
OSTI ID:393293
NH-PINN: Neural homogenization-based physics-informed neural network for multiscale problems
Journal Article
·
Fri Aug 26 20:00:00 EDT 2022
· Journal of Computational Physics
·
OSTI ID:2421760