Derivative-Informed Neural Operators for PDE-Constrained Optimization under Uncertainty

Release Time:2024-06-05Number of visits:10

Speaker:  Peng Chen, Georgia Institute of Technology.

Time:       10:00 am, Jun. 14th

Location: SIST2 415

Host:       Qifeng Liao

Abstract:

In this talk I will present a novel machine learning framework for solving optimization problems governed by large-scale partial differential equations (PDEs) with high-dimensional random parameters. Such optimization problems can be found in Bayesian inverse problems for parameter estimation, optimal experimental design for data acquisition, and stochastic optimization for risk-averse optimal control and design. These problems are computationally prohibitive using classical methods, as the estimation of statistical measures may require many solutions of an expensive-to-solve PDE at every iteration of a sampling or optimization algorithm. To address this challenge, we will present a class of Derivative-Informed Neural Operators (DINO) with the combined merits of (1) being able to accurately approximate not only the mapping from the inputs of random parameters and optimization variables to the PDE state, but also its derivative with respect to the input variables, (2) using a reduced basis architecture that can be efficiently constructed and is scalable to high-dimensional problems, and (3) requiring only a limited number of training data to achieve high accuracy for both the PDE solution and the optimization solution. I will present some applications in material science, computational fluid dynamics, and structure mechanics.

This talk is based on the following papers:

Derivative-Informed Neural Operator: An Efficient Framework for High-dimensional Parametric Derivative Learning

https://www.sciencedirect.com/science/article/pii/S0021999123006502

Efficient PDE-Constrained Optimization under High-dimensional Uncertainty Using Derivative-Informed Neural Operators

https://arxiv.org/abs/2305.20053

Accelerating Bayesian Optimal Experimental Design with Derivative-Informed Neural Operators

https://arxiv.org/abs/2312.14810

Bio:

Peng Chen is a tenure-track assistant professor in the School of Computational Science and Engineering at Georgia Tech. Previously, he was a Research Scientist at the Oden Institute for Computational Engineering and Sciences at University of Texas at Austin. Before joining UT Austin, he spent a year as a lecturer and postdoc at ETH Zurich, 2014-2015. He obtained his Master and Ph.D. degrees in Computational Mathematics from EPFL under the supervision.