Skip to main content
Browse by:
GROUP

Approximating functions, functionals and operators with neural networks for diverse applications

George Karniadakis in front of blackboard
Friday, November 05, 2021
3:30 pm - 4:30 pm
George Em Karniadakis

We will review physics-informed neural network and summarize available extensions for applications in computational mechanics and beyond. We will also introduce new NNs that learn functionals and nonlinear operators from functions and corresponding responses for system identification. The universal approximation theorem of operators is suggestive of the potential of NNs in learning from scattered data any continuous operator or complex system. We first generalize the theorem to deep neural networks, and subsequently we apply it to design a new composite NN with small generalization error, the deep operator network (DeepONet), consisting of a NN for encoding the discrete input function space (branch net) and another NN for encoding the domain of the output functions (trunk net). We demonstrate that DeepONet can learn various explicit operators, e.g., integrals, Laplace transforms and fractional Laplacians, as well as implicit operators that represent deterministic and stochastic differential equations. More generally, DeepOnet can learn multiscale operators spanning across many scales and trained by diverse sources of data simultaneously. Join Zoom Meeting: https://duke.zoom.us/j/99946567210
Meeting ID: 999 4656 7210

Contact: Ariel Dawn