Evaluating the Credible use of Scientific Machine Learning for High Consequence Applications

Abstract:
Machine-learned models are increasingly being used in lieu of, to complement, or as surrogates for classic computational models. The emerging field of scientific machine learning (SciML) seeks to fuse traditional mathematical modeling with advances in machine learning to handle challenges such as the implementation of numerical solvers, model-form error estimations, and the computational expense of high-fidelity models. SciML models balance mechanist equations with data-driven inference, resulting in computational models that preserve scientific knowledge while readily adapting to the unknown through data-driven discovery. The practical adoption of SciML is evident by its impact in a variety of domains that include climate, epidemiology, turbulence, quantum mechanics, and biology. The integration of data-driven methods with mechanistic models has highlighted the need to evaluate the credibility of the resulting SciML models used for high-consequence decisions. 

In this presentation we will articulate why we care about credibility and review the Sandia framework known as the Predictive Capability Maturity Model (PCMM) currently used for evaluating classic computational models. We will further the discussion of SciML credibility with an example of neural network (NN) function approximations of model-form error with application to epidemiology. A valuable tool to reduce model discrepancy, NN approximations to model-form error introduce new challenges with evaluating core mathematical modeling principles in verification, validation, and uncertainty quantification (VVUQ). To mitigate these risks, we propose that the credible use of SciML models may be applied as surrogates for high fidelity heterogenous stochastic models (e.g., agent-based models) where we preserve scientific knowledge while calibrating chaotic systems of human behavior through data-driven discovery.  

Ultimately, this work seeks to motivate the scientific modeling community and decision makers to ask the question: When do the benefits of using the SciML framework outweigh the challenges of evaluating its credibility? 

Bio:
Erin C.S. Acquesta is a Mathematician and Principal Member of the Technical Staff with the Applied Information Sciences Center at Sandia National Laboratories. She received her MS and PhD from North Carolina State University in the field of applied mathematics. Her research areas of interest include scientific machine learning, uncertainty quantification, sensitivity analysis, and machine learning explainability; with an emphasis on providing credible, adaptive, and interpretable modeling capabilities for enhanced situational awareness in support of national security decision-making. Her primary domain area of expertise focuses on the mathematical properties of infectious disease models. As an area of professional service, she is a member of the ASME VVUQ standards committee, writing definitions for verification, validation, uncertainty quantification, and credibility for machine-learned models. She also volunteers her time for educational outreach as a mentor and head referee for the Albuquerque VEX VRC Robotics League and NM State VEX VRC Robotics Competitions. 

Summary: