Machine Learning for Biomedical Simulations

Prof. Daniel M. Tartakovsky, Stanford University

  • Data: 23 novembre 2021 dalle 11:00 alle 12:30

  • Luogo: Aula LAMC - Viale del Risorgimento, 2 - Bologna; online su Teams

Machine Learning for Biomedical Simulations

Prof. Daniel M. Tartakovsky, Stanford University (https://profiles.stanford.edu/daniel-tartakovsky)
November, 23rd, 2021
LAMC (viale Risorgimento 2, Bologna)
Link to attend via Microsoft Teams: teams link

Abstract

Neural networks (NNs) have become ubiquitous in scientific computing, including qualitative predictions of biomedical phenomena. Yet, the question of what they are good for remains unresolved. We discuss several where NNs might offer advantages over the existing alternatives. One is the use of NNs as surrogates or emulators of partial differential equations (PDEs) that describe the dynamics of complex systems. A virtually negligible computational cost of such surrogates renders them an attractive tool for ensemble-based computation, which requires a large number of repeated PDE solves. Since the latter are also needed to generate sufficient data for NN training, the usefulness of NN-based surrogates hinges on the balance between the training cost and the computational gain stemming from their deployment. We rely on multi-fidelity simulations to reduce the cost of data generation for subsequent training of a deep convolutional NN (CNN) using transfer learning. High- and low-fidelity images are generated by solving PDEs on fine and coarse meshes, respectively. We use theoretical results for multilevel Monte Carlo to guide our choice of the numbers of images of each kind. We demonstrate the performance of this multi-fidelity training strategy on the problem of estimation of the distribution of a quantity of interest, whose dynamics is governed by a system of nonlinear PDEs (e.g., parabolic PDEs describing flow and transport in microcirculation) with uncertain/random parameters. Our numerical experiments demonstrate that a mixture of a comparatively large number of low-fidelity data and smaller numbers of high- and low-fidelity data provides an optimal balance of computational speed-up and prediction accuracy. The former is reported relative to both CNN training on high-fidelity images only and Monte Carlo solution of the PDEs. The latter is expressed in terms of both the Wasserstein distance and the Kullback-Leibler divergence.