Satellite and a subset of the main workshop on

Statistical modeling for stochastic processes and related fields online

This is a satellite Bayesian/Monte Carlo workshop on Statistical modeling for stochastic processes and related fields (Sep 27-30) which will be held on 29 Sep 2021.

Register from the zoom form.

Program 29 Sep 2021 (JST = UTC + 9h)

[Satellite workshop]

14:55 - 15:30 Goda Takashi (University of Tokyo)

15:30 - 16:05 Xin Tong (National University of Singapore)

16:05 - 16:40 Alexandre H. Thiery (National University of Singapore)

[Main workshop]

17:00 - 17:35 Kengo Kamatani (ISM, Tokyo)

17:35 - 18:10 Björn Sprungk (TU Bergakademie Freiberg)

18:20 - 18:55 Tony Lelièvre (Ecole des Ponts ParisTech)

18:55 - 19:30 Alexandros Beskos (University College London)

Abstract

Goda Takashi (University of Tokyo)

Multilevel Monte Carlo methods for efficient nested simulations

Nested simulation problem arises in a number of different applications, such as financial risk estimation, medical decision making, and construction of Bayesian experimental designs. Probably the most standard approach to tackle this problem is to consider naive Monte Carlo estimator for each of inner and outer loops, and then ``nest" them. However, it is well-known that the total computational cost required for such a nested Monte Carlo estimator to achieve the root-mean-squared error $\epsilon$ is typically of order $\epsilon^{-3}$ and even increases up to order $\epsilon^{-4}$ in the worse cases.


In this talk, we discuss applying multilevel Monte Carlo (MLMC) methods to reduce this cost significantly. In particular, motivated by applications to medical decision making, we focus on estimating what is called the expected value of information, and prove under some assumptions that the total computational cost for our MLMC estimator is of order $\epsilon^{-2}$. We also present some results of numerical experiments, which confirm the considerable computational savings achieved by our MLMC estimator as compared to the nested Monte Carlo estimator.

Xin Tong (National University of Singapore)

Can Algorithms Collaborate? The Replica Exchange Method and Its Spectral Gap

Abstract: Gradient descent (GD) is known to converge quickly for convex objective functions, but it can be trapped at local minima. On the other hand, Langevin dynamics (LD) can explore the state space and find global minima, but in order to give accurate estimates, LD needs to run with a small discretization step size and weak stochastic force, which in general slows down its convergence. This paper shows that these two algorithms can “collaborate” through a simple exchange mechanism, in which they swap their current positions if LD yields a lower objective function. This idea can be seen as the singular limit of the replica-exchange technique from the sampling literature. We show that this new algorithm converges to the global minimum linearly with high probability, assuming the objective function is strongly convex in a neighborhood of the unique global minimum. By replacing gradients with stochastic gradients, and adding a proper threshold to the exchange mechanism, our algorithm can also be used in online settings. We further verify our theoretical results through some numerical experiments, and observe superior performance of the proposed algorithm over running GD or LD alone. We will further explain how does replica exchange method improve LD’s spectral gap using Poincare type of inequality.

Alexandre H. Thiery (National University of Singapore)

Exploiting geometry for walking larger steps in Bayesian Inverse Problems

Abstract: Consider the observation $y = F(x) + \xi$ of a quantity of interest $x$ -- the random variable $\xi \sim \mathcal{N}(0, \sigma^2 I)$ is a vector of additive noise in the observation. In Bayesian inverse problems, the vector $x$ typically represents the high-dimensional discretization of a continuous and unobserved field while the evaluations of the forward operator $F(\cdot)$ involve solving a system of partial differential equations. In the low-noise regime, i.e. $\sigma \to 0$, the posterior distribution concentrates in the neighborhood of a nonlinear manifold. As a result, the efficiency of standard MCMC algorithms deteriorates due to the need to take increasingly smaller steps.

In this work, we present a constrained HMC algorithm that is robust to small $\sigma$ values, i.e. low noise. Taking the observations generated by the model to be constraints on the prior, we define a manifold on which the constrained HMC algorithm generates samples. By exploiting the geometry of the manifold, our algorithm is able to take larger step sizes than more standard MCMC methods, resulting in a more efficient sampler. If time permits, we will describe how similar ideas can be leveraged within other non-reversible samplers.

Kengo Kamatani (ISM, Tokyo)

Scaling limit analysis of some piecewise deterministic Markov processes

Recently, piecewise deterministic Markov processes have gained interest in the Monte Carlo community in the context of scalable Monte Carlo integration methods. In this talk, we will discuss scaling limits for some piecewise deterministic Markov processes. We will describe these results using multiscale analysis, which is a useful technique for this purpose.


This is joint work with J. Bierkens (TU Delft) and G. O. Roberts (Warwick).

Björn Sprungk (TU Bergakademie Freiberg)

Robust sampling methods for Bayesian inverse problems with small observational noise

The Bayesian approach to inverse problems provides a rigorous framework for the incorporation and quantification of uncertainties in measurements, parameters and models. We are interested in designing numerical methods which are robust w.r.t. the size of the observational noise, i.e., methods which behave well in case of concentrated posterior measures. The concentration of the posterior is a highly desirable situation in practice, since it relates to informative or large data. However, it can pose a computational challenge for numerical methods based on the prior as reference measure. We propose to employ the Laplace approximation of the posterior as the base measure for numerical integration in this context. The Laplace approximation is a Gaussian measure centered at the maximum a-posteriori estimate and with a covariance matrix depending on the Hessian of the log posterior density. We discuss the convergence of the Laplace approximation to the posterior measure in the Hellinger distance and analyze the efficiency of Monte Carlo methods based on it. In particular, we show that Laplace-based importance sampling and Laplace-based Markov chain Monte Carlo methods are robust w.r.t. the concentration of the posterior for large classes of posterior distributions and integrands whereas prior-based sampling methods are not.

This is joint work together with Claudia Schillings (U Mannheim), Daniel Rudolf (U Göttingen), and Philipp Wacker (U Erlangen-Nürnberg).

Tony Lelièvre (Ecole des Ponts ParisTech)

Adaptive importance sampling methods

We will present the general strategy of free energy adaptive importance sampling methods. These are very efficient algorithms to sample multimodal distributions with an information on the modes given by a so-called collective variable which indexes the modes. We will also discuss recent techniques to build this collective variable.


References:

- T. Lelièvre, M. Rousset and G. Stoltz, Long-time convergence of an Adaptive Biasing Force method, Nonlinearity, 21, 1155-1181 (2008).

- G. Fort, B. Jourdain, T. Lelièvre and G. Stoltz, Convergence and efficiency of adaptive importance sampling techniques with partial biasing, Journal of Statistical Physics, 171(2), 220-268 (2018).

- B. Jourdain, T. Lelièvre and P.A. Zitt, Convergence of metadynamics: discussion of the adiabatic hypothesis, to appear in Annals of Applied Probability.

- Z. Belkacemi, P. Gkeka, T. Lelièvre and G. Stoltz, Chasing Collective Variables using Autoencoders and biased trajectories, https://arxiv.org/abs/2104.11061 .

Alexandros Beskos (University College London)

Manifold Markov chain Monte Carlo methods for Bayesian inference in diffusion models

Bayesian inference for nonlinear diffusions, observed at discrete times, is a challenging task that has prompted the development of a number of algorithms, mainly within the computational statistics community. We propose a new direction, and accompanying methodology – borrowing ideas from statistical physics and computational chemistry – for inferring the posterior distribution of latent diffusion paths and model parameters, given observations of the process. Joint configurations of the underlying process noise and of parameters, mapping onto diffusion paths consistent with observations, form an implicitly defined manifold. Then, by making use of a constrained Hamiltonian Monte Carlo algorithm on the embedded manifold, we are able to perform computationally efficient inference for a class of discretely observed diffusion models. Critically, in contrast with other approaches proposed in the literature, our methodology is highly automated, requiring minimal user intervention and applying alike in a range of settings, including: elliptic or hypo-elliptic systems; observations with or without noise; linear or non-linear observation operators. Exploiting Markovianity, we propose a variant of the method with complexity that scales linearly in the resolution of path discretisation and the number of observation times. Example Python code is given at git.io/m-mcmc.

This is joint work with Matthew M Graham and Alexandre H Thiery.