Bayesian Learning with Wasserstein Barycenters
In this work we introduce a novel paradigm for Bayesian learning based on optimal transport theory. Namely, we propose to use the Wasserstein barycenter of the posterior law on models, as an alternative to the maximum a posteriori estimator (MAP) and Bayes predictive distributions. We exhibit conditions granting the existence and consistency of this estimator, discuss some of its basic and specific properties, and propose a numerical approximation relying on standard posterior sampling in general finite-dimensional parameter spaces. We thus also contribute to the recent blooming of applications of optimal transport theory in machine learning, beyond the discrete and semidiscrete settings so far considered. Advantages of the proposed estimator are discussed and illustrated with numerical simulations.
READ FULL TEXT