Appears in collection : Research School - Jean Morlet Chair - Frontiers in Interacting Particle Systems, Aggregation-Diffusion Equations & Collective Behavior / Ecole - Chaire Jean Morlet - Frontières dans les équations de systèmes de particules en interaction. Equations d'agrégation-diffusion et comportement collectif
Sampling is a fundamental task in Machine Learning. For instance in Bayesian Machine Learning, one has to sample from the posterior distribution over the parameters of a learning model, whose density is known up to a normalizing constant. In other settings such as generative modelling, one has to sample from a distribution from which some samples are available (e.g. images). The task of sampling can be seen as an optimization problem over the space of probability measures. The mathematical theory providing the tools and concepts for optimization over the space of probability measures is the theory of optimal transport. The topic of this course will be the connection between optimization and sampling, more precisely, how to solve sampling problems using optimization ideas. The goal of the first part of the course will be to present two important concepts from optimal transport: Wasserstein gradient flows and geodesic convexity. We will introduce them by analogy with their euclidean counterpart that is well known in optimization. The goal of the second part will be to show how these concepts, along with standard optimization techniques, enable to design, improve and analyze various sampling algorithms. In particular. we will focus on several interacting particles schemes that achieve state-of-the-art performance in machine learning.