Sample and Map from a Single Convex Potential: Generation using Conjugate Moment Measures
AuthorsNina Vesseron†, Louis Béthune, Marco Cuturi
Sample and Map from a Single Convex Potential: Generation using Conjugate Moment Measures
AuthorsNina Vesseron†, Louis Béthune, Marco Cuturi
The canonical approach in generative modeling is to split model fitting into two blocks: define first how to sample noise (e.g. Gaussian) and choose next what to do with it (e.g. using a single map or flows). We explore in this work an alternative route that ties sampling and mapping. We find inspiration in moment measures, a result that states that for any measure ρ, there exists a unique convex potential u such that ρ = ∇u♯e-u. While this does seem to tie effectively sampling (from log–concave distribution e-u) and action (pushing particles through ∇u), we observe on simple examples (e.g., Gaussians or 1D distributions) that this choice is ill–suited for practical tasks. We study an alternative factorization, where ρ is factorized as ∇w*♯e-w, where w* is the convex conjugate of a convex potential w. We call this approach conjugate moment measures, and show far more intuitive results on these examples. Because ∇w* is the Monge map between the log–concave distribution e-w and ρ, we rely on optimal transport solvers to propose an algorithm to recover w from samples of ρ, and parameterize w as an input–convex neural network. We also address the common sampling scenario in which the density of ρ is known only up to a normalizing constant, and propose an algorithm to learn w in this setting.
On a Neural Implementation of Brenier’s Polar Factorization
July 16, 2024research area Methods and Algorithmsconference ICML
In 1991, Brenier proved a theorem that generalizes the polar decomposition for square matrices — factored as PSD unitary — to any vector field . The theorem, known as the polar factorization theorem, states that any field can be recovered as the composition of the gradient of a convex function with a measure-preserving map , namely . We propose a practical…
Resolving the Mixing Time of the Langevin Algorithm to its Stationary Distribution for Log-Concave Sampling
July 24, 2023research area Methods and Algorithmsconference COLT
Sampling from a high-dimensional distribution is a fundamental task in statistics, engineering, and the sciences. A canonical approach is the Langevin Algorithm, i.e., the Markov chain for the discretized Langevin Diffusion. This is the sampling analog of Gradient Descent. Despite being studied for several decades in multiple communities, tight mixing bounds for this algorithm remain unresolved even in the seemingly simple setting of log-concave…