View publication

In 1991, Brenier proved a theorem that generalizes the polar decomposition for square matrices -- factored as PSD ×\times unitary -- to any vector field F:RdRdF:\mathbb{R}^d\rightarrow \mathbb{R}^d. The theorem, known as the polar factorization theorem, states that any field FF can be recovered as the composition of the gradient of a convex function uu with a measure-preserving map MM, namely F=uMF=\nabla u \circ M. We propose a practical implementation of this far-reaching theoretical result, and explore possible uses within machine learning. The theorem is closely related to optimal transport (OT) theory, and we borrow from recent advances in the field of neural optimal transport to parameterize the potential uu as an input convex neural network. The map MM can be either evaluated pointwise using uu^*, the convex conjugate of uu, through the identity M=uFM=\nabla u^* \circ F, or learned as an auxiliary network. Because MM is, in general, not injective, we consider the additional task of estimating the ill-posed inverse map that can approximate the pre-image measure M1M^{-1} using a stochastic generator. We illustrate possible applications of Brenier's polar factorization to non-convex optimization problems, as well as sampling of densities that are not log-concave.

Related readings and updates.

Learning Elastic Costs to Shape Monge Displacements

Given a source and a target probability measure supported on Rd\mathbb{R}^dRd, the Monge problem aims for the most efficient way to map one distribution to the other. This efficiency is quantified by defining a cost function between source and target data. Such a cost is often set by default in the machine learning literature to the squared-Euclidean distance, ℓ22(x,y)=12∥x−y∥22\ell^2_2(x,y)=\tfrac12\|x-y\|_2^2ℓ22​(x,y)=21​∥x−y∥22​. The benefits…
See paper details

The Monge Gap: A Regularizer to Learn All Transport Maps

Optimal transport (OT) theory has been been used in machine learning to study and characterize maps that can push-forward efficiently a probability measure onto another. Recent works have drawn inspiration from Brenier's theorem, which states that when the ground cost is the squared-Euclidean distance, the "best" map to morph a continuous measure in P(Rd)\mathcal{P}(\mathbb{R}^d)P(Rd) into another must be the gradient of a convex function. To…
See paper details