Trade-off between dependence and complexity in Wasserstein distance learning

Young-Heon Kim Co-Author
University of British Columbia
 
Soumik Pal Co-Author
University of Washington, Seattle
 
Geoffrey Schiebinger Co-Author
University of British Columbia
 
Nabarun Deb First Author
 
Nabarun Deb Presenting Author
 
Sunday, Aug 4: 3:35 PM - 3:50 PM
3266 
Contributed Papers 
Oregon Convention Center 
The Wasserstein distance is a powerful tool in modern machine learning to metrize the space of probability distributions in a way that takes into account the geometry of the domain.
Therefore, a lot of attention has been devoted in the literature to understanding rates of convergence for Wasserstein distances based on iid data. However, often in machine learning applications, especially in reinforcement learning, object tracking, performative prediction, and other online learning problems, observations are received sequentially, rendering some inherent temporal dependence. Motivated by this observation, we attempt to understand the problem of estimating Wasserstein distances using the natural plug-in estimator based on stationary beta-mixing sequences, a widely used assumption in the study of dependent processes. Our rates of convergence results are applicable under both short and long-range dependence. As expected, under short-range dependence, the rates match those observed in the iid. case. Interestingly, however, even under long-range dependence, we can show that the rates can match those in the iid case provided the (intrinsic) dimension is large enough.

Keywords

Entropy regularized optimal transport

Mckean-Vlasov diffusion

mirror descent

parabolic Monge-Amp`ere

Sinkhorn algorithm

Wasserstein mirror gradient flow 

Abstracts


Main Sponsor

IMS