Sunday, Aug 4: 3:35 PM - 3:50 PM
3266
Contributed Papers
Oregon Convention Center
The Wasserstein distance is a powerful tool in modern machine learning to metrize the space of probability distributions in a way that takes into account the geometry of the domain.
Therefore, a lot of attention has been devoted in the literature to understanding rates of convergence for Wasserstein distances based on iid data. However, often in machine learning applications, especially in reinforcement learning, object tracking, performative prediction, and other online learning problems, observations are received sequentially, rendering some inherent temporal dependence. Motivated by this observation, we attempt to understand the problem of estimating Wasserstein distances using the natural plug-in estimator based on stationary beta-mixing sequences, a widely used assumption in the study of dependent processes. Our rates of convergence results are applicable under both short and long-range dependence. As expected, under short-range dependence, the rates match those observed in the iid. case. Interestingly, however, even under long-range dependence, we can show that the rates can match those in the iid case provided the (intrinsic) dimension is large enough.
Entropy regularized optimal transport
Mckean-Vlasov diffusion
mirror descent
parabolic Monge-Amp`ere
Sinkhorn algorithm
Wasserstein mirror gradient flow
Main Sponsor
IMS