Sample Complexity Bounds for Estimating Probability Divergences under Invariances
Tahmasebi, Behrooz, Jegelka, Stefanie
–arXiv.org Artificial Intelligence
Estimating the optimal transportation distance [50, 51, 40] between probability measures is a fundamental problem in statistics, with many applications in machine learning, from Generative Adversarial Networks (GANs) [22, 2, 39, 31] to domain adaptation and generalization [18, 14, 13], geometric data processing (e.g., Wasserstein barycenters [15] and intrinsic dimension estimation [7]), biomedical research [53], and control and dynamical systems [9]. Estimating the Wasserstein distance is known to be a difficult task in general, and it suffers from the curse of dimensionality [49]. The slow convergence rate is generally unimprovable, as there exist probability measures that are difficult to estimate. However, those hard instances barely appear in practice when we study more structured probability measures. Indeed, in many applications (e.g., graphs, point clouds, molecules, spectral data), the underlying probability measures are invariant with respect to a group action on the input space. As observed in recent works [6, 12], considering the group invariances in the mathematical model can help improve the convergence rate of the Wasserstein distance and the Sobolev Integral Probability Metrics (Sobolev IPMs), with applications, e.g., to generative models for invariant data.
arXiv.org Artificial Intelligence
Nov-5-2023
- Country:
- Europe
- France > Grand Est
- Meurthe-et-Moselle > Nancy (0.04)
- Germany > Bavaria
- Upper Bavaria > Munich (0.04)
- France > Grand Est
- North America > United States
- Massachusetts > Middlesex County > Cambridge (0.04)
- Europe
- Genre:
- Research Report (0.82)
- Technology: