Sliced Gromov-Wasserstein
Titouan, Vayer, Flamary, Rémi, Courty, Nicolas, Tavenard, Romain, Chapel, Laetitia
–Neural Information Processing Systems
Recently used in various machine learning contexts, the Gromov-Wasserstein distance (GW) allows for comparing distributions whose supports do not necessarily lie in the same metric space. However, this Optimal Transport (OT) distance requires solving a complex non convex quadratic program which is most of the time very costly both in time and memory. Contrary to GW, the Wasserstein distance (W) enjoys several properties ({\em e.g.} duality) that permit large scale optimization. Among those, the solution of W on the real line, that only requires sorting discrete samples in 1D, allows defining the Sliced Wasserstein (SW) distance. This paper proposes a new divergence based on GW akin to SW.
Neural Information Processing Systems
Mar-19-2020, 02:46:34 GMT
- Technology: