Goto

Collaborating Authors

 lipschitz continuity


Sinkhorn Barycenters with Free Support via Frank-Wolfe Algorithm

Giulia Luise, Saverio Salzo, Massimiliano Pontil, Carlo Ciliberto

Neural Information Processing Systems

We present a novel algorithm to estimate the barycenter of arbitrary probability distributions with respect to the Sinkhorn divergence. Based on a Frank-Wolfe optimization strategy, our approach proceeds by populating the support of the barycenter incrementally, without requiring any pre-allocation.






TowardsSharperGeneralizationBoundsfor StructuredPrediction

Neural Information Processing Systems

Specifically,inPAC-Bayesian approach, [45,26,4,22]provide the generalization bounds of order O( 1 n). In implicit embedding approach, [12, 13, 52, 11, 58, 7] provide the convergence rate of orderO( 1n1/4), and [53] of orderO( 1 n). In the factor graph decomposition approach, [18, 51] present the generalization upper bounds of orderO( 1 n).


fef6f971605336724b5e6c0c12dc2534-Supplemental.pdf

Neural Information Processing Systems

I W scalars. Taking an expectation on both sides of (17) we obtain { } The next lemma characterizes the spectral properties of the disagreement matrix, used in Lemma 4. W is also a stochastic matrix. W are that of I W, each with multiplicity K. W) = 1 with multiplicity K. Again we can check that the eigenspace of ( λ We prove this result by induction on n. For n = 1 it is trivial. Now assume that the inequality holds for all l n 1. We provide the proof here for completeness.



a29a5ba2cb7bdeabba22de8c83321b46-Paper.pdf

Neural Information Processing Systems

Recently, two families of self-supervised methods, contrastive learning and latent bootstrapping, exemplified by SimCLR and BYOL respectively,havemade significant progress.