Goto

Collaborating Authors

 basin


Asymptotic and Finite-Time Guarantees for Langevin-Based Temperature Annealing in InfoNCE

Chaudhry, Faris

arXiv.org Machine Learning

The InfoNCE loss in contrastive learning depends critically on a temperature parameter, yet its dynamics under fixed versus annealed schedules remain poorly understood. We provide a theoretical analysis by modeling embedding evolution under Langevin dynamics on a compact Riemannian manifold. Under mild smoothness and energy-barrier assumptions, we show that classical simulated annealing guarantees extend to this setting: slow logarithmic inverse-temperature schedules ensure convergence in probability to a set of globally optimal representations, while faster schedules risk becoming trapped in suboptimal minima. Our results establish a link between contrastive learning and simulated annealing, providing a principled basis for understanding and tuning temperature schedules.



Supplementary Material and Datasheet: Off to new Shores: A Dataset & Benchmark for (near-)coastal Flood Inundation Forecasting Contents

Neural Information Processing Systems

This supplementary document follows the Datasheets for Datasets template of (8) to document the Global Flood Forecasting (GFF) dataset and its creation. Further resources are provided: in the accompanying publication https://arxiv.org/abs/2409.18591 in the GitHub repository https://github.com/Multihuntr/GFF







FreeProbabilityforpredictingtheperformanceof feed-forwardfullyconnectedneuralnetworks

Neural Information Processing Systems

We also nuance the idea that learning happens at the edge of chaos by giving evidence that avery desirable feature forneural networks isthehyperbolicity of their Jacobian at initialization.