Goto

Collaborating Authors

 sampa



SAMPa: Sharpness-aware Minimization Parallelized

Neural Information Processing Systems

Sharpness-aware minimization (SAM) has been shown to improve the generalization of neural networks. However, each SAM update requires computing two gradients, effectively doubling the per-iteration cost compared to base optimizers like SGD. We propose a simple modification of SAM, termed SAMPa, which allows us to fully parallelize the two gradient computations. SAMPa achieves a twofold speedup of SAM under the assumption that communication costs between devices are negligible. Empirical results show that SAMPa ranks among the most efficient variants of SAM in terms of computational time. Additionally, our method consistently outperforms SAM across both vision and language tasks. Notably, SAMPa theoretically maintains convergence guarantees even for perturbation sizes, which is established through a novel Lyapunov function. We in fact arrive at SAMPa by treating this convergence guarantee as a hard requirement---an approach we believe is promising for developing SAM-based methods in general.


Sutton's predictions v Zambian rapper Sampa the Great

BBC News

Aston Villa have won nine games in a row in all competitions, but can they reach double figures by beating Manchester United on Sunday? Villa have gone behind in three of those games and haven't kept a clean sheet in their past four matches, said BBC Sport football expert Chris Sutton. But they have been so attacking and Morgan Rogers is absolutely flying. They just never seem to lie down. Sutton is making predictions for all 380 Premier League games this season, against AI, BBC Sport readers and a variety of guests. For week 17, he takes on Zambian musician and rapper Sampa the Great. Sampa the Great's new single, Can't Hold Us, is out now and is included in the EAFC 26 video game. Do you agree with their scores?



SAMPa: Sharpness-aware Minimization Parallelized

Neural Information Processing Systems

Sharpness-aware minimization (SAM) has been shown to improve the generalization of neural networks. However, each SAM update requires sequentially computing two gradients, effectively doubling the per-iteration cost compared to base optimizers like SGD. We propose a simple modification of SAM, termed SAMPa, which allows us to fully parallelize the two gradient computations. SAMPa achieves a twofold speedup of SAM under the assumption that communication costs between devices are negligible. Empirical results show that SAMPa ranks among the most efficient variants of SAM in terms of computational time.


SAMPa: Sharpness-aware Minimization Parallelized

Xie, Wanyun, Pethick, Thomas, Cevher, Volkan

arXiv.org Machine Learning

Sharpness-aware minimization (SAM) has been shown to improve the generalization of neural networks. However, each SAM update requires \emph{sequentially} computing two gradients, effectively doubling the per-iteration cost compared to base optimizers like SGD. We propose a simple modification of SAM, termed SAMPa, which allows us to fully parallelize the two gradient computations. SAMPa achieves a twofold speedup of SAM under the assumption that communication costs between devices are negligible. Empirical results show that SAMPa ranks among the most efficient variants of SAM in terms of computational time. Additionally, our method consistently outperforms SAM across both vision and language tasks. Notably, SAMPa theoretically maintains convergence guarantees even for \emph{fixed} perturbation sizes, which is established through a novel Lyapunov function. We in fact arrive at SAMPa by treating this convergence guarantee as a hard requirement -- an approach we believe is promising for developing SAM-based methods in general. Our code is available at \url{https://github.com/LIONS-EPFL/SAMPa}.


Streamlined Photoacoustic Image Processing with Foundation Models: A Training-Free Solution

Deng, Handi, Zhou, Yucheng, Xiang, Jiaxuan, Gu, Liujie, Luo, Yan, Feng, Hai, Liu, Mingyuan, Ma, Cheng

arXiv.org Artificial Intelligence

Foundation models have rapidly evolved and have achieved significant accomplishments in computer vision tasks. Specifically, the prompt mechanism conveniently allows users to integrate image prior information into the model, making it possible to apply models without any training. Therefore, we propose a method based on foundation models and zero training to solve the tasks of photoacoustic (PA) image segmentation. We employed the segment anything model (SAM) by setting simple prompts and integrating the model's outputs with prior knowledge of the imaged objects to accomplish various tasks, including: (1) removing the skin signal in three-dimensional PA image rendering; (2) dual speed-of-sound reconstruction, and (3) segmentation of finger blood vessels. Through these demonstrations, we have concluded that deep learning can be directly applied in PA imaging without the requirement for network design and training. This potentially allows for a hands-on, convenient approach to achieving efficient and accurate segmentation of PA images. This letter serves as a comprehensive tutorial, facilitating the mastery of the technique through the provision of code and sample datasets.