Newer is not always better: Rethinking transferability metrics, their peculiarities, stability and performance
Ibrahim, Shibal, Ponomareva, Natalia, Mazumder, Rahul
–arXiv.org Artificial Intelligence
Fine-tuning of large pre-trained image and language models on small customized datasets has become increasingly popular for improved prediction and efficient use of limited resources. Fine-tuning requires identification of best models to transfer-learn from and quantifying transferability prevents expensive re-training on all of the candidate models/tasks pairs. In this paper, we show that the statistical problems with covariance estimation drive the poor performance of H-score -- a common baseline for newer metrics -- and propose shrinkage-based estimator. This results in up to 80% absolute gain in H-score correlation performance, making it competitive with the state-of-the-art LogME measure. Our shrinkage-based H-score is $3\times$-10$\times$ faster to compute compared to LogME. Additionally, we look into a less common setting of target (as opposed to source) task selection. We demonstrate previously overlooked problems in such settings with different number of labels, class-imbalance ratios etc. for some recent metrics e.g., NCE, LEEP that resulted in them being misrepresented as leading measures. We propose a correction and recommend measuring correlation performance against relative accuracy in such settings. We support our findings with ~164,000 (fine-tuning trials) experiments on both vision models and graph neural networks.
arXiv.org Artificial Intelligence
May-26-2023
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > France (0.04)
- North America > United States
- Massachusetts > Middlesex County
- Cambridge (0.14)
- New York > New York County
- New York City (0.04)
- Massachusetts > Middlesex County
- Asia > Middle East
- Genre:
- Research Report
- Experimental Study (0.46)
- New Finding (0.66)
- Research Report
- Technology: