Overleaf Example
–Neural Information Processing Systems
A widely used algorithm for transfer learning is fine-tuning, where a pre-trained model is fine-tuned on a target task with a small amount of labeled data. When the capacity of the pre-trained model is much larger than the size of the target data set, fine-tuning is prone to overfitting and "memorizing" the training labels. Hence, an important question is to regularize fine-tuning and ensure its robustness to noise. To address this question, we begin by analyzing the generalization properties of fine-tuning. We present a PAC-Bayes generalization bound that depends on the distance traveled in each layer during fine-tuning and the noise stability of the fine-tuned model.
Neural Information Processing Systems
Mar-22-2025, 18:21:38 GMT
- Genre:
- Research Report > New Finding (0.69)
- Industry:
- Health & Medicine > Diagnostic Medicine > Imaging (0.68)
- Technology: