rethinking transfer learning
Rethinking Transfer Learning for Medical Image Classification
Peng, Le, Liang, Hengyue, Luo, Gaoxiang, Li, Taihui, Sun, Ju
Transfer learning (TL) from pretrained deep models is a standard practice in modern medical image classification (MIC). However, what levels of features to be reused are problem-dependent, and uniformly finetuning all layers of pretrained models may be suboptimal. This insight has partly motivated the recent differential TL strategies, such as TransFusion (TF) and layer-wise finetuning (LWFT), which treat the layers in the pretrained models differentially. In this paper, we add one more strategy into this family, called TruncatedTL, which reuses and finetunes appropriate bottom layers and directly discards the remaining layers. This yields not only superior MIC performance but also compact models for efficient inference, compared to other differential TL methods. Our code is available at: https://github.com/sun-umn/TTL
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- Asia > China > Hubei Province > Wuhan (0.04)
- (2 more...)
- Research Report > New Finding (1.00)
- Instructional Material > Course Syllabus & Notes (0.84)
- Instructional Material > Online (0.61)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.88)
- Information Technology > Artificial Intelligence > Vision > Image Understanding (0.85)