Transfer Learning
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
- Energy (0.46)
- Leisure & Entertainment (0.46)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.41)
Learning with Preserving for Continual Multitask Learning
Wang, Hanchen David, Bae, Siwoo, Chen, Zirong, Ma, Meiyi
Artificial intelligence systems in critical fields like autonomous driving and medical imaging analysis often continually learn new tasks using a shared stream of input data. For instance, after learning to detect traffic signs, a model may later need to learn to classify traffic lights or different types of vehicles using the same camera feed. This scenario introduces a challenging setting we term Continual Multitask Learning (CMTL), where a model sequentially learns new tasks on an underlying data distribution without forgetting previously learned abilities. Existing continual learning methods often fail in this setting because they learn fragmented, task-specific features that interfere with one another. To address this, we introduce Learning with Preserving (LwP), a novel framework that shifts the focus from preserving task outputs to maintaining the geometric structure of the shared representation space. The core of LwP is a Dynamically Weighted Distance Preservation (DWDP) loss that prevents representation drift by regularizing the pairwise distances between latent data representations. This mechanism of preserving the underlying geometric structure allows the model to retain implicit knowledge and support diverse tasks without requiring a replay buffer, making it suitable for privacy-conscious applications. Extensive evaluations on time-series and image benchmarks show that LwP not only mitigates catastrophic forgetting but also consistently outperforms state-of-the-art baselines in CMTL tasks. Notably, our method shows superior robustness to distribution shifts and is the only approach to surpass the strong single-task learning baseline, underscoring its effectiveness for real-world dynamic environments.
- North America > United States > Tennessee > Davidson County > Nashville (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- Europe > Switzerland (0.04)
- Health & Medicine > Diagnostic Medicine > Imaging (0.66)
- Transportation > Ground > Road (0.54)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.49)
Pre - Trained Model Sets for Target Tasks Target Tasks: Model Universe
We study model reusability evaluation (MRE) for source pre-trained models: evaluating their transfer learning performance to new target tasks. In special, we focus on the setting under which the target training datasets are small, making it difficult to produce reliable MRE scores using them. Under this situation, we propose synergistic learning for building the task-model metric, which can be realized by collecting a set of pre-trained models and asking a group of data providers to participate. We provide theoretical guarantees to show that the learned task-model metric distances can serve as trustworthy MRE scores, and propose synergistic learning algorithms and models for general learning tasks. Experiments show that the MRE models learned by synergistic learning can generate significantly more reliable MRE scores than existing approaches for small-data transfer learning.
- Asia > China > Jiangsu Province > Nanjing (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.66)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.55)
- Asia > Middle East > Jordan (0.04)
- North America > Canada (0.04)
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- (2 more...)
Appendix: On Learning Domain-Invariant Representations for Transfer Learning with Multiple Sources
Appendix C contains the proof for trade-off theorem discussed in Section 2.4 of our main In (4), Fubini theorem is invoked to swap the integral signs. This bound is novel since it relates loss on input space and data shift on feature space. Lemma 4. Given a source mixture and a target domain, we have the following d Now we are ready to prove the bound which motivate compressed DI representation. The objective function in Eq. (6) can be viewed as training the optimal hypothesis This estimation is unbiased estimation, i.e., To make it more consistent, we discuss under which condition the Hellinger loss is in the family defined in (10). Model We train a hypothesis ˆ f = ˆ h g and minimize the classification loss w.r.t.
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.40)
- Oceania > Australia (0.04)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- (2 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.90)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.68)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- Europe > Austria (0.04)
- North America > United States > Virginia (0.04)
- (2 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.93)
- Information Technology > Security & Privacy (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.69)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.41)
Transfer learning with a deeper backbone: We have now tested transfer learning with an adversarially trained
We thank the reviewers for their insightful comments. A number of additional experiments were suggested. We have updated our paper to include these results. For example, our adversarially queried R2-D2 model achieves 35.53% robust 5-shot Out-of-distribution testing: We have now evaluated our models on Meta-Dataset. For the same models, on the CUB-200 dataset, the robust accuracies are 29 .04%