fast learner
SAFE TrainedModels
After calibrating in the first session, the slow efficient tuning parameters can capture more informativefeatures, improving generalization to incoming classes. Moreover, to further incorporate novel concepts, we strikeabalance between stability and plasticity byfixing slowefficient tuning parameters and continuously updating the fast ones. Specifically, a cross-classification loss with feature alignment is proposed to circumvent catastrophic forgetting.
SAFE: Slow and Fast Parameter-Efficient Tuning for Continual Learning with Pre-Trained Models
Zhao, Linglan, Zhang, Xuerui, Yan, Ke, Ding, Shouhong, Huang, Weiran
Continual learning aims to incrementally acquire new concepts in data streams while resisting forgetting previous knowledge. With the rise of powerful pre-trained models (PTMs), there is a growing interest in training incremental learning systems using these foundation models, rather than learning from scratch. Existing works often view PTMs as a strong initial point and directly apply parameter-efficient tuning (PET) in the first session for adapting to downstream tasks. In the following sessions, most methods freeze model parameters for tackling forgetting issues. However, applying PET directly to downstream data cannot fully explore the inherent knowledge in PTMs. Additionally, freezing the parameters in incremental sessions hinders models' plasticity to novel concepts not covered in the first session. To solve the above issues, we propose a Slow And Fast parameter-Efficient tuning (SAFE) framework. In particular, to inherit general knowledge from foundation models, we include a transfer loss function by measuring the correlation between the PTM and the PET-applied model. After calibrating in the first session, the slow efficient tuning parameters can capture more informative features, improving generalization to incoming classes. Moreover, to further incorporate novel concepts, we strike a balance between stability and plasticity by fixing slow efficient tuning parameters and continuously updating the fast ones. Specifically, a cross-classification loss with feature alignment is proposed to circumvent catastrophic forgetting. During inference, we introduce an entropy-based aggregation strategy to dynamically utilize the complementarity in the slow and fast learners. Extensive experiments on seven benchmark datasets verify the effectiveness of our method by significantly surpassing the state-of-the-art.
Dynamic Long-Term Time-Series Forecasting via Meta Transformer Networks
Ma'sum, Muhammad Anwar, Sarkar, MD Rasel, Pratama, Mahardhika, Ramasamy, Savitha, Anavatti, Sreenatha, Liu, Lin, Habibullah, null, Kowalczyk, Ryszard
A reliable long-term time-series forecaster is highly demanded in practice but comes across many challenges such as low computational and memory footprints as well as robustness against dynamic learning environments. This paper proposes Meta-Transformer Networks (MANTRA) to deal with the dynamic long-term time-series forecasting tasks. MANTRA relies on the concept of fast and slow learners where a collection of fast learners learns different aspects of data distributions while adapting quickly to changes. A slow learner tailors suitable representations to fast learners. Fast adaptations to dynamic environments are achieved using the universal representation transformer layers producing task-adapted representations with a small number of parameters. Our experiments using four datasets with different prediction lengths demonstrate the advantage of our approach with at least $3\%$ improvements over the baseline algorithms for both multivariate and univariate settings. Source codes of MANTRA are publicly available in \url{https://github.com/anwarmaxsum/MANTRA}.
- Oceania > Australia > South Australia > Adelaide (0.04)
- Oceania > Australia > Australian Capital Territory > Canberra (0.04)
- North America > United States (0.04)
- (3 more...)
- Information Technology > Data Science > Data Mining (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.68)
Continual Learning, Fast and Slow
Pham, Quang, Liu, Chenghao, Hoi, Steven C. H.
According to the Complementary Learning Systems (CLS) theory~\cite{mcclelland1995there} in neuroscience, humans do effective \emph{continual learning} through two complementary systems: a fast learning system centered on the hippocampus for rapid learning of the specifics, individual experiences; and a slow learning system located in the neocortex for the gradual acquisition of structured knowledge about the environment. Motivated by this theory, we propose \emph{DualNets} (for Dual Networks), a general continual learning framework comprising a fast learning system for supervised learning of pattern-separated representation from specific tasks and a slow learning system for representation learning of task-agnostic general representation via Self-Supervised Learning (SSL). DualNets can seamlessly incorporate both representation types into a holistic framework to facilitate better continual learning in deep neural networks. Via extensive experiments, we demonstrate the promising results of DualNets on a wide range of continual learning protocols, ranging from the standard offline, task-aware setting to the challenging online, task-free scenario. Notably, on the CTrL~\cite{veniat2020efficient} benchmark that has unrelated tasks with vastly different visual images, DualNets can achieve competitive performance with existing state-of-the-art dynamic architecture strategies~\cite{ostapenko2021continual}. Furthermore, we conduct comprehensive ablation studies to validate DualNets efficacy, robustness, and scalability. Code will be made available at \url{https://github.com/phquang/DualNet}.
- Health & Medicine > Therapeutic Area > Neurology (0.68)
- Education > Educational Setting (0.46)
- Education > Curriculum (0.46)
DualNet: Continual Learning, Fast and Slow
Pham, Quang, Liu, Chenghao, Hoi, Steven
According to Complementary Learning Systems (CLS) theory~\citep{mcclelland1995there} in neuroscience, humans do effective \emph{continual learning} through two complementary systems: a fast learning system centered on the hippocampus for rapid learning of the specifics and individual experiences, and a slow learning system located in the neocortex for the gradual acquisition of structured knowledge about the environment. Motivated by this theory, we propose a novel continual learning framework named "DualNet", which comprises a fast learning system for supervised learning of pattern-separated representation from specific tasks and a slow learning system for unsupervised representation learning of task-agnostic general representation via a Self-Supervised Learning (SSL) technique. The two fast and slow learning systems are complementary and work seamlessly in a holistic continual learning framework. Our extensive experiments on two challenging continual learning benchmarks of CORE50 and miniImageNet show that DualNet outperforms state-of-the-art continual learning methods by a large margin. We further conduct ablation studies of different SSL objectives to validate DualNet's efficacy, robustness, and scalability. Code will be made available upon acceptance.
SAP Sponsored Machine Learning Study: Lack of Strategic Clarity
A few years ago, machine learning was virtually unheard of outside the geek press; now it's blasted past cutting-edge to the top of the strategic agenda. In fact, in a recent study by SAP and the Economist Intelligence Unit, "Making the Most of Machine Learning: 5 Lessons from Fast Learners," 68 percent of companies surveyed are using machine learning in some form; among procurement companies, it is about 65 percent. These companies are on a path toward automation. Social psychologist and Harvard professor Shoshana Zuboff said, "Everything that can be automated should be automated." Zuboff's viewpoint is particularly true for what we call "knowledge worker" tasks.
Machine Learning: Where Thinking Big Doesn't Mean Being Big
Artificial intelligence, machine learning, and deep learning – these emerging technologies are making headlines with publicity stunts and preliminary breakthroughs for industry giants with deep pockets. While most CEOs and senior leaders are quick to dismiss the next level of predictive analytics as more parlor trick than business case, a growing segment of midsize businesses is beginning to prove them wrong. "Making the Most of Machine Learning: Lessons from 5 Fast Learners," an SAP study conducted by the Economist Intelligence Unit (EIU), reported that small businesses (32%) and midsize companies (42%) are using machine learning for at least one business process. This finding is a stark difference compared to the adoption rates of large enterprises (26%), which traditionally have the resources (that their smaller competitors don't have) to implement such intelligent technology. Contrary to the hype surrounding machine learning, the progress that small and midsize businesses are making in this area is deeply rooted in future growth.