wscl
AI can better retain what it learns by mimicking human sleep
Human brains consolidate memories while sleeping. Could AI systems use the same technique? Building AIs that sleep and dream can lead to better results and more reliable models, according to researchers who aim to replicate the architecture and behaviour of the human brain. But other experts say recreating the intelligence we see within ourselves may not be the most fruitful path for AI research. Concetto Spampinato and his colleagues at the University of Catania, Italy, were looking for ways to avoid a phenomenon known as "catastrophic forgetting", where an AI model trained to do a new task loses the ability to carry out jobs it previously aced.
- Europe > Italy (0.25)
- Europe > United Kingdom (0.05)
Wake-Sleep Consolidated Learning
Sorrenti, Amelia, Bellitto, Giovanni, Salanitri, Federica Proietto, Pennisi, Matteo, Palazzo, Simone, Spampinato, Concetto
We propose Wake-Sleep Consolidated Learning (WSCL), a learning strategy leveraging Complementary Learning System theory and the wake-sleep phases of the human brain to improve the performance of deep neural networks for visual classification tasks in continual learning settings. Our method learns continually via the synchronization between distinct wake and sleep phases. During the wake phase, the model is exposed to sensory input and adapts its representations, ensuring stability through a dynamic parameter freezing mechanism and storing episodic memories in a short-term temporary memory (similarly to what happens in the hippocampus). During the sleep phase, the training process is split into NREM and REM stages. In the NREM stage, the model's synaptic weights are consolidated using replayed samples from the short-term and long-term memory and the synaptic plasticity mechanism is activated, strengthening important connections and weakening unimportant ones. In the REM stage, the model is exposed to previously-unseen realistic visual sensory experience, and the dreaming process is activated, which enables the model to explore the potential feature space, thus preparing synapses to future knowledge. We evaluate the effectiveness of our approach on three benchmark datasets: CIFAR-10, Tiny-ImageNet and FG-ImageNet. In all cases, our method outperforms the baselines and prior work, yielding a significant performance gain on continual visual classification tasks. Furthermore, we demonstrate the usefulness of all processing stages and the importance of dreaming to enable positive forward transfer.
- North America > United States (0.14)
- Europe > Italy (0.05)
- Europe > Austria (0.04)
- (2 more...)
- Research Report (0.64)
- Instructional Material (0.46)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Health & Medicine > Therapeutic Area > Sleep (0.94)
Weakly Supervised Continual Learning
Boschini, Matteo, Buzzega, Pietro, Bonicelli, Lorenzo, Porrello, Angelo, Calderara, Simone
Continual Learning (CL) investigates how to train Deep Networks on a stream of tasks without incurring catastrophic forgetting. CL settings proposed in the literature assume that every incoming example is paired with ground-truth annotations. However, this clashes with many real-world applications: gathering labeled data, which is in itself tedious and expensive, becomes indeed infeasible when data flow as a stream and must be consumed in real-time. This work explores Weakly Supervised Continual Learning (WSCL): here, only a small fraction of labeled input examples are shown to the learner. We assess how current CL methods (e.g.: EWC, LwF, iCaRL, ER, GDumb, DER) perform in this novel and challenging scenario, in which overfitting entangles forgetting. Subsequently, we design two novel WSCL methods which exploit metric learning and consistency regularization to leverage unsupervised data while learning. In doing so, we show that not only our proposals exhibit higher flexibility when supervised information is scarce, but also that less than 25% labels can be enough to reach or even outperform SOTA methods trained under full supervision.
- Europe > Italy > Emilia-Romagna > Modeno Province > Modena (0.04)
- Europe > Italy > Emilia-Romagna > Metropolitan City of Bologna > Bologna (0.04)