Brain-inspired continual pre-trained learner via silent synaptic consolidation
Ran, Xuming, Yao, Juntao, Wang, Yusong, Xu, Mingkun, Liu, Dianbo
–arXiv.org Artificial Intelligence
Pre-trained models have demonstrated impressive generalization capabilities, yet they remain vulnerable to catastrophic forgetting when incrementally trained on new tasks. Existing architecture-based strategies encounter two primary challenges: 1) Integrating a pre-trained network with a trainable sub-network complicates the delicate balance between learning plasticity and memory stability across evolving tasks during learning. In this study, we introduce the Artsy, inspired by the activation mechanisms of silent synapses via spike-timing-dependent plasticity observed in mature brains, to enhance the continual learning capabilities of pre-trained models. The Artsy integrates two key components: During training, the Artsy mimics mature brain dynamics by maintaining memory stability for previously learned knowledge within the pre-trained network while simultaneously promoting learning plasticity in task-specific sub-networks. During inference, artificial silent and functional synapses are utilized to establish precise connections between the pre-synaptic neurons in the pre-trained network and the post-synaptic neurons in the sub-networks, facilitated through synaptic consolidation, thereby enabling effective extraction of relevant information from test samples. Comprehensive experimental evaluations reveal that our model significantly outperforms conventional methods on class-incremental learning tasks, while also providing enhanced biological interpretability for architecture-based approaches. Moreover, we propose that the Artsy offers a promising avenue for simulating biological synaptic mechanisms, potentially advancing our understanding of neural plasticity in both artificial and biological systems. Pre-trained artificial neural networks have demonstrated notable generalization capabilities; however, they are prone to catastrophic forgetting when exposed to sequential training on new datasets, as outlined in previous studies Wang et al. (2024).
arXiv.org Artificial Intelligence
Oct-8-2024
- Country:
- Asia (0.68)
- North America > United States
- Washington > King County > Seattle (0.14)
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Technology: