Goto

Collaborating Authors

 rmm




RMM: Reinforced Memory Management for Class-Incremental Learning

Neural Information Processing Systems

Class-Incremental Learning (CIL) [38] trains classifiers under a strict memory budget: in each incremental phase, learning is done for new data, most of which is abandoned to free space for the next phase. The preserved data are exemplars used for replaying. However, existing methods use a static and ad hoc strategy for memory allocation, which is often sub-optimal. In this work, we propose a dynamic memory management strategy that is optimized for the incremental phases and different object classes. We call our method reinforced memory management (RMM), leveraging reinforcement learning.


CoSP: Reconfigurable Multi-State Metamaterial Inverse Design via Contrastive Pretrained Large Language Model

Yang, Shujie, Zhao, Xuzhe, Zhang, Yuqi, Tang, Yansong, Dong, Kaichen

arXiv.org Artificial Intelligence

Metamaterials, known for their ability to manipulate light at subwavelength scales, face significant design challenges due to their complex and sophisticated structures. Consequently, deep learning has emerged as a powerful tool to streamline their design process. Reconfigurable multi-state metamaterials (RMMs) with adjustable parameters can switch their optical characteristics between different states upon external stimulation, leading to numerous applications. However, existing deep learning-based inverse design methods fall short in considering reconfigurability with multi-state switching. To address this challenge, we propose CoSP, an intelligent inverse design method based on contrastive pretrained large language model (LLM). By performing contrastive pretraining on multi-state spectrum, a well-trained spectrum encoder capable of understanding the spectrum is obtained, and it subsequently interacts with a pretrained LLM. This approach allows the model to preserve its linguistic capabilities while also comprehending Maxwell's Equations, enabling it to describe material structures with target optical properties in natural language. Our experiments demonstrate that CoSP can design corresponding thin-film metamaterial structures for arbitrary multi-state, multi-band optical responses, showing great potentials in the intelligent design of RMMs for versatile applications.


Towards Reversible Model Merging For Low-rank Weights

Alipour, Mohammadsajad, Amiri, Mohammad Mohammadi

arXiv.org Artificial Intelligence

Model merging aims to combine multiple fine-tuned models into a single set of weights that performs well across all source tasks. While prior work has shown that merging can approximate the performance of individual fine-tuned models for each task, it largely overlooks scenarios where models are compressed into low-rank representations, either through low-rank adaptation (LoRA) or post-training singular value decomposition (SVD). We first demonstrate that applying conventional merging methods to low-rank weights leads to severe performance degradation in the merged model. Motivated by this phenomenon, we propose a fundamentally different approach: instead of collapsing all adapters into one set of weights, we construct a compact basis (e.g., an equivalent of holding two or more models) from which original task-specific models can be recovered via linear combination. This reframes merging as generating a reconstruction-capable model space rather than producing a single merged model. Crucially, this allows us to ``revert'' to each individual model when needed, recognizing that no merged model can consistently outperform one specialized for its task. Building on this insight, we introduce our method, Reversible Model Merging (RMM), an efficient, data-free, and flexible method that provides a closed-form solution for selecting the optimal basis of model weights and task-specific coefficients for linear combination. Extensive experiments across diverse datasets and model scales demonstrate that RMM consistently outperforms existing merging approaches, preserving the performance of low-rank compressed models by a significant margin.


Deep learning the sources of MJO predictability: a spectral view of learned features

Yao, Lin, Yang, Da, Duncan, James P. C., Chattopadhyay, Ashesh, Hassanzadeh, Pedram, Bhimji, Wahid, Yu, Bin

arXiv.org Artificial Intelligence

The Madden-Julian oscillation (MJO) is a planetary-scale, intraseasonal tropical rainfall phenomenon crucial for global weather and climate; however, its dynamics and predictability remain poorly understood. Here, we leverage deep learning (DL) to investigate the sources of MJO predictability, motivated by a central difference in MJO theories: which spatial scales are essential for driving the MJO? We first develop a deep convolutional neural network (DCNN) to forecast the MJO indices (RMM and ROMI). Our model predicts RMM and ROMI up to 21 and 33 days, respectively, achieving skills comparable to leading subseasonal-to-seasonal models such as NCEP. To identify the spatial scales most relevant for MJO forecasting, we conduct spectral analysis of the latent feature space and find that large-scale patterns dominate the learned signals. Additional experiments show that models using only large-scale signals as the input have the same skills as those using all the scales, supporting the large-scale view of the MJO. Meanwhile, we find that small-scale signals remain informative: surprisingly, models using only small-scale input can still produce skillful forecasts up to 1-2 weeks ahead. We show that this is achieved by reconstructing the large-scale envelope of the small-scale activities, which aligns with the multi-scale view of the MJO. Altogether, our findings support that large-scale patterns--whether directly included or reconstructed--may be the primary source of MJO predictability.


RMM: Reinforced Memory Management for Class-Incremental Learning Supplementary Materials

Neural Information Processing Systems

This is supplementary to Section 5.2 " This is supplementary to Section 5.2 " This is supplementary to to Section 5.1 " This is supplementary to Section 5.2 " To evaluate the performance of our RMM in unknown scenarios, we supplemented the experiments of using the policy functions trained "in distinct numbers of phases" and "on different datasets" and show the testing results of CIFAR-100 in Table S5. ", using the policy learned on "ImageNet-Subset, For example, Row 5 is for training the policy on "ImageNet-Subset, This is supplementary to Section 5.2 " In Table S6, we can see the clear improvements, e.g., ImageNet-Subset [4]) are available in Table S7. No. Method Policy learned on N =5 N =10 N =25 1 Baseline - 49.02 44.59 38.23 2 w/ RMM ImageNet-Subset 53.15 50.05 42.89 This is supplementary to Section 5.2 " This is supplementary to Section 5.2 " We run our experiments using GPU workstations as follows, 4 No. Method Memory budget of exemplars N =5 N =10 N =25 1 Baseline 1000 64.31 60.97 58.77 2 w/ RMM (ours) 1000 68.20 65.57 Row 1 (baseline) is from the sota method POD-AANets [10].