erging
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Asia > Malaysia (0.04)
- Asia > China > Heilongjiang Province > Harbin (0.04)
- Asia > China > Fujian Province > Xiamen (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.93)
- North America > Canada > Ontario > Toronto (0.04)
- North America > United States > North Carolina (0.04)
- North America > Dominican Republic (0.04)
- (3 more...)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Asia > Malaysia (0.04)
- Asia > China > Heilongjiang Province > Harbin (0.04)
- Asia > China > Fujian Province > Xiamen (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.93)
- North America > Canada > Ontario > Toronto (0.04)
- North America > United States > North Carolina (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (4 more...)
Parameter Competition Balancing for Model Merging
Du, Guodong, Lee, Junlin, Li, Jing, Jiang, Runhua, Guo, Yifei, Yu, Shuyang, Liu, Hanting, Goh, Sim Kuan, Tang, Ho-Kin, He, Daojing, Zhang, Min
While fine-tuning pretrained models has become common practice, these models often underperform outside their specific domains. Recently developed model merging techniques enable the direct integration of multiple models, each fine-tuned for distinct tasks, into a single model. This strategy promotes multitasking capabilities without requiring retraining on the original datasets. However, existing methods fall short in addressing potential conflicts and complex correlations between tasks, especially in parameter-level adjustments, posing a challenge in effectively balancing parameter competition across various tasks. This paper introduces an innovative technique named PCB-Merging (Parameter Competition Balancing), a lightweight and training-free technique that adjusts the coefficients of each parameter for effective model merging. PCB-Merging employs intra-balancing to gauge parameter significance within individual tasks and inter-balancing to assess parameter similarities across different tasks. Parameters with low importance scores are dropped, and the remaining ones are rescaled to form the final merged model. We assessed our approach in diverse merging scenarios, including cross-task, cross-domain, and cross-training configurations, as well as out-of-domain generalization. The experimental results reveal that our approach achieves substantial performance enhancements across multiple modalities, domains, model sizes, number of tasks, fine-tuning forms, and large language models, outperforming existing model merging methods. The code is publicly available at: \url{https://github.com/duguodong7/pcb-merging}.
- Asia > Malaysia (0.04)
- Asia > China > Heilongjiang Province > Harbin (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Asia > China > Fujian Province > Xiamen (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.93)
TIES-Merging: Resolving Interference When Merging Models
Yadav, Prateek, Tam, Derek, Choshen, Leshem, Raffel, Colin, Bansal, Mohit
Transfer learning - i.e., further fine-tuning a pre-trained model on a downstream task - can confer significant advantages, including improved downstream performance, faster convergence, and better sample efficiency. These advantages have led to a proliferation of task-specific fine-tuned models, which typically can only perform a single task and do not benefit from one another. Recently, model merging techniques have emerged as a solution to combine multiple task-specific models into a single multitask model without performing additional training. However, existing merging methods often ignore the interference between parameters of different models, resulting in large performance drops when merging multiple models. In this paper, we demonstrate that prior merging techniques inadvertently lose valuable information due to two major sources of interference: (a) interference due to redundant parameter values and (b) disagreement on the sign of a given parameter's values across models. To address this, we propose our method, TRIM, ELECT SIGN & MERGE (TIES-Merging), which introduces three novel steps when merging models: (1) resetting parameters that only changed a small amount during fine-tuning, (2) resolving sign conflicts, and (3) merging only the parameters that are in alignment with the final agreed-upon sign. We find that TIES-Merging outperforms several existing methods in diverse settings covering a range of modalities, domains, number of tasks, model sizes, architectures, and fine-tuning settings. We further analyze the impact of different types of interference on model parameters, and highlight the importance of resolving sign interference. Our code is available at https://github.com/prateeky2806/ties-merging
- North America > Canada > Ontario > Toronto (0.04)
- North America > United States > North Carolina (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (4 more...)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Model-Based Reasoning (0.34)
- Information Technology > Artificial Intelligence > Machine Learning > Transfer Learning (0.34)