snell
Expanding Sparse Tuning for Low Memory Usage
Parameter-efficient fine-tuning (PEFT) is an effective method for adapting pre-trained vision models to downstream tasks by tuning a small subset of parameters. Among PEFT methods, sparse tuning achieves superior performance by only adjusting the weights most relevant to downstream tasks, rather than densely tuning the whole weight matrix. However, this performance improvement has been accompanied by increases in memory usage, which stems from two factors, i.e., the storage of the whole weight matrix as learnable parameters in the optimizer and the additional storage of tunable weight indexes. In this paper, we propose a method named SNELL (Sparse tuning with kerNELized LoRA) for sparse tuning with low memory usage. To achieve low memory usage, SNELL decomposes the tunable matrix for sparsification into two learnable low-rank matrices, saving from the costly storage of the whole original matrix. A competition-based sparsification mechanism is further proposed to avoid the storage of tunable weight indexes. To maintain the effectiveness of sparse tuning with low-rank matrices, we extend the low-rank decomposition by applying nonlinear kernel functions to the whole-matrix merging. Consequently, we gain an increase in the rank of the merged matrix, enhancing the ability of SNELL in adapting the pre-trained models to downstream tasks. Extensive experiments on multiple downstream tasks show that SNELL achieves state-of-the-art performance with low memory usage, endowing PEFT with sparse tuning to large-scale models.
- North America > United States > California > Los Angeles County > Long Beach (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- (13 more...)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
- Information Technology > Artificial Intelligence > Representation & Reasoning (0.92)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.68)
- North America > United States > California > Los Angeles County > Long Beach (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- (13 more...)
Expanding Sparse Tuning for Low Memory Usage
Parameter-efficient fine-tuning (PEFT) is an effective method for adapting pre-trained vision models to downstream tasks by tuning a small subset of parameters. Among PEFT methods, sparse tuning achieves superior performance by only adjusting the weights most relevant to downstream tasks, rather than densely tuning the whole weight matrix. However, this performance improvement has been accompanied by increases in memory usage, which stems from two factors, i.e., the storage of the whole weight matrix as learnable parameters in the optimizer and the additional storage of tunable weight indexes. In this paper, we propose a method named SNELL (Sparse tuning with kerNELized LoRA) for sparse tuning with low memory usage. To achieve low memory usage, SNELL decomposes the tunable matrix for sparsification into two learnable low-rank matrices, saving from the costly storage of the whole original matrix.
Expanding Sparse Tuning for Low Memory Usage
Shen, Shufan, Sun, Junshu, Ji, Xiangyang, Huang, Qingming, Wang, Shuhui
Parameter-efficient fine-tuning (PEFT) is an effective method for adapting pre-trained vision models to downstream tasks by tuning a small subset of parameters. Among PEFT methods, sparse tuning achieves superior performance by only adjusting the weights most relevant to downstream tasks, rather than densely tuning the whole weight matrix. However, this performance improvement has been accompanied by increases in memory usage, which stems from two factors, i.e., the storage of the whole weight matrix as learnable parameters in the optimizer and the additional storage of tunable weight indexes. In this paper, we propose a method named SNELL (Sparse tuning with kerNELized LoRA) for sparse tuning with low memory usage. To achieve low memory usage, SNELL decomposes the tunable matrix for sparsification into two learnable low-rank matrices, saving from the costly storage of the whole original matrix. A competition-based sparsification mechanism is further proposed to avoid the storage of tunable weight indexes. To maintain the effectiveness of sparse tuning with low-rank matrices, we extend the low-rank decomposition by applying nonlinear kernel functions to the whole-matrix merging. Consequently, we gain an increase in the rank of the merged matrix, enhancing the ability of SNELL in adapting the pre-trained models to downstream tasks. Extensive experiments on multiple downstream tasks show that SNELL achieves state-of-the-art performance with low memory usage, endowing PEFT with sparse tuning to large-scale models. Codes are available at https://github.com/ssfgunner/SNELL.
- North America > United States > California > Los Angeles County > Long Beach (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- (13 more...)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (0.92)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
Prototypical Fine-tuning: Towards Robust Performance Under Varying Data Sizes
Jin, Yiqiao, Wang, Xiting, Hao, Yaru, Sun, Yizhou, Xie, Xing
In this paper, we move towards combining large parametric models with non-parametric prototypical networks. We propose prototypical fine-tuning, a novel prototypical framework for fine-tuning pretrained language models (LM), which automatically learns a bias to improve predictive performance for varying data sizes, especially low-resource settings. Our prototypical fine-tuning approach can automatically adjust the model capacity according to the number of data points and the model's inherent attributes. Moreover, we propose four principles for effective prototype fine-tuning towards the optimal solution. Experimental results across various datasets show that our work achieves significant performance improvements under various low-resource settings, as well as comparable and usually better performances in high-resource scenarios.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Asia > Middle East > Jordan (0.05)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
Lack of diversity in AI development causes serious real-life harm for people of color
Every time you ask Alexa to turn on your lights or play a song, you're using AI. But AI is also put to work in more serious ways, like facial recognition software by law enforcement. Some critics say there's a troubling lack of diversity among those who create the programs, and that is causing serious harm for people of color. We're joined now by Angle Bush. ANGLE BUSH: Thank you for having me.
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.36)
- Law (0.35)
- Government (0.31)
Getting Started With AI? Consider These Simple Marketing Projects
Marketing teams are increasingly turning to artificial intelligence (AI) to improve results, with marketers investing over $227 million in AI-based technologies in 2018 alone, according to Statista. Yet many companies have yet to use AI or have just started exploring it recently. So what are the best entry-level AI projects in the CX/martech stack to establish success? "With the rise of low-code technologies, we're witnessing a Cambrian explosion of AI projects that are considered low-hanging fruit for marketers and customer experience professionals," said Jen Snell, vice president, product strategy and marketing, Intelligent Self-Service, Verint. "We're now seeing immense success with automated interactions at the start of customer journeys, which is an area where most marketers are focused."
AI (Artificial Intelligence) Lessons From The World Series
Snell, second from left, comes out of the game against the Dodgers in the 6th inning in Game 6 of the World Series at Globe Life Field on October 27, 2020 in Arlington, Texas. I'm a lifelong Dodgers fan and I waited for 32 years for the team to win another World Series. But during this period of time, the sport has certainly seen much change. With the availability of huge amounts of data, sophisticated computers and advanced analytics, the strategies have become increasingly based on the numbers. It seems that AI (Artificial Intelligence) has dominated the decision making process.
- North America > United States > Texas > Tarrant County > Arlington (0.25)
- North America > United States > Florida (0.07)
- North America > United States > California > Los Angeles County > Los Angeles (0.06)
- (2 more...)
AI (Artificial Intelligence) Lessons From The World Series
ARLINGTON, TEXAS - OCTOBER 27:Rays pitcher Blake Snell, second from left, comes out of the game ... [ ] against the Dodgers in the 6th inning in Game 6 of the World Series at Globe Life Field on October 27, 2020 in Arlington, Texas. I'm a lifelong Dodgers fan and I waited for 32 years for the team to win another World Series. But during this period of time, the sport has certainly seen much change. With the availability of huge amounts of data, sophisticated computers and advanced analytics, the strategies have become increasingly based on the numbers. It seems that AI (Artificial Intelligence) has dominated the decision making process.
- North America > United States > Texas > Tarrant County > Arlington (0.45)
- North America > United States > Florida (0.07)
- North America > United States > California > Los Angeles County > Los Angeles (0.06)
- (2 more...)