Goto

Collaborating Authors

 low-resource task



where we cannot manually access and annotate a lot of data, as well as for low-resource tasks in different languages

Neural Information Processing Systems

We thank all the reviewers for their time and insightful feedback about our work. Many of the recent few-shot learning works focus on computer vision compared to NLU tasks. We leverage self-training with several advances to bridge this gap. Similar baselines reported for active learning [Gal et al., 2017] and preference learning [Houlsby et al., UDA [Xie et al., 2019] and self-training with noisy student [Xie et al., 2020] show these techniques to work best with Additionally, for IMDB longer sequence length plays a big role. Sample mixing based on easy and hard examples is an interesting idea.


Using Temperature Sampling to Effectively Train Robot Learning Policies on Imbalanced Datasets

Patil, Basavasagar, Belt, Sydney, Lee, Jayjun, Fazeli, Nima, Bucher, Bernadette

arXiv.org Artificial Intelligence

Increasingly large datasets of robot actions and sensory observations are being collected to train ever-larger neural networks. These datasets are collected based on tasks and while these tasks may be distinct in their descriptions, many involve very similar physical action sequences (e.g., 'pick up an apple' versus'pick up an orange'). As a result, many datasets of robotic tasks are substantially imbalanced in terms of the physical robotic actions they represent. In this work, we propose a simple sampling strategy for policy training that mitigates this imbalance. Our method requires only a few lines of code to integrate into existing code-bases and improves generalization. We evaluate our method in both pre-training small models and fine-tuning large foundational models. Our results show substantial improvements on low-resource tasks compared to prior state-of-the-art methods, without degrading performance on high-resource tasks. This enables more effective use of model capacity for multi-task policies.


Order Matters in the Presence of Dataset Imbalance for Multilingual Learning

Neural Information Processing Systems

We present a simple yet effective method of pre-training on high-resource tasks, followed by fine-tuning on a mixture of high/low-resource tasks. We provide a thorough empirical study and analysis of this method's benefits


where we cannot manually access and annotate a lot of data, as well as for low-resource tasks in different languages

Neural Information Processing Systems

We thank all the reviewers for their time and insightful feedback about our work. Many of the recent few-shot learning works focus on computer vision compared to NLU tasks. We leverage self-training with several advances to bridge this gap. Similar baselines reported for active learning [Gal et al., 2017] and preference learning [Houlsby et al., UDA [Xie et al., 2019] and self-training with noisy student [Xie et al., 2020] show these techniques to work best with Additionally, for IMDB longer sequence length plays a big role. Sample mixing based on easy and hard examples is an interesting idea.


Enhancing Cross-task Transfer of Large Language Models via Activation Steering

Tang, Xinyu, Lv, Zhihao, Cheng, Xiaoxue, Li, Junyi, Zhao, Wayne Xin, Wen, Zujie, Zhang, Zhiqiang, Zhou, Jun

arXiv.org Artificial Intelligence

Large language models (LLMs) have shown impressive abilities in leveraging pretrained knowledge through prompting, but they often struggle with unseen tasks, particularly in data-scarce scenarios. While cross-task in-context learning offers a direct solution for transferring knowledge across tasks, it still faces critical challenges in terms of robustness, scalability, and efficiency. In this paper, we investigate whether cross-task transfer can be achieved via latent space steering without parameter updates or input expansion. Through an analysis of activation patterns in the latent space of LLMs, we observe that the enhanced activations induced by in-context examples have consistent patterns across different tasks. Inspired by these findings, we propose CAST, a novel Cross-task Activation Steering Transfer framework that enables effective transfer by manipulating the model's internal activation states. Our approach first selects influential and diverse samples from high-resource tasks, then utilizes their contrastive representation-enhanced activations to adapt LLMs to low-resource tasks. Extensive experiments across both cross-domain and cross-lingual transfer settings show that our method outperforms competitive baselines and demonstrates superior scalability and lower computational costs.


Token-Efficient Leverage Learning in Large Language Models

Zeng, Yuanhao, Wang, Min, Wang, Yihang, Shao, Yingxia

arXiv.org Artificial Intelligence

Large Language Models (LLMs) have excelled in various tasks but perform better in high-resource scenarios, which presents challenges in low-resource scenarios. Data scarcity and the inherent difficulty of adapting LLMs to specific tasks compound the challenge. To address the twin hurdles, we introduce \textbf{Leverage Learning}. We present a streamlined implement of this methodology called Token-Efficient Leverage Learning (TELL). TELL showcases the potential of Leverage Learning, demonstrating effectiveness across various LLMs and low-resource tasks, ranging from $10^4$ to $10^6$ tokens. It reduces task data requirements by up to nearly an order of magnitude compared to conventional Supervised Fine-Tuning (SFT) while delivering competitive performance. With the same amount of task data, TELL leads in improving task performance compared to SFT. We discuss the mechanism of Leverage Learning, suggesting it aligns with quantization hypothesis and explore its promising potential through empirical testing.


Order Matters in the Presence of Dataset Imbalance for Multilingual Learning

Choi, Dami, Xin, Derrick, Dadkhahi, Hamid, Gilmer, Justin, Garg, Ankush, Firat, Orhan, Yeh, Chih-Kuan, Dai, Andrew M., Ghorbani, Behrooz

arXiv.org Artificial Intelligence

In this paper, we empirically study the optimization dynamics of multi-task learning, particularly focusing on those that govern a collection of tasks with significant data imbalance. We present a simple yet effective method of pre-training on high-resource tasks, followed by fine-tuning on a mixture of high/low-resource tasks. We provide a thorough empirical study and analysis of this method's benefits showing that it achieves consistent improvements relative to the performance trade-off profile of standard static weighting. We analyze under what data regimes this method is applicable and show its improvements empirically in neural machine translation (NMT) and multi-lingual language modeling.


Toward Efficient Language Model Pretraining and Downstream Adaptation via Self-Evolution: A Case Study on SuperGLUE

Zhong, Qihuang, Ding, Liang, Zhan, Yibing, Qiao, Yu, Wen, Yonggang, Shen, Li, Liu, Juhua, Yu, Baosheng, Du, Bo, Chen, Yixin, Gao, Xinbo, Miao, Chunyan, Tang, Xiaoou, Tao, Dacheng

arXiv.org Artificial Intelligence

This technical report briefly describes our JDExplore d-team's Vega v2 submission on the SuperGLUE leaderboard. SuperGLUE is more challenging than the widely used general language understanding evaluation (GLUE) benchmark, containing eight difficult language understanding tasks, including question answering, natural language inference, word sense disambiguation, coreference resolution, and reasoning. [Method] Instead of arbitrarily increasing the size of a pretrained language model (PLM), our aim is to 1) fully extract knowledge from the input pretraining data given a certain parameter budget, e.g., 6B, and 2) effectively transfer this knowledge to downstream tasks. To achieve goal 1), we propose self-evolution learning for PLMs to wisely predict the informative tokens that should be masked, and supervise the masked language modeling (MLM) process with rectified smooth labels. For goal 2), we leverage the prompt transfer technique to improve the low-resource tasks by transferring the knowledge from the foundation model and related downstream tasks to the target task. [Results] According to our submission record (Oct. 2022), with our optimized pretraining and fine-tuning strategies, our 6B Vega method achieved new state-of-the-art performance on 4/8 tasks, sitting atop the SuperGLUE leaderboard on Oct. 8, 2022, with an average score of 91.3.