Inductive biases of multi-task learning and finetuning: multiple regimes of feature reuse

Neural Information Processing Systems 

Neural networks are often trained on multiple tasks, either simultaneously (multi-task learning, MTL) or sequentially (pretraining and subsequent finetuning, PT+FT). In particular, it is common practice to pretrain neural networks on a large auxiliary task before finetuning on a downstream task with fewer samples. Despite the prevalence of this approach, the inductive biases that arise from learning multiple tasks are poorly characterized. In this work, we address this gap.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found