Insights into Pre-training via Simpler Synthetic Tasks Yuhuai Wu

Neural Information Processing Systems 

Pre-training produces representations that are effective for a wide range of downstream tasks, but it is still unclear what properties of pre-training are necessary for effective gains. Notably, recent work shows that even pre-training on synthetic tasks can achieve significant gains in downstream tasks. In this work, we perform three experiments that iteratively simplify pre-training and show that the simplifications still retain much of its gains.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found