Insights into Pre-training via Simpler Synthetic Tasks Yuhuai Wu
–Neural Information Processing Systems
Pre-training produces representations that are effective for a wide range of downstream tasks, but it is still unclear what properties of pre-training are necessary for effective gains. Notably, recent work shows that even pre-training on synthetic tasks can achieve significant gains in downstream tasks. In this work, we perform three experiments that iteratively simplify pre-training and show that the simplifications still retain much of its gains.
Neural Information Processing Systems
Aug-16-2025, 17:42:40 GMT
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Asia > China
- Hong Kong (0.04)
- Europe
- France > Provence-Alpes-Côte d'Azur
- Bouches-du-Rhône > Marseille (0.04)
- Germany > Berlin (0.04)
- France > Provence-Alpes-Côte d'Azur
- North America
- Canada (0.04)
- Dominican Republic (0.04)
- United States > California
- Santa Clara County > Palo Alto (0.04)
- Africa > Ethiopia
- Genre:
- Research Report > New Finding (0.67)
- Industry:
- Information Technology (0.46)
- Technology: