Insights into Pre-training via Simpler Synthetic Tasks Yuhuai Wu
–Neural Information Processing Systems
Pre-training produces representations that are effective for a wide range of downstream tasks, but it is still unclear what properties of pre-training are necessary for effective gains. Notably, recent work shows that even pre-training on synthetic tasks can achieve significant gains in downstream tasks. In this work, we perform three experiments that iteratively simplify pre-training and show that the simplifications still retain much of its gains.
Neural Information Processing Systems
Nov-15-2025, 09:48:02 GMT
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Asia > China
- Hong Kong (0.04)
- Europe
- France > Provence-Alpes-Côte d'Azur
- Bouches-du-Rhône > Marseille (0.04)
- Germany > Berlin (0.04)
- France > Provence-Alpes-Côte d'Azur
- North America
- Canada (0.04)
- Dominican Republic (0.04)
- United States > California
- Santa Clara County > Palo Alto (0.04)
- Africa > Ethiopia
- Genre:
- Research Report > New Finding (0.67)
- Industry:
- Information Technology > Security & Privacy (0.46)
- Technology: