Rare Gems: Finding Lottery Tickets at Initialization Kartik Sreenivasan
–Neural Information Processing Systems
Large neural networks can be pruned to a small fraction of their original size, with little loss in accuracy, by following a time-consuming "train, prune, re-train" approach. Frankle & Carbin [9] conjecture that we can avoid this by training lottery tickets, i.e., special sparse subnetworks found at initialization, that can be trained to high accuracy. However, a subsequent line of work [11, 41] presents concrete evidence that current algorithms for finding trainable networks at initialization, fail simple baseline comparisons, e.g., against training random sparse subnetworks. Finding lottery tickets that train to better accuracy compared to simple baselines remains an open problem.
Neural Information Processing Systems
Mar-22-2025, 22:54:10 GMT
- Genre:
- Contests & Prizes (1.00)
- Research Report > New Finding (0.46)
- Industry:
- Leisure & Entertainment > Gambling (1.00)
- Technology: