Hot papers on arXiv from 2021
Reproduced under a CC BY 4.0 license. We've collated the most tweeted papers for each month that were uploaded onto arXiv during 2021. Results are powered by Arxiv Sanity Preserver. Abstract: Large-scale model training has been a playing ground for a limited few requiring complex model refactoring and access to prohibitively expensive GPU clusters. ZeRO-Offload changes the large model training landscape by making large model training accessible to nearly everyone.
Dec-27-2021, 10:25:24 GMT