6d538a6e667960b168d3d947eb6207a6-Paper-Conference.pdf

Neural Information Processing Systems 

Prior work tries to improve the sampling locality by enforcing all the training jobs loading the same dataset in the same order and pace. However, such a solution isonly efficient under strong constraints: alljobs are trained onthe same dataset with the same starting moment and training speed. In this paper, we propose a new data loading method for efficiently training parallel DNNs with much flexible constraints. Our method is still highly efficient when different training jobs use different but overlapped datasets and have different starting moments andtrainingspeeds.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found