[1605.06465] Swapout: Learning an ensemble of deep architectures (samples from dropout, ResNets, stochastic depth) • /r/MachineLearning

@machinelearnbot 

Remarkably, our 32 layer wider model performs similar to a 1001 layer ResNet model. Could this mean potentially ResNet like accuracy, available on less advanced infrastructure (e.g. This is parameter number dependent, ResNets have few parameters. Authors of this paper show that it is competitive with ResNets using (approximately) the same number of parameters, though. It is a poor advertising for a good paper.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found