Review for NeurIPS paper: Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping
–Neural Information Processing Systems
Summary and Contributions: This paper proposes to accelerate training of Transformer networks by progressively reducing Transformer layers from the network during training. First, it compares two different architectures of BERT, PostLN and PreLN. PostLN applies layer normalization after the element-wise addition in Transformer blocks. The PreLN changes the placement of the location of layer normalization by placing it only on the input stream of the sublayers. It finds that PostLN is more sensitive to the choice of hyperparameters, and training often diverges with more aggressive learning rates whereas PreLN avoids vanishing gradients and leads to more stable optimization.
Neural Information Processing Systems
Jan-27-2025, 04:06:59 GMT
- Technology: