Ouroboros: On Accelerating Training of Transformer-Based Language Models