MiniPLM: Knowledge Distillation for Pre-Training Language Models

Open in new window