Compact Language Models via Pruning and Knowledge Distillation
–Neural Information Processing Systems
Large language models (LLMs) targeting different deployment scales and sizes are currently produced by training each variant from scratch; this is extremely compute-intensive. In this paper, we investigate if pruning an existing LLM and then re-training it with a fraction (<3%) of the original training data can be a suitable alternative to repeated, full retraining.
Neural Information Processing Systems
Nov-16-2025, 23:45:03 GMT
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe
- North America > United States
- Virginia (0.04)
- Asia > Middle East
- Genre:
- Research Report
- Experimental Study (0.93)
- New Finding (0.68)
- Research Report
- Industry:
- Education (0.46)
- Technology: