Token-Scaled Logit Distillation for Ternary Weight Generative Language Models Minsoo Kim 1 Sihwa Lee 1 Janghwan Lee

Neural Information Processing Systems 

Generative Language Models (GLMs) have shown impressive performance in tasks such as text generation, understanding, and reasoning. However, the large model size poses challenges for practical deployment.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found