MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization

Neural Information Processing Systems 

For example, we achieve a Wikitext2 perplexity of 5.95 on the LLaMA2-70B

Similar Docs  Excel Report  more

TitleSimilaritySource
None found