Huff-LLM: End-to-End Lossless Compression for Efficient LLM Inference