Soft-to-Hard Vector Quantization for End-to-End Learning Compressible Representations

Neural Information Processing Systems 

We present a new approach to learn compressible representations in deep architectures with an end-to-end training strategy. Our method is based on a soft (continuous) relaxation of quantization and entropy, which we anneal to their discrete counterparts throughout training.