Unified Embedding: Battle-Tested Feature Representations for Web-Scale ML Systems

Neural Information Processing Systems 

Learning high-quality feature embeddings efficiently and effectively is critical for the performance of web-scale machine learning systems. The standard approach is to represent each feature value as a d -dimensional embedding, which introduces hundreds of billions of parameters for extremely high-cardinality features. This bottleneck has led to substantial progress in alternative embedding algorithms. Many of these methods, however, make the assumption that each feature uses an independent embedding table. This work introduces a simple yet highly effective framework, Feature Multiplexing, where one single representation space is used for many different categorical features.