On Affine Homotopy between Language Encoders Robin S. M. Chan

Neural Information Processing Systems 

Pre-trained language encoders--functions that represent text as vectors--are an integral component of many NLP tasks. We tackle a natural question in language encoder analysis: What does it mean for two encoders to be similar? We contend that a faithful measure of similarity needs to be intrinsic, that is, task-independent, yet still be informative of extrinsic similarity--the performance on downstream tasks. It is common to consider two encoders similar if they are homotopic, i.e., if they can be aligned through some transformation.