LangVAE and LangSpace: Building and Probing for Language Model VAEs
Carvalho, Danilo S., Zhang, Yingji, Unsworth, Harriet, Freitas, André
–arXiv.org Artificial Intelligence
We present LangVAE, a novel framework for modular construction of variational autoencoders (VAEs) on top of pre-trained large language models (LLMs). Such language model VAEs can encode the knowledge of their pre-trained components into more compact and semantically disentangled representations. The representations obtained in this way can be analysed with the LangVAE companion framework: LangSpace, which implements a collection of probing methods, such as vector traversal and interpolation, disentanglement measures, and cluster visualisations. LangVAE and LangSpace offer a flexible, efficient and scalable way of building and analysing textual representations, with simple integration for models available on the HuggingFace Hub. Additionally, we conducted a set of experiments with different encoder and decoder combinations, as well as annotated inputs, revealing a wide range of interactions across architectural families and sizes w.r.t. generalisation and disentanglement. Our findings demonstrate a promising framework for systematising the experimentation and understanding of textual representations.
arXiv.org Artificial Intelligence
May-2-2025
- Country:
- Asia
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.04)
- Singapore (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- Middle East > UAE
- Europe
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Switzerland (0.04)
- United Kingdom > England
- Greater Manchester > Manchester (0.04)
- Ireland > Leinster
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- United States > Minnesota
- Hennepin County > Minneapolis (0.14)
- Canada > Ontario
- Asia
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Education > Educational Setting (0.47)
- Technology: