Set-Theoretic Compositionality of Sentence Embeddings
Bansal, Naman, mahajan, Yash, Sinha, Sanjeev, Karmaker, Santu
–arXiv.org Artificial Intelligence
Sentence encoders play a pivotal role in various NLP tasks; hence, an accurate evaluation of their compositional properties is paramount. However, existing evaluation methods predominantly focus on goal task-specific performance. This leaves a significant gap in understanding how well sentence embeddings demonstrate fundamental compositional properties in a task-independent context. Leveraging classical set theory, we address this gap by proposing six criteria based on three core "set-like" compositions/operations: \textit{TextOverlap}, \textit{TextDifference}, and \textit{TextUnion}. We systematically evaluate $7$ classical and $9$ Large Language Model (LLM)-based sentence encoders to assess their alignment with these criteria. Our findings show that SBERT consistently demonstrates set-like compositional properties, surpassing even the latest LLMs. Additionally, we introduce a new dataset of ~$192$K samples designed to facilitate future benchmarking efforts on set-like compositionality of sentence embeddings.
arXiv.org Artificial Intelligence
Feb-28-2025
- Country:
- Asia (1.00)
- Europe (1.00)
- North America > United States
- Minnesota > Hennepin County > Minneapolis (0.14)
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Government (0.96)
- Health & Medicine (0.93)
- Law (0.67)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.46)
- Leisure & Entertainment > Sports
- Olympic Games (0.46)
- Transportation
- Air (0.68)
- Infrastructure & Services (0.46)
- Technology: