Challenging Assumptions in Learning Generic Text Style Embeddings
Ostheimer, Phil, Kloft, Marius, Fellenz, Sophie
–arXiv.org Artificial Intelligence
Recent advancements in language representation learning primarily emphasize language modeling for deriving meaningful representations, often neglecting style-specific considerations. This study addresses this gap by creating generic, sentence-level style embeddings crucial for style-centric tasks. Our approach is grounded on the premise that low-level text style changes can compose any high-level style. We hypothesize that applying this concept to representation learning enables the development of versatile text style embeddings. By fine-tuning a general-purpose text encoder using contrastive learning and standard cross-entropy loss, we aim to capture these low-level style shifts, anticipating that they offer insights applicable to high-level text styles. The outcomes prompt us to reconsider the underlying assumptions as the results do not always show that the learned style representations capture high-level text styles.
arXiv.org Artificial Intelligence
Jan-27-2025
- Country:
- Asia > Middle East
- Qatar (0.14)
- North America > United States
- Louisiana (0.15)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Asia > Middle East
- Genre:
- Research Report (1.00)
- Technology: