Text Simplification with Sentence Embeddings
–arXiv.org Artificial Intelligence
Sentence embeddings can be decoded to give approximations of the original texts used to create them. We explore this effect in the context of text simplification, demonstrating that reconstructed text embeddings preserve complexity levels. We experiment with a small feed forward neural network to effectively learn a transformation between sentence embeddings representing high-complexity and low-complexity texts. We provide comparison to a Seq2Seq and LLM-based approach, showing encouraging results in our much smaller learning setting. Finally, we demonstrate the applicability of our transformation to an unseen simplification dataset (MedEASI), as well as datasets from languages outside the training data (ES,DE). We conclude that learning transformations in sentence embedding space is a promising direction for future research and has potential to unlock the ability to develop small, but powerful models for text simplification and other natural language generation tasks.
arXiv.org Artificial Intelligence
Oct-29-2025
- Country:
- Asia
- China > Hong Kong (0.04)
- Middle East
- Singapore (0.04)
- Atlantic Ocean > North Atlantic Ocean
- Baltic Sea (0.04)
- Europe
- France > Provence-Alpes-Côte d'Azur
- Bouches-du-Rhône > Marseille (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Italy > Tuscany
- Florence (0.04)
- Netherlands (0.04)
- Ukraine (0.04)
- France > Provence-Alpes-Côte d'Azur
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- Mexico > Mexico City
- Mexico City (0.04)
- United States > Oregon
- Multnomah County > Portland (0.04)
- Canada > Ontario
- Asia
- Genre:
- Research Report > New Finding (0.68)
- Technology: