Discourse Structure Extraction from Pre-Trained and Fine-Tuned Language Models in Dialogues
Li, Chuyuan, Huber, Patrick, Xiao, Wen, Amblard, Maxime, Braud, Chloé, Carenini, Giuseppe
–arXiv.org Artificial Intelligence
Discourse processing suffers from data sparsity, especially for dialogues. As a result, we explore approaches to build discourse structures for dialogues, based on attention matrices from Pre-trained Language Models (PLMs). We investigate multiple tasks for fine-tuning and show that the dialogue-tailored Sentence Ordering task performs best. To locate and exploit discourse information in PLMs, we propose an unsupervised and a semi-supervised method. Our proposals achieve encouraging results on the STAC corpus, with F1 scores of 57.2 and 59.3 for unsupervised and semi-supervised methods, respectively. When restricted to projective trees, our scores improved to 63.3 and 68.1.
arXiv.org Artificial Intelligence
Jun-25-2023
- Country:
- Asia
- China
- India > Maharashtra
- Mumbai (0.04)
- Japan > Honshū
- Kansai > Osaka Prefecture > Osaka (0.04)
- Taiwan > Taiwan Province
- Taipei (0.04)
- Europe
- Czechia > Prague (0.04)
- France
- Grand Est > Meurthe-et-Moselle
- Nancy (0.04)
- Occitanie > Haute-Garonne
- Toulouse (0.04)
- Grand Est > Meurthe-et-Moselle
- Germany > Berlin (0.04)
- Italy > Tuscany
- Florence (0.04)
- Portugal > Lisbon
- Lisbon (0.04)
- Slovenia (0.04)
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- North America
- Canada > British Columbia
- Dominican Republic (0.04)
- United States
- California > San Diego County
- San Diego (0.04)
- Maryland > Baltimore (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- California > San Diego County
- Oceania > Australia
- Asia
- Genre:
- Research Report (0.82)
- Technology: