PROM: A Phrase-level Copying Mechanism with Pre-training for Abstractive Summarization
Ma, Xinbei, Gong, Yeyun, He, Pengcheng, Zhao, Hai, Duan, Nan
–arXiv.org Artificial Intelligence
Based on the remarkable achievements of pre-trained language models in abstractive summarization, the copying mechanism has proved helpful by improving the factuality, stability, and overall performance. This work proposes PROM, a new PhRase-level cOpying Mechanism that enhances attention on n-grams, which can be applied to zero-shot summarization with pre-training. PROM adds an indicator layer to explicitly pick up tokens in n-gram that can be copied from the source, and calculates an auxiliary loss for the copying prediction. Empirical studies show that PROM makes significant improvements in fine-tuning on benchmarks. In zero-shot setting, PROM is utilized in the self-supervised pre-training on raw corpora and provides new general baselines on a wide range of summarization datasets. Further analysis shows that PROM performs more reasonable copying and contributes to faithfulness.
arXiv.org Artificial Intelligence
May-11-2023
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Asia
- Afghanistan (0.04)
- China
- Middle East > Iraq (0.04)
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Denmark > Capital Region
- Copenhagen (0.04)
- Germany > Berlin (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Italy > Tuscany
- Florence (0.04)
- Portugal > Lisbon
- Lisbon (0.04)
- Sweden (0.05)
- Belgium > Brussels-Capital Region
- North America
- Canada > British Columbia
- Dominican Republic (0.04)
- United States
- Louisiana > Orleans Parish
- New Orleans (0.05)
- Minnesota > Hennepin County
- Minneapolis (0.28)
- Washington > King County
- Seattle (0.04)
- Louisiana > Orleans Parish
- Oceania > Australia
- South America > Ecuador (0.04)
- Africa > Ethiopia
- Genre:
- Research Report (0.64)
- Industry:
- Government (0.46)
- Information Technology (0.47)
- Law (0.69)
- Media > News (0.46)
- Technology: