Unsupervised Summarization Re-ranking
Ravaut, Mathieu, Joty, Shafiq, Chen, Nancy
–arXiv.org Artificial Intelligence
With the rise of task-specific pre-training objectives, abstractive summarization models like PEGASUS offer appealing zero-shot performance on downstream summarization tasks. However, the performance of such unsupervised models still lags significantly behind their supervised counterparts. Similarly to the supervised setup, we notice a very high variance in quality among summary candidates from these models while only one candidate is kept as the summary output. In this paper, we propose to re-rank summary candidates in an unsupervised manner, aiming to close the performance gap between unsupervised and supervised models. Our approach improves the unsupervised PEGASUS by up to 7.27% and ChatGPT by up to 6.86% relative mean ROUGE across four widely-adopted summarization benchmarks ; and achieves relative gains of 7.51% (up to 23.73% from XSum to WikiHow) averaged over 30 zero-shot transfer setups (finetuning on a dataset, evaluating on another).
arXiv.org Artificial Intelligence
May-26-2023
- Country:
- Asia > Middle East
- Republic of Türkiye (0.92)
- Europe (1.00)
- North America > United States (1.00)
- Asia > Middle East
- Genre:
- Personal (0.93)
- Research Report (1.00)
- Industry:
- Energy > Oil & Gas
- Upstream (1.00)
- Government
- Health & Medicine (1.00)
- Information Technology (0.67)
- Law (1.00)
- Energy > Oil & Gas
- Technology: