Abstract
Abstractive summarisation is notoriously hard to evaluate since standard word-overlap-based metrics are biased towards specific words in the human reference. We introduce a new evaluation metric which abstracts away from the word-level and instead is based on fact-level content weighting, i.e. relating the facts of the document to the facts of the summary. We follow the assumption that a good summary will reflect all relevant facts, i.e. the ones present in the ground truth (human-generated reference summary). We confirm this hypothesis by showing that our weightings are highly correlated to human perception and compare favourably to the recent manual highlight-based metric of Hardy et al. (2019).
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics |
| Editors | Dan Jurafsky, Joyce Chai, Natalie Schluter, Joel Tetreault |
| Publisher | Association for Computational Linguistics |
| Pages | 5071-5081 |
| Number of pages | 11 |
| ISBN (Electronic) | 9781952148255 |
| DOIs | |
| Publication status | Published - Jul 2020 |
| Event | 58th Annual Meeting of the Association for Computational Linguistics 2020 - Virtual, Online, United States Duration: 5 Jul 2020 → 10 Jul 2020 |
Conference
| Conference | 58th Annual Meeting of the Association for Computational Linguistics 2020 |
|---|---|
| Abbreviated title | ACL 2020 |
| Country/Territory | United States |
| City | Virtual, Online |
| Period | 5/07/20 → 10/07/20 |
ASJC Scopus subject areas
- Computer Science Applications
- Linguistics and Language
- Language and Linguistics