Models for Sentence Compression: A Comparison across Domains, Training Requirements and Evaluation Measures

James Clarke, Mirella Lapata

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Sentence compression is the task of producing a summary at the sentence
level. This paper focuses on three aspects of this task which have not received detailed treatment in the literature: training requirements, scalability, and automatic evaluation. We provide a novel comparison between a supervised constituent-based and an weakly supervised word-based compression algorithm and examine how these models port to different domains (written vs. spoken text). To achieve this, a human-authored compression corpus has been created and our study high-lights potential problems with the automatically gathered compression corpora currently used. Finally, we assess whether automatic evaluation measures can be used to determine compression quality.
Original languageEnglish
Title of host publicationACL 2006, 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference, Sydney, Australia, 17-21 July 2006
PublisherAssociation for Computational Linguistics
Pages377-384
Number of pages8
Publication statusPublished - 2006

Fingerprint

Dive into the research topics of 'Models for Sentence Compression: A Comparison across Domains, Training Requirements and Evaluation Measures'. Together they form a unique fingerprint.

Cite this