Projects per year
In this paper we generalise the sentence compression task. Rather than simply shorten a sentence by deleting words or constituents, as in previous work, we rewrite it using additional operations such as substitution, reordering, and insertion. We present a new corpus that is suited to our task and a discriminative tree-to-tree transduction model that can naturally account for structural and lexical mismatches. The model incorporates a novel grammar extraction method, uses a language model for coherent output, and can be easily tuned to a wide range of compression specific loss functions.
|Title of host publication||Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)|
|Publisher||Association for Computational Linguistics|
|Number of pages||8|
|Publication status||Published - 2008|