Abstract / Description of output
Tree substitution grammars (TSGs) are a compelling alternative to context-free grammars for modelling syntax. However, many popular techniques for estimating weighted TSGs (under the moniker of Data Oriented Parsing) suffer from the problems of inconsistency and over-fitting. We present a theoretically principled model which solves these problems using a Bayesian non-parametric formulation. Our model learns compact and simple grammars, uncovering latent linguistic structures (e.g., verb subcategorisation), and in doing so far out-performs a standard PCFG.
Original language | English |
---|---|
Title of host publication | Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics |
Place of Publication | Boulder, Colorado |
Publisher | Association for Computational Linguistics |
Pages | 548-556 |
Number of pages | 9 |
ISBN (Print) | 978-1-932432-41-1 |
Publication status | Published - 1 Jun 2009 |