Abstract
We prove that log-linearly interpolated backoff language models can be efficiently and exactly collapsed into a single normalized backoff model, contradicting Hsu (2007). While prior work reported that log-linear interpolation yields lower perplexity than linear interpolation, normalizing at query time was impractical. We normalize the model offline in advance, which is efficient due to a recurrence relationship between the normalizing factors. To tune interpolation weights, we apply Newton’s method to this convex problem and show that the derivatives can be computed efficiently in a batch process. These findings are combined in new open-source interpolationtool, which is distributed with KenLM. With 21 out-of-domain corpora,log-linear interpolation yields 72.58 perplexity on TED talks, compared to 75.91 for linear interpolation.
Original language | English |
---|---|
Title of host publication | Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics |
Place of Publication | Berlin, Germany |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 876-886 |
Number of pages | 11 |
ISBN (Print) | 978-1-945626-00-5 |
DOIs | |
Publication status | Published - 12 Aug 2016 |
Event | 54th Annual Meeting of the Association for Computational Linguistics - Berlin, Germany Duration: 7 Aug 2016 → 12 Aug 2016 https://mirror.aclweb.org/acl2016/ |
Conference
Conference | 54th Annual Meeting of the Association for Computational Linguistics |
---|---|
Abbreviated title | ACL 2016 |
Country/Territory | Germany |
City | Berlin |
Period | 7/08/16 → 12/08/16 |
Internet address |