Abstract
We validated various novel and recently proposed methods for statistical machine translation on 10 language pairs, using large data resources. We saw gains from optimizing parameters, training with sparse features, the operation sequence model, and domain adaptation techniques. We also report on utilizing a huge language model trained on 126 billion tokens.
Original language | English |
---|---|
Title of host publication | Proceedings of the Eighth Workshop on Statistical Machine Translation |
Place of Publication | Sofia, Bulgaria |
Publisher | Association for Computational Linguistics |
Pages | 114-121 |
Number of pages | 8 |
Publication status | Published - 1 Aug 2013 |