Abstract
This paper investigates semi-supervised methods for discriminative language modeling, whereby n-best lists are “hallucinated” for given reference text and are then used for training n-gram language models using the perceptron algorithm. We perform controlled experiments on a very strong baseline English CTS system, comparing three methods for simulating ASR output, and compare the results with training with “real” n-best list output from the baseline recognizer. We find that methods based on extracting phrasal cohorts - similar to methods from machine translation for extracting phrase tables - yielded the largest gains of our three methods, achieving over half of the WER reduction of the fully supervised methods.
Original language | English |
---|---|
Title of host publication | 2012 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2012, Kyoto, Japan, March 25-30, 2012 |
Publisher | Institute of Electrical and Electronics Engineers (IEEE) |
Pages | 5001-5004 |
Number of pages | 4 |
ISBN (Electronic) | 978-1-4673-0044-5 |
ISBN (Print) | 978-1-4673-0045-2 |
DOIs | |
Publication status | Published - 2012 |