Projects per year
Abstract
We investigate cross-lingual acoustic modelling for low resource languages using the subspace Gaussian mixture model (SGMM). We assume the presence of acoustic models trained on multiple source languages, and use the global subspace parameters from those models for improved modelling in a target language with limited amounts of transcribed speech. Experiments on the GlobalPhone corpus using Spanish, Portuguese, and Swedish as source languages and German as target language (with 1 hour and 5 hours of transcribed audio) show that multilingually trained SGMM shared parameters result in lower word error rates (WERs) than using those from a single source language. We also show that regularizing the estimation of the SGMM state vectors by penalizing their l1-norm help to overcome numerical instabilities and lead to lower WER.
Original language | English |
---|---|
Title of host publication | Automatic Speech Recognition and Understanding (ASRU), 2011 IEEE Workshop on |
Publisher | Institute of Electrical and Electronics Engineers |
Pages | 365-370 |
Number of pages | 6 |
ISBN (Electronic) | 978-1-4673-0366-8 |
ISBN (Print) | 978-1-4673-0365-1 |
DOIs | |
Publication status | Published - 2011 |
Keywords / Materials (for Non-textual outputs)
- Gaussian processes
- acoustic signal processing
- natural language processing
- speech recognition
- German
- GlobalPhone corpus
- Portuguese
- Spanish
- Swedish
- cross-lingual acoustic modelling
- cross-lingual speech recognition
- global subspace parameter
- low resource language
- regularized subspace Gaussian mixture model
- word error rates
- Acoustics
- Data models
- Estimation
- Hidden Markov models
- Speech recognition
- Training data
- Vectors
Fingerprint
Dive into the research topics of 'Regularized subspace Gaussian mixture models for cross-lingual speech recognition'. Together they form a unique fingerprint.Projects
- 2 Finished