Abstract
We investigate the robustness and training dynamics of raw waveform acoustic models for automatic speech recognition (ASR). It is known that the first layer of such models learn a set of filters, performing a form of time-frequency analysis. This layer is liable to be under-trained owing to gradient vanishing, which can negatively affect the network performance. Through a set of experiments on TIMIT, Aurora-4 and WSJ datasets, we investigate the training dynamics of the first layer by measuring the evolution of its average frequency response over different epochs. We demonstrate that the network efficiently learns an optimal set of filters with a high spectral resolution and the dynamics of the first layer highly correlates with the dynamics of the cross entropy (CE) loss and word error rate (WER). In addition, we study the robustness of raw waveform models in both matched and mismatched conditions. The accuracy of these models is found to be comparable to, or better than, their MFCC-based counterparts in matched conditions and notably improved by using a better alignment. The role of raw waveform normalisation was also examined and up to 4.3% absolute WER reduction in mismatched conditions was achieved.
Original language | English |
---|---|
Title of host publication | Proceedings of Interspeech 2020 |
Publisher | International Speech Communication Association |
Pages | 1001-1005 |
Number of pages | 5 |
DOIs | |
Publication status | Published - 25 Oct 2020 |
Event | Interspeech 2020 - Virtual Conference, China Duration: 25 Oct 2020 → 29 Oct 2020 http://www.interspeech2020.org/ |
Publication series
Name | |
---|---|
ISSN (Electronic) | 1990-9772 |
Conference
Conference | Interspeech 2020 |
---|---|
Abbreviated title | INTERSPEECH 2020 |
Country/Territory | China |
City | Virtual Conference |
Period | 25/10/20 → 29/10/20 |
Internet address |
Keywords
- ASR
- acoustic modelling
- raw waveform
- training dynamics
- average frequency response