Kernel Approximation Methods for Speech Recognition - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Article Dans Une Revue Journal of Machine Learning Research Année : 2019

Kernel Approximation Methods for Speech Recognition

Résumé

We study the performance of kernel methods on the acoustic modeling task for automatic speech recognition, and compare their performance to deep neural networks (DNNs). To scale the kernel methods to large data sets, we use the random Fourier feature method of Rahimi and Recht (2007). We propose two novel techniques for improving the performance of kernel acoustic models. First, we propose a simple but effective feature selection method which reduces the number of random features required to attain a fixed level of performance. Second, we present a number of metrics which correlate strongly with speech recognition performance when computed on the heldout set; we attain improved performance by using these metrics to decide when to stop training. Additionally, we show that the linear bottleneck method of Sainath et al. (2013a) improves the performance of our kernel models significantly, in addition to speeding up training and making the models more compact. Leveraging these three methods, the kernel methods attain token error rates between 0.5% better and 0.1% worse than fully-connected DNNs across four speech recognition data sets, including the TIMIT and Broadcast News benchmark tasks.
Fichier principal
Vignette du fichier
jmlr19.pdf (538.76 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-02166422 , version 1 (26-06-2019)

Identifiants

  • HAL Id : hal-02166422 , version 1

Citer

Avner May, Alireza Bagheri Garakani, Zhiyun Lu, Dong Guo, Kuan Liu, et al.. Kernel Approximation Methods for Speech Recognition. Journal of Machine Learning Research, 2019, 20, pp.1 - 36. ⟨hal-02166422⟩
83 Consultations
161 Téléchargements

Partager

Gmail Facebook X LinkedIn More