Title
Kernel methods match Deep Neural Networks on TIMIT
Abstract
Despite their theoretical appeal and grounding in tractable convex optimization techniques, kernel methods are often not the first choice for large-scale speech applications due to their significant memory requirements and computational expense. In recent years, randomized approximate feature maps have emerged as an elegant mechanism to scale-up kernel methods. Still, in practice, a large number of random features is required to obtain acceptable accuracy in predictive tasks. In this paper, we develop two algorithmic schemes to address this computational bottleneck in the context of kernel ridge regression. The first scheme is a specialized distributed block coordinate descent procedure that avoids the explicit materialization of the feature space data matrix, while the second scheme gains efficiency by combining multiple weak random feature models in an ensemble learning framework. We demonstrate that these schemes enable kernel methods to match the performance of state of the art Deep Neural Networks on TIMIT for speech recognition and classification tasks. In particular, we obtain the best classification error rates reported on TIMIT using kernel methods.
Year
DOI
Venue
2014
10.1109/ICASSP.2014.6853587
ICASSP
Keywords
Field
DocType
optimisation,deep neural networks,kernel ridge regression,speech recognition,kernel methods,timit,ensemble learning framework,large-scale speech applications,learning (artificial intelligence),regression analysis,feature space data matrix,speech classification tasks,tractable convex optimization techniques,deep learning,random features,multiple weak random feature models,large-scale kernel machines,distributed computing,specialized distributed block coordinate descent procedure,randomized approximate feature maps,neural nets,hidden markov models,computational modeling,kernel,neural networks,training data,learning artificial intelligence
Graph kernel,TIMIT,Radial basis function kernel,Pattern recognition,Kernel embedding of distributions,Computer science,Tree kernel,Polynomial kernel,Artificial intelligence,Kernel method,Ensemble learning,Machine learning
Conference
ISSN
Citations 
PageRank 
1520-6149
31
1.25
References 
Authors
11
5
Name
Order
Citations
PageRank
Po-Sen Huang192644.01
Avron, Haim231628.52
Tara N. Sainath33497232.43
Vikas Sindhwani43423154.85
Bhuvana Ramabhadran51779153.83