Support Vector Machines (SVMs) have recently proved to be powerful pattern classification tools with a strong connection to statistical learning theory. One of the hurdles to using SVMs in speech recognition, and a crucial aspect of SVM design in general, is the choice of the kernel function for non-separable data, and the setting of its parameters. This is often based on experience or a potentially costly search. This paper gives some experimental justification for the Fisher kernels proposed in [1]; kernels are obtained and their extra regularisation and use of labelled and unlabelled data discussed. Fisher kernels are derived from generative probability models of the data, and are a first step to implementing kernels for variable length sequences.
T. Jaakkola and D. Haussler. Exploiting generative models in discriminative classifiers. In M.S. Kearns, S.A. Solla, and D.A. Cohn, editors, Advances in Neural Information Processing Systems 11. MIT Press, 1999.
Cite as: Smith, N., Niranjan, M. (2000) Data-dependent kernels in svm classification of speech patterns. Proc. 6th International Conference on Spoken Language Processing (ICSLP 2000), vol. 1, 297-300, doi: 10.21437/ICSLP.2000-74
@inproceedings{smith00_icslp, author={Nathan Smith and Mahesan Niranjan}, title={{Data-dependent kernels in svm classification of speech patterns}}, year=2000, booktitle={Proc. 6th International Conference on Spoken Language Processing (ICSLP 2000)}, pages={vol. 1, 297-300}, doi={10.21437/ICSLP.2000-74} }