ISCA Archive ICSLP 2000
ISCA Archive ICSLP 2000

Minimum Bayes error feature selection

George Saon, Mukund Padmanabhan

We consider the problem of designing a linear transformation Θ ∈ Rpxn, of rank p ≤ n, which projects the features of a classifier x ∈ Rn onto y = Θx ∈ p such as to achieve minimum Bayes error (or probability of misclassification). Two avenues will be explored: the first is to maximize the Θ-average divergence between the class densities and the second is to minimize the union Bhattacharyya bound in the range of Θ. While both approaches yield similar performance in practice, they outperform standard LDA features and show a 10% relative improvement in the word error rate over state-of-the-art cepstral features on a large vocabulary telephony speech recognition task.

Cite as: Saon, G., Padmanabhan, M. (2000) Minimum Bayes error feature selection. Proc. 6th International Conference on Spoken Language Processing (ICSLP 2000), vol. 3, 75-78

  author={George Saon and Mukund Padmanabhan},
  title={{Minimum Bayes error feature selection}},
  booktitle={Proc. 6th International Conference on Spoken Language Processing (ICSLP 2000)},
  pages={vol. 3, 75-78}