ISCA Archive Eurospeech 1999
ISCA Archive Eurospeech 1999

Linguistic features for whole sentence maximum entropy language models

Xiaojin Zhu, Stanley F. Chen, Ronald Rosenfeld

Whole sentence maximum entropy models directly model the probability of a sentence using features – arbitrary computable properties of the sentence. We investigate whether linguistic features that capture the underlying linguistic structure of a sentence can improve modeling. We use a shallow parser to parse sentences into linguistic constituents in two corpora; one is the original training corpus, and the other is an artificial corpus generated from an initial trigram model. We define three sets of candidate linguistic features based on these constituents, and compute the prevalence of each feature in the two data sets. We select features with significantly different frequencies. These correspond to phenomena poorly modeled by traditional trigrams, and reveal interesting linguistic deficiencies of the initial model. We found 6798 linguistic features in the Switchboard domain and achieved small improvements in perplexity and speech recognition accuracy with these features.


doi: 10.21437/Eurospeech.1999-363

Cite as: Zhu, X., Chen, S.F., Rosenfeld, R. (1999) Linguistic features for whole sentence maximum entropy language models. Proc. 6th European Conference on Speech Communication and Technology (Eurospeech 1999), 1807-1810, doi: 10.21437/Eurospeech.1999-363

@inproceedings{zhu99b_eurospeech,
  author={Xiaojin Zhu and Stanley F. Chen and Ronald Rosenfeld},
  title={{Linguistic features for whole sentence maximum entropy language models}},
  year=1999,
  booktitle={Proc. 6th European Conference on Speech Communication and Technology (Eurospeech 1999)},
  pages={1807--1810},
  doi={10.21437/Eurospeech.1999-363}
}