Speech and Language Technology in Education (SLaTE 2013)
This paper presents approaches to automated content scoring of spoken language test responses from non-native speakers of English which contain multiple parts addressing factual information that the test taker has previously heard via auditory stimulus materials. While previous work relating to content scoring of spontaneous, unpredictable speech has focused only on entire responses and on general topic matching approaches, such as content vector analysis, the specific nature of spoken responses in our data requires response segmentation and extraction of features that indicate the relevance and correctness of the facts contained in the different parts of the response. Our best content features, based on similarity with key facts and concepts, achieve correlations of r=0.615 (for speech recognition output) and r=0.637 (using human transcriptions) with expert human rater scores. Furthermore, we show that these content features outperform traditional vector space based features. Finally, we demonstrate that the performance of a scoring model based on a combination of features developed previously and some of the newly designed content features improves significantly from r=0.624 to r=0.664 on an unseen evaluation set when using speech recognition output.
Index Terms: spoken language assessment, automated scoring, content appropriateness
Bibliographic reference. Xiong, Wenting / Evanini, Keelan / Zechner, Klaus / Chen, Lei (2013): "Automated content scoring of spoken responses containing multiple parts with factual information", In SLaTE-2013, 137-142.