Development of Robust Automated Scoring Models Using Adversarial Input for Oral Proficiency Assessment

Su-Youn Yoon, Chong Min Lee, Klaus Zechner, Keelan Evanini


In this study, we developed an automated scoring model for an oral proficiency test eliciting spontaneous speech from non-native speakers of English. In a large-scale oral proficiency test, a small number of responses may have atypical characteristics that make it difficult even for state-of-the-art automated scoring models to assign fair scores. The oral proficiency test in this study consisted of questions asking about content in materials provided to the test takers, and the atypical responses frequently had serious content abnormalities. In order to develop an automated scoring system that is robust to these atypical responses, we first developed a set of content features to capture content abnormalities. Next, we trained scoring models using the augmented training dataset, including synthetic atypical responses. Compared to the baseline scoring model, the new model showed comparable performance in scoring normal responses, while it assigned fairer scores for authentic atypical responses extracted from operational test administrations.


 DOI: 10.21437/Interspeech.2019-1711

Cite as: Yoon, S., Lee, C.M., Zechner, K., Evanini, K. (2019) Development of Robust Automated Scoring Models Using Adversarial Input for Oral Proficiency Assessment. Proc. Interspeech 2019, 1871-1875, DOI: 10.21437/Interspeech.2019-1711.


@inproceedings{Yoon2019,
  author={Su-Youn Yoon and Chong Min Lee and Klaus Zechner and Keelan Evanini},
  title={{Development of Robust Automated Scoring Models Using Adversarial Input for Oral Proficiency Assessment}},
  year=2019,
  booktitle={Proc. Interspeech 2019},
  pages={1871--1875},
  doi={10.21437/Interspeech.2019-1711},
  url={http://dx.doi.org/10.21437/Interspeech.2019-1711}
}