In everyday life, speech is often accompanied by a situation-specific acoustic cue; a hungry bark as you ask ‘Has anyone fed the dog?’. This paper investigates the effect such cues have on speech intelligibility in noise and evaluates their interaction with the established effect of situation-specific semantic cues. This work is motivated by the introduction of new object-based broadcast formats, which have the potential to optimise intelligibility by controlling the level of individual broadcast audio elements, at point of service. Results of this study show that situation-specific acoustic cues alone can improve word recognition in multi-talker babble by 69.5%, a similar amount to semantic cues. The combination of both semantic and acoustic cues provide further improvement of 106.0% compared with no cues, and 18.7% compared with semantic cues only. Interestingly, whilst increasing subjective intelligibility of the target word, the presence of acoustic cues degraded the objective intelligibility of the speech-based semantic cues by 47.0% (equivalent to reducing the speech level by 4.5 dB). This paper discusses the interactions between the two types of cues and the implications that these results have for assessing and improving the intelligibility of broadcast speech.
Cite as: Ward, L., Shirley, B., Tang, Y., Davies, W.J. (2017) The Effect of Situation-Specific Non-Speech Acoustic Cues on the Intelligibility of Speech in Noise. Proc. Interspeech 2017, 2958-2962, doi: 10.21437/Interspeech.2017-500
@inproceedings{ward17b_interspeech, author={Lauren Ward and Ben Shirley and Yan Tang and William J. Davies}, title={{The Effect of Situation-Specific Non-Speech Acoustic Cues on the Intelligibility of Speech in Noise}}, year=2017, booktitle={Proc. Interspeech 2017}, pages={2958--2962}, doi={10.21437/Interspeech.2017-500} }