ISCA Archive Interspeech 2021
ISCA Archive Interspeech 2021

Contextual Density Ratio for Language Model Biasing of Sequence to Sequence ASR Systems

Jesús Andrés-Ferrer, Dario Albesano, Puming Zhan, Paul Vozila

End-2-end (E2E) models have become increasingly popular in some ASR tasks because of their performance and advantages. These E2E models directly approximate the posterior distribution of tokens given the acoustic inputs. Consequently, the E2E systems implicitly define a language model (LM) over the output tokens, which makes the exploitation of independently trained language models less straightforward than in conventional ASR systems. This makes it difficult to dynamically adapt E2E ASR system to contextual profiles for better recognizing special words such as named entities. In this work, we propose a contextual density ratio approach for both training a contextual aware E2E model and adapting the language model to named entities. We apply the aforementioned technique to an E2E ASR system, which transcribes doctor and patient conversations, for better adapting the E2E system to the names in the conversations. Our proposed technique achieves a relative improvement of up to 46.5% on the names over an E2E baseline without degrading the overall recognition accuracy of the whole test set. Moreover, it also surpasses a contextual shallow fusion baseline by 22.1% relative.


doi: 10.21437/Interspeech.2021-443

Cite as: Andrés-Ferrer, J., Albesano, D., Zhan, P., Vozila, P. (2021) Contextual Density Ratio for Language Model Biasing of Sequence to Sequence ASR Systems. Proc. Interspeech 2021, 2007-2011, doi: 10.21437/Interspeech.2021-443

@inproceedings{andresferrer21_interspeech,
  author={Jesús Andrés-Ferrer and Dario Albesano and Puming Zhan and Paul Vozila},
  title={{Contextual Density Ratio for Language Model Biasing of Sequence to Sequence ASR Systems}},
  year=2021,
  booktitle={Proc. Interspeech 2021},
  pages={2007--2011},
  doi={10.21437/Interspeech.2021-443}
}