Semi-Supervised Adaptation of RNNLMs by Fine-Tuning with Domain-Specific Auxiliary Features

Salil Deena, Raymond W.M. Ng, Pranava Madhyastha, Lucia Specia, Thomas Hain


Recurrent neural network language models (RNNLMs) can be augmented with auxiliary features, which can provide an extra modality on top of the words. It has been found that RNNLMs perform best when trained on a large corpus of generic text and then fine-tuned on text corresponding to the sub-domain for which it is to be applied. However, in many cases the auxiliary features are available for the sub-domain text but not for the generic text. In such cases, semi-supervised techniques can be used to infer such features for the generic text data such that the RNNLM can be trained and then fine-tuned on the available in-domain data with corresponding auxiliary features.

In this paper, several novel approaches are investigated for dealing with the semi-supervised adaptation of RNNLMs with auxiliary features as input. These approaches include: using zero features during training to mask the weights of the feature sub-network; adding the feature sub-network only at the time of fine-tuning; deriving the features using a parametric model and; back-propagating to infer the features on the generic text. These approaches are investigated and results are reported both in terms of PPL and WER on a multi-genre broadcast ASR task.


 DOI: 10.21437/Interspeech.2017-1598

Cite as: Deena, S., Ng, R.W., Madhyastha, P., Specia, L., Hain, T. (2017) Semi-Supervised Adaptation of RNNLMs by Fine-Tuning with Domain-Specific Auxiliary Features. Proc. Interspeech 2017, 2715-2719, DOI: 10.21437/Interspeech.2017-1598.


@inproceedings{Deena2017,
  author={Salil Deena and Raymond W.M. Ng and Pranava Madhyastha and Lucia Specia and Thomas Hain},
  title={Semi-Supervised Adaptation of RNNLMs by Fine-Tuning with Domain-Specific Auxiliary Features},
  year=2017,
  booktitle={Proc. Interspeech 2017},
  pages={2715--2719},
  doi={10.21437/Interspeech.2017-1598},
  url={http://dx.doi.org/10.21437/Interspeech.2017-1598}
}