An important goal of an automatic classifier is to learn the best possible generalization from given training material. One possible improvement over a standard learning algorithm is to train several related tasks in parallel. We apply the multi-task learning scheme to a recurrent neural network estimating phoneme posterior probabilities and HMM state posterior probabilities, respectively. A comparison of networks with different additional tasks within a hybrid NN/HMM acoustic model is presented. The evaluation has been performed using the WSJ0 speaker independent test set with a closed vocabulary of 5000 words and shows a significant improvement compared to a standard hybrid acoustic model if gender classification is used as additional task.
Cite as: Stadermann, J., Koska, W., Rigoll, G. (2005) Multi-task learning strategies for a recurrent neural net in a hybrid tied-posteriors acoustic model. Proc. Interspeech 2005, 2993-2996, doi: 10.21437/Interspeech.2005-137
@inproceedings{stadermann05_interspeech, author={Jan Stadermann and Wolfram Koska and Gerhard Rigoll}, title={{Multi-task learning strategies for a recurrent neural net in a hybrid tied-posteriors acoustic model}}, year=2005, booktitle={Proc. Interspeech 2005}, pages={2993--2996}, doi={10.21437/Interspeech.2005-137} }