INTERSPEECH 2015
16th Annual Conference of the International Speech Communication Association

Dresden, Germany
September 6-10, 2015

Two-Stage Multi-Target Joint Learning for Monaural Speech Separation

Shuai Nie (1), Shan Liang (1), Wei Xue (1), Xueliang Zhang (2), Wenju Liu (1), Like Dong (3), Hong Yang (3)

(1) Chinese Academy of Sciences, China
(2) Inner Mongolia University, China
(3) SGCC, China

Recently, supervised speech separation has been extensively studied and shown considerable promise. Due to the temporal continuity of speech, speech auditory features and separation targets present prominent spectro-temporal structures and strong correlations over the time-frequency (T-F) domain, which can be exploited for speech separation. However, many supervised speech separation methods independently model each T-F unit with only one target and much ignore these useful information. In this paper, we propose a two-stage multi-target joint learning method to jointly model the related speech separation targets at the frame level. Systematic experiments show that the proposed approach consistently achieves better separation and generalization performances in the low signal-to-noise ratio (SNR) conditions.

Full Paper

Bibliographic reference.  Nie, Shuai / Liang, Shan / Xue, Wei / Zhang, Xueliang / Liu, Wenju / Dong, Like / Yang, Hong (2015): "Two-stage multi-target joint learning for monaural speech separation", In INTERSPEECH-2015, 1503-1507.