Sixth European Conference on Speech Communication and Technology
(EUROSPEECH'99)

Budapest, Hungary
September 5-9, 1999

Segregation of Vowel in Background Noise Using the Model of Segregating Two Acoustic Sources Based on Auditory Scene Analysis

Masashi Unoki, Masato Akagi

Japan Advanced Institute of Science and Technology Tatsunokuchi, Nomi, Ishikawa, Japan

This paper proposes an auditory sound segregation model based on auditory scene analysis. It solves the problem of segregating two acoustic sources by using constraints related to the heuristic regularities proposed by Bregman and by making an improvement to our previously proposed model. The improvement is to reconsider constraints on the continuity of instantaneous phases as well as constraints on the continuity of instantaneous amplitudes and fundamental frequencies in order to segregate the desired signal from a noisy signal precisely even in waveforms. Simulations performed to segregate a real vowel from a noisy vowel and to compare the results of using all or only some constraints showed that our improved model can segregate real speech precisely even in waveforms using all the constraints related to the four regularities, and that the absence of some constraints reduces the segregation accuracy.


Full Paper (PDF)

Bibliographic reference.  Unoki, Masashi / Akagi, Masato (1999): "Segregation of vowel in background noise using the model of segregating two acoustic sources based on auditory scene analysis", In EUROSPEECH'99, 2575-2578.