Joint Phoneme Alignment and Text-Informed Speech Separation on Highly Corrupted Speech

Kilian Schulze-Forster, Clement Doire, Gaël Richard, Roland Badeau
IEEE International Conference on Acoustics, Speech, and Signal Processing, 2020

Paper       bibtex       Code       Slides       Poster

Welcome to the website of the paper. Here you find some audio examples. To access the paper, code, or poster click on the links above.

Abstract

Speech separation quality can be improved by exploiting textual information. However, this usually requires text-to-speech alignment at phoneme level. Classical alignment methods are made for rather clean speech and do not work as well on corrupted speech. We propose to perform text-informed speech-music separation and phoneme alignment jointly using recurrent neural networks and the attention mechanism. We show that it leads to benefits for both tasks. In experiments, phoneme transcripts are used to improve the perceived quality of separated speech over a non-informed baseline. Moreover, our novel phoneme alignment method based on the attention mechanism achieves state-of-the-art alignment accuracy on clean and on heavily corrupted speech.

Audio Examples

The audio examples are taken from the test set which is described in the paper.


Example 1

Corrupted speech:
Clean speech:
BL:
V1:
V2:
V3:
OA:

Example 2

Corrupted speech:
Clean speech:
BL:
V1:
V2:
V3:
OA: