Divide and Conquer: A Deep CASA Approach to Talker-independent Monaural Speaker Separation
About
We address talker-independent monaural speaker separation from the perspectives of deep learning and computational auditory scene analysis (CASA). Specifically, we decompose the multi-speaker separation task into the stages of simultaneous grouping and sequential grouping. Simultaneous grouping is first performed in each time frame by separating the spectra of different speakers with a permutation-invariantly trained neural network. In the second stage, the frame-level separated spectra are sequentially grouped to different speakers by a clustering network. The proposed deep CASA approach optimizes frame-level separation and speaker tracking in turn, and produces excellent results for both objectives. Experimental results on the benchmark WSJ0-2mix database show that the new approach achieves the state-of-the-art results with a modest model size.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Speech Separation | WSJ0-2Mix (test) | SDRi (dB)18 | 141 | |
| Speech Separation | WSJ0-2Mix | SI-SNRi (dB)17.7 | 65 | |
| Source Separation | WSJ0-2Mix (test) | SI-SNRi17.7 | 17 | |
| Speaker Separation | WSJ0-2mix OC (test) | PESQ3.51 | 15 | |
| Speaker Separation | WSJ0-2mix 8kHz (test) | ΔSDR18 | 14 |