Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SpatialNet: Extensively Learning Spatial Information for Multichannel Joint Speech Separation, Denoising and Dereverberation

About

This work proposes a neural network to extensively exploit spatial information for multichannel joint speech separation, denoising and dereverberation, named SpatialNet. In the short-time Fourier transform (STFT) domain, the proposed network performs end-to-end speech enhancement. It is mainly composed of interleaved narrow-band and cross-band blocks to respectively exploit narrow-band and cross-band spatial information. The narrow-band blocks process frequencies independently, and use self-attention mechanism and temporal convolutional layers to respectively perform spatial-feature-based speaker clustering and temporal smoothing/filtering. The cross-band blocks process frames independently, and use full-band linear layer and frequency convolutional layers to respectively learn the correlation between all frequencies and adjacent frequencies. Experiments are conducted on various simulated and real datasets, and the results show that 1) the proposed network achieves the state-of-the-art performance on almost all tasks; 2) the proposed network suffers little from the spectral generalization problem; and 3) the proposed network is indeed performing speaker clustering (demonstrated by attention maps).

Changsheng Quan, Xiaofei Li• 2023

Related benchmarks

TaskDatasetResultRank
Speech EnhancementAd-hoc microphone array 8-channel (test)
STOI0.959
59
Speech EnhancementAd-hoc microphone array 4-channel (test)
STOI0.933
38
Multi-channel speech enhancementSimulated adhoc microphone array dataset 4-channel (test)
STOI94.4
21
Universal Sound SeparationMC-FUSS
SI-SDRi (J=2)17.8
10
Universal Source SeparationASA2
SI-SDRi9.6
7
Speech SeparationWHAMR! 2CH
SI-SNRi (dB)20.2
6
Showing 6 of 6 rows

Other info

Follow for update