Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Speech Denoising in the Waveform Domain with Self-Attention

About

In this work, we present CleanUNet, a causal speech denoising model on the raw waveform. The proposed model is based on an encoder-decoder architecture combined with several self-attention blocks to refine its bottleneck representations, which is crucial to obtain good results. The model is optimized through a set of losses defined over both waveform and multi-resolution spectrograms. The proposed method outperforms the state-of-the-art models in terms of denoised speech quality from various objective and subjective evaluation metrics. We release our code and models at https://github.com/nvidia/cleanunet.

Zhifeng Kong, Wei Ping, Ambrish Dantrey, Bryan Catanzaro• 2022

Related benchmarks

TaskDatasetResultRank
Speech EnhancementMultilingual low-SNR (evaluation set)
PESQ2.82
23
Speech DenoisingDNS no-reverb (test)
PESQ (WB)3.146
16
Speech Denoisinginternal dataset (test)
PESQ (WB)2.45
5
Speech DenoisingValentini (test)
PESQ WB2.905
4
Speech DenoisingDNS
RTF0.0034
4
Showing 5 of 5 rows

Other info

Code

Follow for update