Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Structured State Space Decoder for Speech Recognition and Synthesis

About

Automatic speech recognition (ASR) systems developed in recent years have shown promising results with self-attention models (e.g., Transformer and Conformer), which are replacing conventional recurrent neural networks. Meanwhile, a structured state space model (S4) has been recently proposed, producing promising results for various long-sequence modeling tasks, including raw speech classification. The S4 model can be trained in parallel, same as the Transformer model. In this study, we applied S4 as a decoder for ASR and text-to-speech (TTS) tasks by comparing it with the Transformer decoder. For the ASR task, our experimental results demonstrate that the proposed model achieves a competitive word error rate (WER) of 1.88%/4.25% on LibriSpeech test-clean/test-other set and a character error rate (CER) of 3.80%/2.63%/2.98% on the CSJ eval1/eval2/eval3 set. Furthermore, the proposed model is more robust than the standard Transformer model, particularly for long-form speech on both the datasets. For the TTS task, the proposed method outperforms the Transformer baseline.

Koichi Miyazaki, Masato Murata, Tomoki Koriyama• 2022

Related benchmarks

TaskDatasetResultRank
Automatic Speech RecognitionLibriSpeech (test-other)
WER5.13
966
Automatic Speech RecognitionLibriSpeech clean (test)
WER2.29
833
Automatic Speech RecognitionLibrispeech (test-clean)
WER2.29
84
Automatic Speech RecognitionAISHELL-1 (test)
CER4.9
71
Automatic Speech RecognitionWenetSpeech Meeting (test)
CER15.9
45
Automatic Speech RecognitionAISHELL-1 (dev)
CER4.5
34
Automatic Speech RecognitionWenetSpeech Net (test)
CER8.9
25
Automatic Speech RecognitionWenetSpeech (dev)
CER9.7
14
Showing 8 of 8 rows

Other info

Follow for update