Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

AISHELL

Benchmarks

Task NameDataset NameSOTA ResultTrend
Speech RecognitionAISHELL-1 (dev)
WER1.2
28
Automatic Speech RecognitionAISHELL-1
WER0.6
22
Speaker DiarizationAISHELL-4
DER (%)1.6
20
Automatic Speech RecognitionAISHELL (test)
CER1.95
20
Speaker-Attributed Automatic Speech RecognitionAISHELL-4 (test)
CER0.1543
18
Speech SynthesisAISHELL3 Mandarin
UTMOS2.7
14
Automatic Speech RecognitionSLR93 (AISHELL-3) (test)
CER4.55
10
Automatic Speech RecognitionAISHELL-2 (ios)
CER2.33
10
Automatic Speech RecognitionAISHELL Mandarin 3
CER1.86
9
Automatic Speech RecognitionAISHELL D 2021 (test)
CER1.66
7
Automatic Speech RecognitionAISHELL C 2021 (Eval)
CER1.71
7
Automatic Speech RecognitionAISHELL Eval A 2021
CER3.45
7
Automatic Speech RecognitionAISHELL-2 ios (dev)
CER2.08
7
Automatic Speech RecognitionAISHELL-2
Word Error Rate (WER)2.16
7
Automatic Speech RecognitionAISHELL-1 1.0 (test)
CER (Offline, Rescoring)5.25
7
ASR Error CorrectionAISHELL-1 (dev)
WER3.8
6
Target Speaker ExtractionAISHELL Noisy zero-shot
SI-SDR10.2
5
Target Speaker ExtractionAISHELL zero-shot Clean
SI-SDR13.4
5
Text-to-SpeechAISHELL-1
Error Rate1.9
4
Automatic Speech RecognitionAISHELL-1
Error Rate2.5
4
Contextual Automatic Speech RecognitionAISHELL-1-NE (test)
CER0.92
4
Speech WatermarkingAiShell3 (OOD)
GN+Ec99.33
4
Speech WatermarkingAiShell3 (out-of-distribution)
Robustness (Gaussian Noise 5 dB)98.68
4
Automatic Speech RecognitionAISHELL
CER0.54
4
Automatic Speech RecognitionAISHELL-3
Error Rate9.2
3
Showing 25 of 31 rows