Powerset multi-class cross entropy loss for neural speaker diarization
About
Since its introduction in 2019, the whole end-to-end neural diarization (EEND) line of work has been addressing speaker diarization as a frame-wise multi-label classification problem with permutation-invariant training. Despite EEND showing great promise, a few recent works took a step back and studied the possible combination of (local) supervised EEND diarization with (global) unsupervised clustering. Yet, these hybrid contributions did not question the original multi-label formulation. We propose to switch from multi-label (where any two speakers can be active at the same time) to powerset multi-class classification (where dedicated classes are assigned to pairs of overlapping speakers). Through extensive experiments on 9 different benchmarks, we show that this formulation leads to significantly better performance (mostly on overlapping speech) and robustness to domain mismatch, while eliminating the detection threshold hyperparameter, critical for the multi-label formulation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Speaker Diarization | AISHELL-4 | DER (%)11.9 | 20 | |
| Speaker Diarization | RAMC | DER18.4 | 9 | |
| Speaker Diarization | AliMeeting far | DER22.5 | 6 | |
| Joint ASR and Diarization | AliMeeting Far-field Mandarin (test) | Error Rate2.52 | 5 | |
| Speaker Diarization | VoxConverse v0.3 | DER (%)0.094 | 5 | |
| Joint ASR and Diarization | AMI SDM English (test) | Fail Rate4.6 | 5 | |
| Speaker Diarization | AMI Channel 1 | DER (%)20.9 | 5 | |
| Speaker Diarization | MSDWild Few | DER (%)19.8 | 4 |