The Achilles' Heel of Angular Margins: A Chebyshev Polynomial Fix for Speaker Verification
About
Angular margin losses, such as AAM-Softmax, have become the de facto in speaker and face verification. Their success hinges on directly manipulating the angle between features and class prototypes. However, this manipulation relies on the arccos function to recover the angle, introducing a significant yet overlooked source of training instability. The derivative of arccos explodes at its boundaries, causing gradient peaks during optimisation. Furthermore, the formulation fails to generate a sufficiently sharp gradient for hard-to-classify examples. We address these issues by proposing ChebyAAM, a loss that replaces the arccos operation with its Chebyshev polynomial approximation. This substitution eliminates gradient explosion and applies a stronger corrective signal to hard examples, leading to more effective optimisation. Experiments on three benchmarks (VoxCeleb, SITW, and CN-Celeb) demonstrate that our method resolves the instability and consistently improves performance. Our work suggests that approximating angular operations, rather than calculating them explicitly, offers a more robust path for designing future metric learning losses. Code is available at https://github.com/ExtraOrdinaryLab/vibe.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Speaker Verification | VoxCeleb1 (Vox1-O) | EER0.8453 | 33 | |
| Speaker Verification | VoxCeleb1 (Vox1-H) | EER1.9008 | 20 | |
| Speaker Verification | VoxCeleb-E | -- | 15 | |
| Speaker Verification | CNCeleb (eval) | EER7.1099 | 12 | |
| Speaker Recognition | SITW (Speakers In The Wild) core-core protocol | EER1.3122 | 9 | |
| Speaker Verification | SITW core-multi | EER0.0435 | 4 |