Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SelectMix: Enhancing Label Noise Robustness through Targeted Sample Mixing

About

Deep neural networks tend to memorize noisy labels, severely degrading their generalization performance. Although Mixup has demonstrated effectiveness in improving generalization and robustness, existing Mixup-based methods typically perform indiscriminate mixing without principled guidance on sample selection and mixing strategy, inadvertently propagating noisy supervision. To overcome these limitations, we propose SelectMix, a confidence-guided mixing framework explicitly tailored for noisy labels. SelectMix first identifies potentially noisy or ambiguous samples through confidence based mismatch analysis using K-fold cross-validation, then selectively blends identified uncertain samples with confidently predicted peers from their potential classes. Furthermore, SelectMix employs soft labels derived from all classes involved in the mixing process, ensuring the labels accurately represent the composition of the mixed samples, thus aligning supervision signals closely with the actual mixed inputs. Through extensive theoretical analysis and empirical evaluations on multiple synthetic (MNIST, Fashion-MNIST, CIFAR-10, CIFAR-100) and real-world benchmark datasets (CIFAR-N, MNIST and Clothing1M), we demonstrate that SelectMix consistently outperforms strong baseline methods, validating its effectiveness and robustness in learning with noisy labels.

Qiuhao Liu, Ling Li, Yao Lu, Qi Xuan, Zhaowei Zhu, Jiaheng Wei• 2025

Related benchmarks

TaskDatasetResultRank
Reward ModelingHelpSteer (test)
MAE0.2624
48
Safety EvaluationWildGuardMix
Safety Score0.8672
22
Reward ModelingUltraFeedback (test)
MAE0.1679
21
Reward ModelingPKU-SafeRLHF (test)
MAE0.1422
19
Safety AlignmentStrongREJECT--
18
Safety AlignmentHarmBench
Score95.78
10
Showing 6 of 6 rows

Other info

Follow for update