Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

C-Mixup: Improving Generalization in Regression

About

Improving the generalization of deep networks is an important open challenge, particularly in domains without plentiful data. The mixup algorithm improves generalization by linearly interpolating a pair of examples and their corresponding labels. These interpolated examples augment the original training set. Mixup has shown promising results in various classification tasks, but systematic analysis of mixup in regression remains underexplored. Using mixup directly on regression labels can result in arbitrarily incorrect labels. In this paper, we propose a simple yet powerful algorithm, C-Mixup, to improve generalization on regression tasks. In contrast with vanilla mixup, which picks training examples for mixing with uniform probability, C-Mixup adjusts the sampling probability based on the similarity of the labels. Our theoretical analysis confirms that C-Mixup with label similarity obtains a smaller mean square error in supervised regression and meta-regression than vanilla mixup and using feature similarity. Another benefit of C-Mixup is that it can improve out-of-distribution robustness, where the test distribution is different from the training distribution. By selectively interpolating examples with similar labels, it mitigates the effects of domain-associated information and yields domain-invariant representations. We evaluate C-Mixup on eleven datasets, ranging from tabular to video data. Compared to the best prior approach, C-Mixup achieves 6.56%, 4.76%, 5.82% improvements in in-distribution generalization, task generalization, and out-of-distribution robustness, respectively. Code is released at https://github.com/huaxiuyao/C-Mixup.

Huaxiu Yao, Yiping Wang, Linjun Zhang, James Zou, Chelsea Finn• 2022

Related benchmarks

TaskDatasetResultRank
Age EstimationAFAD B
MRAE (%)2.74
33
LVEF PredictionEchoNet-Dynamic (test)
MAE7.59
28
RegressionRCF-MNIST
RMSE (Avg)0.146
24
Age EstimationIMDB-WIKI-B
MRAE-5.26
17
RegressionMSD-B (test)
MRAE (Symmetric, 20%)3.75
17
Age EstimationUTKFace-B 20% symmetric noise
MRAE17.76
16
Age EstimationUTKFace-B 80% symmetric noise
MRAE117.7
16
Age EstimationUTKFace-B 40% symmetric noise
MRAE34
16
Age EstimationUTKFace-B 60% symmetric noise
MRAE74.29
16
RegressionSHIFT15M-B (test)
MRAE (Symmetric, 20%)9.47
16
Showing 10 of 29 rows

Other info

Code

Follow for update