Tunable Soft Equivariance with Guarantees
About
Equivariance is a fundamental property in computer vision models, yet strict equivariance is rarely satisfied in real-world data, which can limit a model's performance. Controlling the degree of equivariance is therefore desirable. We propose a general framework for constructing soft equivariant models by projecting the model weights into a designed subspace. The method applies to any pre-trained architecture and provides theoretical bounds on the induced equivariance error. Empirically, we demonstrate the effectiveness of our method on multiple pre-trained backbones, including ViT and ResNet, across image classification, semantic segmentation, and human-trajectory prediction tasks. Notably, our approach improves the performance while simultaneously reducing equivariance error on the competitive ImageNet benchmark.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Trajectory Prediction | ETH UCY (test) | -- | 72 | |
| Trajectory Prediction | Hotel ETH-UCY (test) | ADE5.69 | 58 | |
| Trajectory Prediction | UNIV ETH-UCY (test) | ADE7.85 | 44 | |
| Regression | Synthetic O(5) | Relative MSE (10^-1)0.72 | 12 | |
| Image Classification | CIFAR10 (test) | Accuracy (Acc)98.82 | 9 | |
| Image Classification | CIFAR100 (test) | Accuracy (Acc)91.03 | 9 | |
| Semantic segmentation | PASCAL VOC 15 (test) | mIoU89.48 | 9 | |
| Human Trajectory Prediction | ZARA1 ETH-UCY (test) | cADE3.4 | 3 | |
| Human Trajectory Prediction | ZARA2 ETH/UCY (test) | cADE2.91 | 3 |