Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Equivariance with Learned Canonicalization Functions

About

Symmetry-based neural networks often constrain the architecture in order to achieve invariance or equivariance to a group of transformations. In this paper, we propose an alternative that avoids this architectural constraint by learning to produce canonical representations of the data. These canonicalization functions can readily be plugged into non-equivariant backbone architectures. We offer explicit ways to implement them for some groups of interest. We show that this approach enjoys universality while providing interpretable insights. Our main hypothesis, supported by our empirical results, is that learning a small neural network to perform canonicalization is better than using predefined heuristics. Our experiments show that learning the canonicalization function is competitive with existing techniques for learning equivariant functions across many tasks, including image classification, $N$-body dynamics prediction, point cloud classification and part segmentation, while being faster across the board.

S\'ekou-Oumar Kaba, Arnab Kumar Mondal, Yan Zhang, Yoshua Bengio, Siamak Ravanbakhsh• 2022

Related benchmarks

TaskDatasetResultRank
Shape Part SegmentationShapeNet (test)
Mean IoU81.07
164
ClassificationModelNet40
Accuracy88.87
108
Node ClassificationPATTERN (test)
Test Accuracy86.534
88
Image ClassificationCIFAR-10 original (test)
Accuracy95
87
Graph ClassificationEXP (test)
Accuracy50
33
Image ClassificationCIFAR100 original (test)
Accuracy80.86
20
Image ClassificationCIFAR-10 data-augmented (+) (test)--
16
Image ClassificationSTL10 original (test)
Accuracy95.3
15
Graph SeparationGRAPH8c random initialization
Non-Separated Pairs0.00e+0
11
Graph SeparationEXP random initialization
Non-separated Graph Pairs0.00e+0
11
Showing 10 of 15 rows

Other info

Follow for update