Learning Dense Object Descriptors from Multiple Views for Low-shot Category Generalization
About
A hallmark of the deep learning era for computer vision is the successful use of large-scale labeled datasets to train feature representations for tasks ranging from object recognition and semantic segmentation to optical flow estimation and novel view synthesis of 3D scenes. In this work, we aim to learn dense discriminative object representations for low-shot category recognition without requiring any category labels. To this end, we propose Deep Object Patch Encodings (DOPE), which can be trained from multiple views of object instances without any category or semantic object part labels. To train DOPE, we assume access to sparse depths, foreground masks and known cameras, to obtain pixel-level correspondences between views of an object, and use this to formulate a self-supervised learning task to learn discriminative object patches. We find that DOPE can directly be used for low-shot classification of novel categories using local-part matching, and is competitive with and outperforms supervised and self-supervised learning baselines. Code and data available at https://github.com/rehg-lab/dope_selfsup.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Low-shot recognition | CO3D-LS (test) | Accuracy75.16 | 32 | |
| Low-shot recognition | ModelNet40-LS (test) | Acc (5-way 1-shot)62.76 | 12 | |
| Image-only low-shot recognition | ShapeNet LS | 5-way 1-shot Accuracy62 | 10 | |
| Multi-object Category Recognition (Categ-MObj) | Toys4k multi-object setting | LSA56.99 | 10 | |
| Low-shot object recognition | Toys4k Inst-SObj | Accuracy95.8 | 6 | |
| Low-shot recognition | Toys4k Inst-SObj 1.0 (test) | Accuracy95.8 | 6 | |
| Low-shot object recognition | Toys4k Categ-SObj | Accuracy73.06 | 6 | |
| Low-shot object recognition | Toys4k Categ-SObj-PoseVar | Accuracy68.84 | 6 | |
| Low-shot recognition | Toys4k Categ-SObj 1.0 (test) | Accuracy73.06 | 6 | |
| Low-shot recognition | Toys4k Categ-SObj-Pose Var 1.0 (test) | Accuracy68.84 | 6 |