Hallucination Improves Few-Shot Object Detection
About
Learning to detect novel objects from few annotated examples is of great practical importance. A particularly challenging yet common regime occurs when there are extremely limited examples (less than three). One critical factor in improving few-shot detection is to address the lack of variation in training data. We propose to build a better model of variation for novel classes by transferring the shared within-class variation from base classes. To this end, we introduce a hallucinator network that learns to generate additional, useful training examples in the region of interest (RoI) feature space, and incorporate it into a modern object detection model. Our approach yields significant performance improvements on two state-of-the-art few-shot detectors with different proposal generation procedures. In particular, we achieve new state of the art in the extremely-few-shot regime on the challenging COCO benchmark.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Object Detection | MS COCO novel classes | nAP6.9 | 132 | |
| Object Detection | MS COCO novel classes 2017 (val) | AP6.9 | 123 | |
| Object Detection | PASCAL VOC Set 2 (novel) | -- | 110 | |
| Object Detection | PASCAL VOC Set 3 (novel) | AP50 (shot=1)40.4 | 71 | |
| Object Detection | PASCAL VOC (Novel Set 1) | AP50 (shot=1)47 | 71 | |
| Few-shot Object Detection | Pascal VOC | mAP48.1 | 65 | |
| Object Detection | Pascal VOC Overall Average 2007 (test) | mAP@0.540.6 | 20 | |
| Object Detection | MS COCO (novel split) | 1-Shot nAP4.4 | 15 | |
| Object Detection | MS COCO novel classes 1-shot | AP4.4 | 6 | |
| Object Detection | MS COCO 2-shot (novel classes) | AP5.6 | 6 |