CD-FSOD: A Benchmark for Cross-domain Few-shot Object Detection
About
In this paper, we propose a study of the cross-domain few-shot object detection (CD-FSOD) benchmark, consisting of image data from a diverse data domain. On the proposed benchmark, we evaluate state-of-art FSOD approaches, including meta-learning FSOD approaches and fine-tuning FSOD approaches. The results show that these methods tend to fall, and even underperform the naive fine-tuning model. We analyze the reasons for their failure and introduce a strong baseline that uses a mutually-beneficial manner to alleviate the overfitting problem. Our approach is remarkably superior to existing approaches by significant margins (2.0\% on average) on the proposed benchmark. Our code is available at \url{https://github.com/FSOD/CD-FSOD}.
Wuti Xiong• 2022
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot Object Detection | CD-FSOD | ArTaxOr Score18.1 | 200 | |
| Object Detection | Cross-Domain Object Detection (COCO to ArTaxOr, Clipart1k, DIOR, DeepFish, NEU-DET, UODD) 10-shot | mAP (ArTaxOr)12.5 | 17 | |
| Object Detection | UODD one-shot 14 (test) | nAP (1-shot)5.9 | 14 | |
| Object Detection | ArTaxOr | nAP (1-shot)5.1 | 14 | |
| Object Detection | DIOR | nAP (1-shot)10.5 | 14 | |
| Object Detection | NEU-DET | mAP21.1 | 13 | |
| Object Detection | Clipart1k | mAP27.3 | 13 | |
| Object Detection | UODD | mAP14.5 | 13 | |
| Object Detection | DIOR | mAP26.5 | 13 | |
| Object Detection | DeepFish | mAP15.5 | 13 |
Showing 10 of 13 rows