Cross-domain Few-shot Learning with Task-specific Adapters
About
In this paper, we look at the problem of cross-domain few-shot classification that aims to learn a classifier from previously unseen classes and domains with few labeled samples. Recent approaches broadly solve this problem by parameterizing their few-shot classifiers with task-agnostic and task-specific weights where the former is typically learned on a large training set and the latter is dynamically predicted through an auxiliary network conditioned on a small support set. In this work, we focus on the estimation of the latter, and propose to learn task-specific weights from scratch directly on a small support set, in contrast to dynamically estimating them. In particular, through systematic analysis, we show that task-specific weights through parametric adapters in matrix form with residual connections to multiple intermediate layers of a backbone network significantly improves the performance of the state-of-the-art models in the Meta-Dataset benchmark with minor additional cost.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot classification | Meta-Dataset (test) | Omniglot78.2 | 48 | |
| Few-shot classification | Meta-Dataset | Avg Seen Accuracy80.4 | 45 | |
| Few-shot classification | Meta-Dataset 1.0 (test) | ILSVRC Accuracy57.35 | 42 | |
| Few-shot Image Classification | Meta-Dataset (test) | Omniglot Accuracy96.8 | 40 | |
| Few-shot Image Classification | Aircraft (test) | Mean Accuracy89.9 | 28 | |
| Few-shot classification | Textures Meta-Dataset (test) | Mean Accuracy77.5 | 10 | |
| Few-shot classification | CIFAR-10 Meta-Dataset (test) | Mean Accuracy82.9 | 10 | |
| Few-shot classification | CIFAR100 (test) | Accuracy70.4 | 10 | |
| Few-shot classification | ImageNet Meta-Dataset (test) | Mean Accuracy59.5 | 10 | |
| Few-shot classification | Omniglot Meta-Dataset (test) | Mean Accuracy94.9 | 10 |