Learning Like Humans: Analogical Concept Learning for Generalized Category Discovery
About
Generalized Category Discovery (GCD) seeks to uncover novel categories in unlabeled data while preserving recognition of known categories, yet prevailing visual-only pipelines and the loose coupling between supervised learning and discovery often yield brittle boundaries on fine-grained, look-alike categories. We introduce the Analogical Textual Concept Generator (ATCG), a plug-and-play module that analogizes from labeled knowledge to new observations, forming textual concepts for unlabeled samples. Fusing these analogical textual concepts with visual features turns discovery into a visual-textual reasoning process, transferring prior knowledge to novel data and sharpening category separation. ATCG attaches to both parametric and clustering style GCD pipelines and requires no changes to their overall design. Across six benchmarks, ATCG consistently improves overall, known-class, and novel-class performance, with the largest gains on fine-grained data. Our code is available at: https://github.com/zhou-9527/AnaLogical-GCD.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Generalized Category Discovery | ImageNet-100 | All Accuracy92.7 | 208 | |
| Generalized Category Discovery | CIFAR-100 | Accuracy (All)84.7 | 185 | |
| Generalized Category Discovery | Stanford Cars | Accuracy (All)80 | 160 | |
| Generalized Category Discovery | CUB | Accuracy (All)84.1 | 133 | |
| Generalized Category Discovery | FGVC Aircraft | Accuracy (All)66.6 | 105 | |
| Generalized Category Discovery | Herbarium19 | Score (All Categories)50.3 | 71 | |
| Generalized Category Discovery | Fine-grained Avg | Overall Accuracy76.6 | 12 | |
| Generalized Category Discovery | All Datasets Avg | Overall Accuracy75.1 | 12 | |
| Generalized Category Discovery | Classification Avg | Overall Accuracy88.7 | 10 | |
| Generalized Category Discovery | CUB-200 | Accuracy (All)73.7 | 5 |