Zero-Shot Learning with Common Sense Knowledge Graphs
About
Zero-shot learning relies on semantic class representations such as hand-engineered attributes or learned embeddings to predict classes without any labeled examples. We propose to learn class representations by embedding nodes from common sense knowledge graphs in a vector space. Common sense knowledge graphs are an untapped source of explicit high-level knowledge that requires little human effort to apply to a range of tasks. To capture the knowledge in the graph, we introduce ZSL-KG, a general-purpose framework with a novel transformer graph convolutional network (TrGCN) for generating class representations. Our proposed TrGCN architecture computes non-linear combinations of node neighbourhoods. Our results show that ZSL-KG improves over existing WordNet-based methods on five out of six zero-shot benchmark datasets in language and vision.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | ImageNet 1k (test) | Top-1 Accuracy3 | 798 | |
| Zero-shot Learning | AWA2 | Top-1 Accuracy0.781 | 95 | |
| Classification | AWA2 (test) | MCA (unseen)66.8 | 22 | |
| Zero-Shot Object Classification | aPY | U Score55.2 | 16 | |
| Image Classification | ImageNet 2-hop split | Flat Hit@126.3 | 15 | |
| Image Classification | ImageNet 3-hop split | Flat Hit@16.3 | 15 | |
| Object Classification | ImageNet All | Top-1 Accuracy3 | 8 | |
| Intent Classification | SNIPS-NLU (test) | Accuracy88.98 | 7 | |
| Fine-Grained Entity Typing | OntoNotes | -- | 6 | |
| Object Classification | APY (test) | U Score55.2 | 5 |