Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CoKE: Contextualized Knowledge Graph Embedding

About

Knowledge graph embedding, which projects symbolic entities and relations into continuous vector spaces, is gaining increasing attention. Previous methods allow a single static embedding for each entity or relation, ignoring their intrinsic contextual nature, i.e., entities and relations may appear in different graph contexts, and accordingly, exhibit different properties. This work presents Contextualized Knowledge Graph Embedding (CoKE), a novel paradigm that takes into account such contextual nature, and learns dynamic, flexible, and fully contextualized entity and relation embeddings. Two types of graph contexts are studied: edges and paths, both formulated as sequences of entities and relations. CoKE takes a sequence as input and uses a Transformer encoder to obtain contextualized representations. These representations are hence naturally adaptive to the input, capturing contextual meanings of entities and relations therein. Evaluation on a wide variety of public benchmarks verifies the superiority of CoKE in link prediction and path query answering. It performs consistently better than, or at least equally well as current state-of-the-art in almost every case, in particular offering an absolute improvement of 21.0% in H@10 on path query answering. Our code is available at \url{https://github.com/PaddlePaddle/Research/tree/master/KG/CoKE}.

Quan Wang, Pingping Huang, Haifeng Wang, Songtai Dai, Wenbin Jiang, Jing Liu, Yajuan Lyu, Yong Zhu, Hua Wu• 2019

Related benchmarks

TaskDatasetResultRank
Link PredictionFB15k-237 (test)
Hits@1054.9
419
Link PredictionWN18RR (test)
Hits@1055.3
380
Knowledge Graph CompletionWN18RR
Hits@145
165
Knowledge Graph CompletionFB15k-237
Hits@100.549
108
Link PredictionFB15k-237 filtered (test)
Hits@100.549
60
Link PredictionWN18RR filtered (test)
Hits@100.553
57
Link PredictionKinship
MRR0.793
36
Showing 7 of 7 rows

Other info

Follow for update