Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Knowledge Enhanced Contextual Word Representations

About

Contextual word representations, typically trained on unstructured, unlabeled text, do not contain any explicit grounding to real world entities and are often unable to remember facts about those entities. We propose a general method to embed multiple knowledge bases (KBs) into large scale models, and thereby enhance their representations with structured, human-curated knowledge. For each KB, we first use an integrated entity linker to retrieve relevant entity embeddings, then update contextual word representations via a form of word-to-entity attention. In contrast to previous approaches, the entity linkers and self-supervised language modeling objective are jointly trained end-to-end in a multitask setting that combines a small amount of entity linking supervision with a large amount of raw text. After integrating WordNet and a subset of Wikipedia into BERT, the knowledge enhanced BERT (KnowBert) demonstrates improved perplexity, ability to recall facts as measured in a probing task and downstream performance on relationship extraction, entity typing, and word sense disambiguation. KnowBert's runtime is comparable to BERT's and it scales to large KBs.

Matthew E. Peters, Mark Neumann, Robert L. Logan IV, Roy Schwartz, Vidur Joshi, Sameer Singh, Noah A. Smith• 2019

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningPIQA
Accuracy66.61
647
Relation ExtractionTACRED (test)
F1 Score71.5
194
Common Sense ReasoningCOPA
Accuracy69.4
138
Relation ExtractionTACRED
Micro F171.5
97
Commonsense ReasoningSocialIQA
Accuracy65.4
97
Temporal Knowledge Graph Question AnsweringCRONQUESTIONS (test)
Hits@1 (Overall)22.6
77
Commonsense ReasoningOBQA
Accuracy58.5
75
Relation ExtractionSemEval (test)
Micro F189.1
55
Commonsense ReasoningCommonsenseQA (CSQA) v1.0 (test)
Accuracy53.88
46
Entity LinkingAIDA A
Micro F182.1
43
Showing 10 of 49 rows

Other info

Follow for update