Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Graph Language Models

About

While Language Models (LMs) are the workhorses of NLP, their interplay with structured knowledge graphs (KGs) is still actively researched. Current methods for encoding such graphs typically either (i) linearize them for embedding with LMs -- which underutilize structural information, or (ii) use Graph Neural Networks (GNNs) to preserve the graph structure -- but GNNs cannot represent text features as well as pretrained LMs. In our work we introduce a novel LM type, the Graph Language Model (GLM), that integrates the strengths of both approaches and mitigates their weaknesses. The GLM parameters are initialized from a pretrained LM to enhance understanding of individual graph concepts and triplets. Simultaneously, we design the GLM's architecture to incorporate graph biases, thereby promoting effective knowledge distribution within the graph. This enables GLMs to process graphs, texts, and interleaved inputs of both. Empirical evaluations on relation classification tasks show that GLM embeddings surpass both LM- and GNN-based baselines in supervised and zero-shot setting, demonstrating their versatility.

Moritz Plenz, Anette Frank• 2024

Related benchmarks

TaskDatasetResultRank
Relation label classificationConceptNet (CN)
Accuracy65.3
120
Source ClassificationWikidata (test)
Macro F186.2
12
Relation ClassificationWikidata (test)
Macro F185.28
12
Showing 3 of 3 rows

Other info

Code

Follow for update