Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LIME: Making LLM Data More Efficient with Linguistic Metadata Embeddings

About

Pre-training decoder-only language models relies on vast amounts of high-quality data, yet the availability of such data is increasingly reaching its limits. While metadata is commonly used to create and curate these datasets, its potential as a direct training signal remains under-explored. We challenge this status quo and propose LIME (Linguistic Metadata Embeddings), a method that enriches token embeddings with metadata capturing syntax, semantics, and contextual properties. LIME substantially improves pre-training efficiency. Specifically, it adapts up to 56% faster to the training data distribution, while introducing only 0.01% additional parameters at negligible compute overhead. Beyond efficiency, LIME improves tokenization, leading to remarkably stronger language modeling capabilities and generative task performance. These benefits persist across model scales (500M to 2B). In addition, we develop a variant with shifted metadata, LIME+1, that can guide token generation. Given prior metadata for the next token, LIME+1 improves reasoning performance by up to 38% and arithmetic accuracy by up to 35%.

Sebastian Sztwiertnia, Felix Friedrich, Kristian Kersting, Patrick Schramowski, Bj\"orn Deiseroth• 2025

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy43.1
1460
Commonsense ReasoningWinoGrande
Accuracy54.6
776
Commonsense ReasoningPIQA
Accuracy69.4
647
Question AnsweringARC Easy
Normalized Acc58.1
385
Question AnsweringBoolQ
Accuracy59.4
240
Language ModelingLAMBADA
Accuracy49
183
Common Sense ReasoningCOPA
Accuracy64
138
Question AnsweringTriviaQA
Accuracy19.5
85
ReasoningFLenQA 250 tokens
Accuracy80
15
ReasoningFLenQA 500 tokens
Accuracy74
15
Showing 10 of 17 rows

Other info

Follow for update