Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Token Distillation: Attention-aware Input Embeddings For New Tokens

About

Current language models rely on static vocabularies determined at pretraining time, which can lead to decreased performance and increased computational cost for domains underrepresented in the original vocabulary. New tokens can be added to solve this problem, when coupled with a good initialization for their new embeddings. However, existing embedding initialization methods require expensive further training or pretraining of additional modules. In this paper, we propose Token Distillation and show that by distilling representations obtained using the original tokenization, we can quickly learn high-quality input embeddings for new tokens. Experimental results with a wide range of open-weight models show that Token Distillation outperforms even strong baselines.

Konstantin Dobler, Desmond Elliott, Gerard de Melo• 2025

Related benchmarks

TaskDatasetResultRank
Biomedical domain adaptationOpen Medical-LLM leaderboard
Macro Average71
84
Biomedical domain adaptationOpen Medical-LLM leaderboard macro-average
Macro Average Score71
75
Definition GenerationBiomedical domain tokens
Similarity Score83.8
75
Definition GenerationMulti-word tokens famous people, places, entities, sayings and concepts
Correctness86.3
66
Multiple-choice tasksFrenchBench
Accuracy81.5
61
Showing 5 of 5 rows

Other info

Follow for update