Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CoPE: Clipped RoPE as A Scalable Free Lunch for Long Context LLMs

About

Rotary Positional Embedding (RoPE) is a key component of context scaling in Large Language Models (LLMs). While various methods have been proposed to adapt RoPE to longer contexts, their guiding principles generally fall into two categories: (1) out-of-distribution (OOD) mitigation, which scales RoPE frequencies to accommodate unseen positions, and (2) Semantic Modeling, which posits that the attention scores computed with RoPE should always prioritize semantically similar tokens. In this work, we unify these seemingly distinct objectives through a minimalist intervention, namely CoPE: soft clipping lowfrequency components of RoPE. CoPE not only eliminates OOD outliers and refines semantic signals, but also prevents spectral leakage caused by hard clipping. Extensive experiments demonstrate that simply applying our soft clipping strategy to RoPE yields significant performance gains that scale up to 256k context length, validating our theoretical analysis and establishing CoPE as a new state-of-the-art for length generalization. Our code, data, and models are available at https://github.com/hrlics/CoPE.

Haoran Li, Sucheng Ren, Alan Yuille, Feng Wang• 2026

Related benchmarks

TaskDatasetResultRank
Language UnderstandingMMLU
Accuracy62.37
756
ReasoningBBH
Accuracy64.51
507
Question AnsweringGPQA
Accuracy29.31
258
Language UnderstandingMMLU-Pro
Accuracy34.05
70
Long-context language modelingHELMET
Summarization Score32.78
27
Math FindInfiniteBench
Performance (8k Context)35.43
3
KVInfiniteBench
KV Retrieval Score (8k)6.2
3
MKRULER
Performance @ 8k Context100
3
NIAHRULER
NIAH Score (8k Context)100
3
Long-context language modelingRULER
Accuracy (8k Context)81.5
2
Showing 10 of 10 rows

Other info

GitHub

Follow for update