Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Spectral Attention Steering for Prompt Highlighting

About

Attention steering is an important technique for controlling model focus, enabling capabilities such as prompt highlighting, where the model prioritises user-specified text. However, existing attention steering methods require explicit storage of the full attention matrix, making them incompatible with memory-efficient implementations like FlashAttention. We introduce Spectral Editing Key Amplification (SEKA), a training-free steering method that tackles this by directly editing key embeddings before attention computation. SEKA uses spectral decomposition to steer key embeddings towards latent directions that amplify attention scores for certain tokens. We extend this to Adaptive SEKA (AdaSEKA), a query-adaptive variant that uses a training-free routing mechanism to dynamically combine multiple expert subspaces based on the prompt's semantic intent. Our experiments show both methods significantly outperform strong baselines on standard steering benchmarks while adding much lower latency and memory overhead, in compatibility with optimised attention.

Weixian Waylon Li, Yuchen Niu, Yongxin Yang, Keshuang Li, Tiejun Ma, Shay B. Cohen• 2026

Related benchmarks

TaskDatasetResultRank
Knowledge EditingCounterFact
Efficacy99.08
301
Instruction FollowingPronoun Changing
P. Score99.88
40
Bias MitigationBias-in-Bios
Accuracy93.04
40
Bias classificationBiasBios
Accuracy93.04
35
Gender bias evaluationPronoun Change
Performance Score (P)98.66
35
Long-context retrievalLost-in-the-Middle 30-passage contexts
Average Exact Match60.86
20
Factual Knowledge EditingCounterFact (indices 0-5000)
SEKA99.08
5
Gender Bias MitigationBiasBios (indices 0-4999)
SEKA93.04
5
Pronoun SteeringPronoun Change (test)
SEKA98.66
5
Showing 9 of 9 rows

Other info

GitHub

Follow for update