Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Transformers without Normalization

About

Normalization layers are ubiquitous in modern neural networks and have long been considered essential. This work demonstrates that Transformers without normalization can achieve the same or better performance using a remarkably simple technique. We introduce Dynamic Tanh (DyT), an element-wise operation $DyT($x$) = \tanh(\alpha $x$)$, as a drop-in replacement for normalization layers in Transformers. DyT is inspired by the observation that layer normalization in Transformers often produces tanh-like, $S$-shaped input-output mappings. By incorporating DyT, Transformers without normalization can match or exceed the performance of their normalized counterparts, mostly without hyperparameter tuning. We validate the effectiveness of Transformers with DyT across diverse settings, ranging from recognition to generation, supervised to self-supervised learning, and computer vision to language models. These findings challenge the conventional understanding that normalization layers are indispensable in modern neural networks, and offer new insights into their role in deep networks.

Jiachen Zhu, Xinlei Chen, Kaiming He, Yann LeCun, Zhuang Liu• 2025

Related benchmarks

TaskDatasetResultRank
Question Answering and ReasoningDownstream Reasoning Suite (Arc-e, PIQA, Hellaswag, OpenBookQA, Winogrande, MMLU, BoolQ)
ARC-e29.78
14
Language ModelingPretraining Dataset
Train Loss (PT)3.709
10
Language Modeling and Zero-shot ReasoningStandard LLM Evaluation Suite ARC-e, PIQA, Hellaswag, OpenBookQA, Winogrande, MMLU, BoolQ
PT Eval Loss3.696
5
Pre-trainingPre-training (evaluation)
Pre-training Eval Loss3.855
5
Supervised Fine-tuningSFT (train)
SFT Train Loss3.361
5
Supervised Fine-tuningSFT (evaluation)
SFT Evaluation Loss3.971
5
Zero-shot EvaluationZero-shot Downstream Tasks (Arc-e, PIQA, Hellaswag, OpenBookQA, Winogrande, MMLU, BoolQ) Llama-1B Benchmark Suite (test)
Arc-e Accuracy29.28
5
Showing 7 of 7 rows

Other info

Follow for update