Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

EMFormer: Efficient Multi-Scale Transformer for Accumulative Context Weather Forecasting

About

Long-term weather forecasting is critical for socioeconomic planning and disaster preparedness. While recent approaches employ finetuning to extend prediction horizons, they remain constrained by the issues of catastrophic forgetting, error accumulation, and high training overhead. To address these limitations, we present a novel pipeline across pretraining, finetuning and forecasting to enhance long-context modeling while reducing computational overhead. First, we introduce an Efficient Multi-scale Transformer (EMFormer) to extract multi-scale features through a single convolution in both training and inference. Based on the new architecture, we further employ an accumulative context finetuning to improve temporal consistency without degrading short-term accuracy. Additionally, we propose a composite loss that dynamically balances different terms via a sinusoidal weighting, thereby adaptively guiding the optimization trajectory throughout pretraining and finetuning. Experiments show that our approach achieves strong performance in weather forecasting and extreme event prediction, substantially improving long-term forecast accuracy. Moreover, EMFormer demonstrates strong generalization on vision benchmarks (ImageNet-1K and ADE20K) while delivering a 5.69x speedup over conventional multi-scale modules.

Hao Chen, Tao Han, Jie Zhang, Song Guo, Fenghua Ling, Lei Bai• 2026

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K (val)
mIoU49.6
2731
Weather forecastingERA5 1.4° grid (test)
RMSE0.0599
30
Typhoon Track ForecastingCMA Typhoon Track Forecasts 2024
AMPIL66.7
10
Weather forecastingWeather data 0.25° resolution, 721x1440
GPU Cost (Inference, M)962.2
4
Weather forecastingWeather data 1.4° resolution, 128x256
Inference GPU Cost (M)7.03e+8
4
Showing 5 of 5 rows

Other info

Follow for update