KFS: KAN based adaptive Frequency Selection learning architecture for long term time series forecasting
About
Multi-scale decomposition architectures have emerged as predominant methodologies in time series forecasting. However, real-world time series exhibit noise interference across different scales, while heterogeneous information distribution among frequency components at varying scales leads to suboptimal multi-scale representation. Inspired by Kolmogorov-Arnold Networks (KAN) and Parseval's theorem, we propose a KAN based adaptive Frequency Selection learning architecture (KFS) to address these challenges. This framework tackles prediction challenges stemming from cross-scale noise interference and complex pattern modeling through its FreK module, which performs energy-distribution-based dominant frequency selection in the spectral domain. Simultaneously, KAN enables sophisticated pattern representation while timestamp embedding alignment synchronizes temporal representations across scales. The feature mixing module then fuses scale-specific patterns with aligned temporal features. Extensive experiments across multiple real-world time series datasets demonstrate that KT achieves state-of-the-art performance as a simple yet effective architecture.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multivariate long-term forecasting | ETTh1 | MSE0.428 | 394 | |
| Multivariate long-term series forecasting | ETTh2 | MSE0.367 | 367 | |
| Multivariate long-term series forecasting | Weather | MSE0.159 | 359 | |
| Multivariate long-term series forecasting | ETTm1 | MSE0.38 | 305 | |
| Multivariate long-term forecasting | Electricity | MSE0.178 | 236 | |
| Multivariate long-term series forecasting | ETTm2 | MSE0.274 | 223 | |
| Long-term time-series forecasting | Weather | Memory (MB)116 | 5 |