This paper presents , a -based multivariate time series forecasting model that operates in an rthogonally transformed domain. Recent forecasting models typically adopt the temporal forecast (TF) paradigm, which directly encode and decode time series in the time domain. However, the entangled step-wise dependencies in series data can hinder the performance of TF. To address this, some forecasters conduct encoding and decoding in the transformed domain using fixed, dataset-independent bases (e.g., sine and cosine signals in the Fourier transform). In contrast, we utilize , a data-adaptive transformation based on an orthogonal matrix that diagonalizes the series' temporal Pearson correlation matrix. This approach enables more effective encoding and decoding in the decorrelated feature domain and can serve as a plug-in module to enhance existing forecasters. To enhance the representation learning for multivariate time series, we introduce a customized linear layer, , which employs a normalized weight matrix to capture multivariate dependencies. Empirically, the NormLin module shows a surprising performance advantage over multi-head self-attention, while requiring nearly half the FLOPs. Extensive experiments on 24 benchmarks and 140 forecasting tasks demonstrate that OLinear consistently achieves state-of-the-art performance with high efficiency. Notably, as a plug-in replacement for self-attention, the NormLin module consistently enhances Transformer-based forecasters. The code and datasets are available atthis https URL
View on arXiv@article{yue2025_2505.08550, title={ OLinear: A Linear Model for Time Series Forecasting in Orthogonally Transformed Domain }, author={ Wenzhen Yue and Yong Liu and Haoxuan Li and Hao Wang and Xianghua Ying and Ruohao Guo and Bowei Xing and Ji Shi }, journal={arXiv preprint arXiv:2505.08550}, year={ 2025 } }