Junliang Tao, Hongbing WangThis email address is being protected from spambots. You need JavaScript enabled to view it., Li Cao, Chenhao Xie, CJian Li, and Liang Zhou

School of Big Data and Computer Science, Guizhou Normal University, Guiyang, 550025, China


 

Received: November 25, 2025
Accepted: December 24, 2025
Publication Date: March 2, 2026

 Copyright The Author(s). This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are cited.


Download Citation: ||https://doi.org/10.6180/jase.202608_31.045  


Recently, Transformer-based models have achieved remarkable progress in time series forecasting, yet they still suffer from limitations in efficiency and scalability due to quadratic complexity. In contrast, the newly proposed Mamba model demonstrates strong potential with its linear-time complexity and selective state space mechanism, making it more suitable for long-sequence modeling. Meanwhile, convolutional neural networks (CNNs) have long been recognized for their effectiveness in capturing local dependencies and fine grained temporal patterns. To integrate the complementary advantages of these two paradigms, we propose PaDuM, a Patch-Based Dual-Stream Network that jointly leverages CNN and Mamba. Specifically, PaDuM first applies an exponential moving average (EMA) to adaptively decouple input sequences into trend and seasonal components, ensuring better interpretability and reduced noise. Each component is then patchified and modeled through dual streams, where CNN focuses on local seasonal variations while Mamba captures global trends with efficiency. To further enhance stability, we introduce a novel Sigmoid-based weight decay loss, which emphasizes recent predictions while preventing overfitting to distant horizons. Extensive experiments conducted on eight diversereal-world datasets spanning electricity consumption, traffic flow, and meteorological data consistently demonstrate that PaDuM achieves state-of-the-art forecasting performance, while maintaining strong robustness, scalability, and generalization ability across domains. The implementation and resources are publicly available at https://github.com/T-DXVN/PaDuM.


Keywords: Time Series Forecasting; State space model; CNN; Mamba; Patch-based Modeling


  1. [1] H. Zhou, S. Zhang, J. Peng, S. Zhang, J. Li, H. Xiong, and W. Zhang. “Informer: Beyond efficient trans former for long sequence time-series forecasting”. In: Proceedings of the AAAI conference on artificial intelligence. 35. 2021, 11106–11115. DOI: 10.1609/aaai.v35i12.17325.
  2. [2] Z. Karevan and J. A. Suykens, (2020) “Transductive LSTM for time-series prediction: An application to weather forecasting" Neural Networks 125: 1–9. DOI: 10.1016/j.neunet.2019.12.030.
  3. [3] L. Martín, L. F. Zarzalejo, J. Polo, A. Navarro, R. Marchante, and M. Cony, (2010) “Prediction of global solar irradiance based on time series analysis: Application to solar thermal power plants energy production planning" Solar energy 84(10): 1772–1781. DOI: 10.1016/j.solener.2010.07.002.
  4. [4] X.Yin,G.Wu,J.Wei,Y.Shen,H.Qi,andB.Yin,(2021) “Deeplearning on traffic prediction: Methods, analysis, and future directions" IEEE Transactions on Intelligent Transportation Systems 23(6): 4927–4943. DOI: 10.1109/TITS.2021.3054840.
  5. [5] A. A. Laghari, V. V. Estrela, and S. Yin, (2024) “How to collect and interpret medical pictures captured in highly challenging environments that range from nanoscale to hyperspectral imaging" Current Medical Imaging 20(1): e281222212228. DOI: 10.2174/1573405619666221228094228.
  6. [6] A. A. Laghari, Y. Sun, M. Alhussein, K. Aurangzeb, M. S. Anwar, and M. Rashid, (2023) “Deep residual dense network based on bidirectional recurrent neural net work for atrial fibrillation detection" Scientific Reports 13(1): 15109. DOI: 10.1038/s41598-023-40343-x.
  7. [7] A. A. Laghari, S. Shahid, R. Yadav, S. Karim, A. Khan, H. Li, andY.Shoulin, (2023) “The state of art and review on video streaming" Journal of High Speed Networks 29(3): 211–236. DOI: 10.3233/JHS-222087.
  8. [8] A. A. Laghari, V. V. Estrela, H. Li, Y. Shoulin, A. A. Khan, M. S. Anwar, A. Wahab, and K. Bouraqia, (2024) “Quality of experience assessment in virtual/augmented reality serious games for healthcare: A systematic literature review" Technology and Disability 36(1-2): 17–28. DOI: 10.3233/TAD-230035.
  9. [9] S. Yin, H. Li, L. Teng, A. A. Laghari, A. Almadhor, M. Gregus, and G. A. Sampedro, (2024) “Brain CT image classification based on mask RCNN and attention mechanism" Scientific Reports 14(1): 29300. DOI: 10.1038/s41598-024-78566-1.
  10. [10] M. A. Munir, R. A. Shah, M. Ali, A. A. Laghari, A. Almadhor, and T. R. Gadekallu, (2024) “Enhancing Gene Mutation Prediction with Sparse Regularized Au toencoders in Lung Cancer Radiomics Analysis" IEEE Access: DOI: 10.1109/ACCESS.2024.3523330.
  11. [11] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin. “Attention is all you need”. In: Advances in Neural Information Processing Systems. 30. 2017.
  12. [12] A. Gu, K. Goel, and C. Ré, (2021) “Efficiently modeling long sequences with structured state spaces" arXiv preprint arXiv:2111.00396: DOI: 10.48550/arXiv.2111.00396.
  13. [13] J. T. Smith, A. Warrington, and S. W. Linderman, (2022) “Simplified state space layers for sequence modeling" arXiv preprint arXiv:2208.04933: DOI: 10.48550/arXiv.2208.04933.
  14. [14] A. Gu, K. Goel, A. Gupta, and C. Ré, (2022) “On the Parameterization and Initialization of Diagonal State Space Models" Advances in Neural Information Processing Systems 35: 35971–35983.
  15. [15] A. Gu, I. Johnson, K. Goel, K. Saab, T. Dao, A. Rudra, and C. Ré, (2021) “Combining recurrent, convolutional, and continuous-time models with linear state space layers" Advances in neural information processing systems 34: 572–585.
  16. [16] A. Katharopoulos, A. Vyas, N. Pappas, and F. Fleuret. “Transformers are rnns: Fast autoregressive trans formers with linear attention”. In: International conference on machine learning. PMLR. 2020, 5156–5165.
  17. [17] T. Dao, D. Y. Fu, K. K. Saab, A. W. Thomas, A. Rudra, and C. Ré. “Hungry Hungry Hippos: Towards Lan guage Modeling with State Space Models”. In: Inter national Conference on Learning Representations. 2023. DOI: 10.48550/arXiv.2212.14052.
  18. [18] J. T. H. Smith, A. Warrington, and S. W. Linderman. “Simplified State Space Layers for Sequence Model ing”. In: International Conference on Learning Representations. 2023. DOI: 10.48550/arXiv.2208.04933.
  19. [19] A.GuandT.Dao,(2024)“Mamba:Linear-timesequence modeling with selective state spaces" Conference on Language Modeling:
  20. [20] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al. “An Image is Worth 16x16 Words: Transformers for Image Recogni tion at Scale”. In: International Conference on Learning Representations. 2020.
  21. [21] E. S. Gardner Jr, (1985) “Exponential smoothing: The state of the art" Journal of Forecasting 4(1): 1–28. DOI: 10.1002/for.3980040103.
  22. [22] J. Tao, L. Cao, H. Wang, C. Xie, J. Li, and L. Zhou, (2026) “PaDuM:Patch-Based Dual-Stream Network with CNN and Mambafor Time Series Forecasting" Submit ted to the 2026 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2026): DOI: 10.48550/arXiv.2411.05793.
  23. [23] H.Wu,J.Xu,J.Wang,andM.Long.“Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting”. In: Advances in Neural Information Processing Systems. 34. 2021, 22419–22430.
  24. [24] T. Zhou, Z. Ma, Q. Wen, X. Wang, L. Sun, and R. Jin. “Fed former: Frequency enhanced decomposed transformer for long-term series forecasting”. In: International Conference on Machine Learning. 162. Proceedings of Machine Learning Research. PMLR. 2022, 27268–27286.
  25. [25] A. Zeng, M. Chen, L. Zhang, and Q. Xu. “Are trans formers effective for time series forecasting?” In: Proceedings of the AAAI Conference on Artificial Intelligence. 37. 2023, 11121–11128. DOI: 10.1609/aaai.v37i9.26317.
  26. [26] A. Stitsyuk and J. Choi. “xPatch: Dual-Stream Time Series Forecasting with Exponential Seasonal-Trend Decomposition”. In: Proceedings of the AAAI Conference on Artificial Intelligence. 39. 2025, 20601–20609. DOI: 10.1609/aaai.v39i19.34270.
  27. [27] Y. Nie, H. N. Nguyen, P. Sinthong, and J. Kalagnanam. “A Time Series is Worth 64 Words: Long-term Forecasting with Transformers”. In: International Conference on Learning Representations. 2023. DOI: 10.48550/arXiv.2211.14730.
  28. [28] Y. Zhang and J. Yan. “Cross former: Transformer utilizing cross-dimension dependency for multivariate time series forecasting”. In: The Eleventh International Conference on Learning Representations (ICLR). 2022.
  29. [29] L. Sifre and S. Mallat, (2014) “Rigid-motion scattering for texture classification" International Journal of Computer Vision: DOI: 10.48550/arXiv.1403.1687.
  30. [30] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabi novich. “Going deeper with convolutions”. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2015.
  31. [31] S. Ioffe and C. Szegedy. “Batch normalization: Accelerating deep network training by reducing internal covariate shift”. In: International conference on machine learning. PMLR. 2015, 448–456.
  32. [32] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang,T.Weyand,M.Andreetto,andH.Adam,(2017) “Mobilenets: Efficient convolutional neural networks for mobile vision applications" Computer Vision and Pat tern Recognition: DOI: 10.48550/arXiv.1704.04861.
  33. [33] F. Chollet. “Xception: Deep learning with depth wise separable convolutions”. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2017.
  34. [34] A. Trockman and J. Z. Kolter, (2022) “Patches Are All You Need?" Transactions on Machine Learning Research: DOI: 10.48550/arXiv.2201.09792.
  35. [35] Z. Wanget al., (2025) “Is mamba effective for time series forecasting?" Neurocomputing 619: 129178. DOI: 10.1016/j.neucom.2024.129178.
  36. [36] M.A.AhamedandQ.Cheng,(2024) “TimeMachine: A Time Series is Worth 4 Mambas for Long-term Fore casting" ECAI 2024: 27th European Conference on Artificial Intelligence: DOI: 10.3233/faia240677.
  37. [37] A. Trindade, (2015) “ElectricityLoadDia grams20112014" UCI Machine Learning Repository 10: C58C86.
  38. [38] G. Lai, W.-C. Chang, Y. Yang, and H. Liu. “Modeling long-and short-term temporal patterns with deep neural networks”. In: The 41st international ACM SI GIR conference on research & development in information retrieval. 2018, 95–104. DOI: 10.1145/3209978.3210006.
  39. [39] X. Wang, T. Zhou, Q. Wen, J. Gao, B. Ding, and R. Jin. “CARD: Channel Aligned Robust Blend Trans former for Time Series Forecasting”. In: The Twelfth International Conference on Learning Representations. 2024. DOI: 10.48550/arXiv.2305.12095.
  40. [40] S. Wang, H. Wu, X. Shi, T. Hu, H. Luo, L. Ma, J. Y. Zhang, and J. ZHOU. “Time Mixer: Decomposable Multiscale Mixing for Time Series Forecasting”. In: International Conference on Learning Representations (ICLR). 2024. DOI: 10.48550/arXiv.2405.14616.
  41. [41] H. Wu, T. Hu, Y. Liu, H. Zhou, J. Wang, and M. Long. “TimesNet:Temporal2D-Variation Modeling for General Time Series Analysis”. In: International Conference on Learning Representations. 2023. DOI: 10.48550/arXiv.2210.02186.