Transformer-Based Deep Learning for Financial Time-Series Forecasting: A Multi-Horizon Prediction Framework

Authors

  • Julian R. Sterling, PhD Department of Systems Engineering, Colorado School of Mines
  • Elara V. Montgomery, PhD School of Computing and Information, University of Pittsburgh

Abstract

The rapid evolution of deep learning architectures has fundamentally altered the landscape of financial econometrics and predictive modeling. Traditional linear and autoregressive models, while foundational to financial theory, often fail to capture the high-frequency volatility, non-linear dependencies, and long-range temporal correlations inherent in modern globalized markets. This paper explores the transition toward attention-based mechanisms, specifically focusing on Transformer architectures as a robust framework for multi-horizon financial time-series forecasting. Unlike recurrent structures that suffer from vanishing gradients and sequential processing bottlenecks, the self-attention mechanism enables the simultaneous processing of vast historical datasets, facilitating the identification of structural breaks and regime shifts across multiple temporal scales. This research provides a comprehensive systems-level analysis of the integration of Transformer models within socio-technical financial infrastructures. We examine the architectural trade-offs between computational complexity and predictive accuracy, the role of positional encoding in preserving temporal order, and the systemic implications of deploying such models in high-stakes trading environments. Furthermore, the paper addresses critical dimensions of algorithmic governance, including the interpretability of attention weights, the ethical considerations of market-wide model convergence, and the environmental sustainability of large-scale deep learning deployments. By synthesizing insights from computer science, financial engineering, and public policy, we propose a multi-horizon framework that balances predictive power with systemic stability, offering a roadmap for the next generation of resilient financial AI systems.

References

1.Abu-Mostafa, Y. S., & Atiya, A. F. (1996). Introduction to financial forecasting. Applied Intelligence, 6(3), 205-213.

2.Arratia, A. (2014). Computational Finance: An Introductory Course with R. Atlantis Press.

3.Bai, S., Kolter, J. Z., & Koltun, V. (2018). An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271.

4.Bollerslev, T. (1986). Generalized autoregressive conditional heteroskedasticity. Journal of Econometrics, 31(3), 307-327.

5.Zhou, D. (2026). AI-Driven Hybrid SAST–DAST–SCA–IAST Framework for Risk-Based Vulnerability Prioritization in Microservice Architectures.

6.Box, G. E., Jenkins, G. M., Reinsel, G. C., & Ljung, G. M. (2015). Time Series Analysis: Forecasting and Control. John Wiley & Sons.

7.Brock, W. A., Lakonishok, J., & LeBaron, B. (1992). Simple technical trading rules and the stochastic properties of stock returns. The Journal of Finance, 47(5), 1731-1764.

8.Brownlee, J. (2018). Deep Learning for Time Series Forecasting. Machine Learning Mastery.

9.Qi, R. (2025). AUBIQ: A generative AI-powered framework for automating business intelligence requirements in resource-constrained enterprises. Frontiers in Business and Finance, 2(01), 66-86.

10.Chen, T., & Guestrin, C. (2016). XGBoost: A scalable tree boosting system. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.

11.Cont, R. (2001). Empirical properties of asset returns: Stylized facts and statistical issues. Quantitative Finance, 1(2), 223-236.

12.Dauphin, Y. N., Fan, A., Auli, M., & Grangier, D. (2017). Language modeling with gated convolutional networks. International Conference on Machine Learning.

13.Zhang, T. (2025). A Knowledge Graph-Enhanced Multimodal AI Framework for Intelligent Tax Data Integration and Compliance Enhancement. Frontiers in Business and Finance, 2(02), 247-261.

14.Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.

15.Diebold, F. X., & Mariano, R. S. (1995). Comparing predictive accuracy. Journal of Business & Economic Statistics, 13(3), 253-263.

16.Fischer, T., & Krauss, C. (2018). Deep learning with long short-term memory networks for financial market predictions. European Journal of Operational Research, 270(2), 654-669.

17.Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep Learning. MIT Press.

18.Qi, R. (2025, August). Interpretable Slow-Moving Inventory Forecasting: A Hybrid Neural Network Approach with Interactive Visualization. In Proceedings of the 2025 International Conference on Generative Artificial Intelligence for Business (pp. 41-46).

19.He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.

20.Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735-1780.

21.Yi, X. (2025, October). Real-Time Fair-Exposure Ad Allocation for SMBs and Underserved Creators via Contextual Bandits-with-Knapsacks. In Proceedings of the 2025 2nd International Conference on Digital Economy and Computer Science (pp. 1602-1607).

22.Hyndman, R. J., & Athanasopoulos, G. (2018). Forecasting: Principles and Practice. OTexts.

23.Kim, S. (2017). Financial series prediction using attention-based LSTM. arXiv preprint arXiv:1701.01887.

24.Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.

25.Liu, T. (2022, December). Financial Constraint’Impact on Firms’ ESG Rating Based on Chinese Stock Market. In 2022 4th International Conference on Economic Management and Cultural Industry (ICEMCI 2022) (pp. 1085-1095). Atlantis Press.

26.Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194), 20200209.

27.Lopez de Prado, M. (2018). Advances in Financial Machine Learning. John Wiley & Sons.

28.Makridakis, S., Spiliotis, E., & Assimakopoulos, V. (2018). The M4 Competition: Results, findings, conclusion and way forward. International Journal of Forecasting, 34(4), 596-608.

29.Tang, Y., Kojima, K., Gotoda, M., Nishikawa, S., Hayashi, S., Koike-Akino, T., ... & Klamkin, J. (2020, February). InP grating coupler design for vertical coupling of InP and silicon chips. In Integrated Optics: Devices, Materials, and Technologies XXIV (Vol. 11283, pp. 33-38). SPIE.

30.Paszke, A., et al. (2019). PyTorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems.

31.Rangapuram, S. S., et al. (2018). Deep state space models for time series forecasting. Advances in Neural Information Processing Systems.

32.Liu, T. (2026). Volatility Forecasting and Early-Warning Market Stress Detection: A Leakage-Safe Evaluation with Tree Ensembles and Transformers.

33.Salinas, D., Flunkert, V., Gasthaus, J., & Januschowski, T. (2020). DeepAR: Probabilistic forecasting with autoregressive recurrent networks. International Journal of Forecasting, 36(3), 1181-1191.

34.Schwartz, R., Dodge, J., Smith, N. A., & Etzioni, O. (2020). Green AI. Communications of the ACM, 63(12), 54-63.

35.Taylor, S. J. (2011). Asset Price Dynamics, Volatility, and Prediction. Princeton University Press.

36.Vaswani, A., et al. (2017). Attention is all you need. Advances in Neural Information Processing Systems.

37.Wen, R., et al. (2017). A multi-horizon quantile recurrent forecasting network. arXiv preprint arXiv:1711.11053.

38.Zhang, G. P. (2003). Time series forecasting using a hybrid ARIMA and neural network model. Neurocomputing, 50, 159-175.

39.Yi, X. (2026). A Federated and Differentially Private Incentive–Marketing Framework for Privacy-Preserving Cross-Channel Measurement in AI-Powered Digital Commerce.

40.Zhou, H., et al. (2021). Informer: Beyond efficient transformer for long sequence time-series forecasting. The Thirty-Fifth AAAI Conference on Artificial Intelligence.

41.Zoran, D., et al. (2020). Towards learning interpretable representations of self-attention. arXiv preprint arXiv:2004.14300.

Downloads

Published

2026-03-16

How to Cite

Julian R. Sterling, PhD, & Elara V. Montgomery, PhD. (2026). Transformer-Based Deep Learning for Financial Time-Series Forecasting: A Multi-Horizon Prediction Framework. International Journal of Artificial Intelligence Research, 1(1). Retrieved from https://isipress.org/index.php/IJAIR/article/view/80