Accelerating Financial Intelligence via High Throughput Distributed Systems for Large Language Model Augmented Time Series Forecasting

Authors

  • Simon Ellsworth School of Engineering and Applied Sciences, Gonzaga University
  • Trevor Kingsley Department of Electrical Engineering and Computer Science, University of New Mexico

Keywords:

Distributed Systems, Financial Intelligence, Large Language Models, Time Series Forecasting, High-Throughput Inference, Socio-Technical Infrastructure, Algorithmic Governance.

Abstract

The integration of Large Language Models (LLMs) into financial time series forecasting represents a paradigm shift from purely frequentist econometric models to context-aware reasoning systems. While traditional quantitative methods excel at identifying statistical patterns in numerical data, they often fail to capture the nuanced causal drivers found in unstructured textual narratives. LLM-augmented forecasting addresses this gap by synthesizing market microstructure signals with macroeconomic sentiment. However, the deployment of such models in high-frequency financial environments is hindered by the significant computational latency of transformer-based architectures. This paper proposes a high-throughput distributed system architecture specifically optimized for LLM-augmented financial intelligence. We investigate the structural trade-offs between model quantization, speculative decoding, and distributed inference across heterogeneous compute clusters. The proposed framework emphasizes system-level robustness, hardware-aware orchestration, and the socio-technical implications of automated financial decision-making. By aligning high-throughput engineering with advanced linguistic reasoning, the system enables real-time forecasting that remains resilient to market non-stationarity. Furthermore, we examine the governance requirements for these systems, focusing on algorithmic fairness, environmental sustainability, and the evolving regulatory landscape for autonomous financial agents. The discussion concludes with a forward-looking perspective on the role of distributed systems in achieving equitable and stable global financial markets.

References

1.Abadi, M., Chu, A., Goodfellow, I., McMahan, H. B., Mironov, I., Talwar, K., & Zhang, L. (2016). Deep learning with differential privacy. Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, 308-318.

2.Acharya, V. V., & Richardson, M. (2009). Causes of the financial crisis. Critical Review, 21(2-3), 195-210.

3.Arumugam, R., & Bhargavi, R. (2019). A survey on modern trainable systems for time series forecasting. IEEE Access, 7, 70113-70135.

4.Bommasani, R., et al. (2021). On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258.

5.Brown, T., et al. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33, 1877-1901.

6.Cartea, A., Jaimungal, S., & Penalva, J. (2015). Algorithmic and High-Frequency Trading. Cambridge University Press.

7.Chen, L., & Zheng, Z. (2023). LLM-augmented financial analysis: Challenges and opportunities. Journal of Financial Data Science, 5(4), 12-28.

8.Dean, J., & Ghemawat, S. (2008). MapReduce: Simplified data processing on large clusters. Communications of the ACM, 51(1), 107-113.

9.Engle, R. F. (1982). Autoregressive conditional heteroscedasticity with estimates of the variance of United Kingdom inflation. Econometrica, 987-1007.

10.Ghoshal, B., & Tucker, A. (2022). Scalable inference for deep learning in finance. Quantitative Finance, 22(10), 1845-1860.

11.Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep Learning. MIT Press.

12.Goyal, N., et al. (2023). High-throughput inference for large language models: A systems perspective. ACM SIGOPS Operating Systems Review, 57(1), 45-56.

13.Hendershott, T., Jones, C. M., & Menkveld, A. J. (2011). Does algorithmic trading improve liquidity? The Journal of Finance, 66(1), 1-33.

14.Kaplan, J., et al. (2020). Scaling laws for neural language models. arXiv preprint arXiv:2001.08361.

15.Kirilenko, A. S., et al. (2017). The Flash Crash: High-frequency trading in an electronic market. The Journal of Finance, 72(3), 967-998.

16.Lo, A. W. (2017). Adaptive Markets: Financial Evolution at the Speed of Thought. Princeton University Press.

17.Lopez de Prado, M. (2018). Advances in Financial Machine Learning. Wiley.

18.Liu, T. (2026). Leakage-Safe Benchmark Design for Market-Stress Early Warning: An Economically Credible Evaluation.

19.Narayanan, D., et al. (2019). PipeDream: Generalized pipeline parallelism for DNN training. Proceedings of the 27th ACM Symposium on Operating Systems Principles.

20.O’Neil, C. (2016). Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown.

21.Pasquale, F. (2015). The Black Box Society: The Secret Algorithms That Control Money and Information. Harvard University Press.

22.Rajbhandari, S., et al. (2020). ZeRO: Memory optimizations toward training trillion parameter models. SC20: International Conference for High Performance Computing, Networking, Storage and Analysis.

23.Shalf, J. (2020). The future of computing beyond Moore’s Law. Philosophical Transactions of the Royal Society A, 378(2166).

24.Stoica, I., et al. (2017). Ray: A distributed framework for emerging AI applications. 13th USENIX Symposium on Operating Systems Design and Implementation.

25.Vaswani, A., et al. (2017). Attention is all you need. Advances in Neural Information Processing Systems, 30.

26.Wu, S., et al. (2023). BloombergGPT: A large language model for finance. arXiv preprint arXiv:2303.17564.

27.Zaharia, M., et al. (2012). Resilient distributed datasets: A fault-tolerant abstraction for in-memory cluster computing. 9th USENIX Symposium on Networked Systems Design and Implementation.

28.Zhang, L., et al. (2021). Deep reinforcement learning for automated stock trading: An ensemble strategy. SSRN Electronic Journal.

29.Zhou, Y., et al. (2022). Mixture-of-experts with exponential selection. arXiv preprint arXiv:2202.08906.

30.Mo, F., Haddadi, H., Katiyar, K., Ansari, R., & Chuah, C. N. (2021). PPFL: Privacy-preserving federated learning with trusted execution environments. Proceedings of the 19th Annual International Conference on Mobile Systems, Applications, and Services, 94-108.

31.Wang, J., et al. (2021). A field guide to federated optimization. arXiv preprint arXiv:2107.06917.

32.Rothchild, D., et al. (2020). FetchSGD: Communication-efficient federated learning with sketching. Proceedings of the 37th International Conference on Machine Learning.

33.Kairouz, P., et al. (2021). Advances and open problems in federated learning. Foundations and Trends in Machine Learning, 14(1-2), 1-210.

34.Yang, Q., Liu, Y., Chen, T., & Tong, Y. (2019). Federated machine learning: Concept and applications. ACM Transactions on Intelligent Systems and Technology (TIST), 10(2), 1-19.

35.Zhao, Y., Li, M., Lai, L., Suda, N., Civin, D., & Chandra, V. (2018). Federated learning with non-iid data. arXiv preprint arXiv:1806.00582.

36.Zuboff, S. (2019). The Age of Surveillance Capitalism: The Fight for a Human Future at the New Frontier of Power. PublicAffairs.

Downloads

Published

2026-04-09

How to Cite

Simon Ellsworth, & Trevor Kingsley. (2026). Accelerating Financial Intelligence via High Throughput Distributed Systems for Large Language Model Augmented Time Series Forecasting. International Journal of Artificial Intelligence Research, 1(1). Retrieved from https://isipress.org/index.php/IJAIR/article/view/112