Skip to main content
Erschienen in: Progress in Artificial Intelligence 1/2021

05.01.2021 | Regular Paper

A synchronous deep reinforcement learning model for automated multi-stock trading

verfasst von: Rasha AbdelKawy, Walid M. Abdelmoez, Amin Shoukry

Erschienen in: Progress in Artificial Intelligence | Ausgabe 1/2021

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Automated trading is one of the research areas that has benefited from the recent success of deep reinforcement learning (DRL) in solving complex decision-making problems. Despite the large number of researches done, casting the stock trading problem in a DRL framework still remains an open research area due to many reasons, including dynamic extraction of financial data features instead of handcrafted features, applying a scalable DRL technique that can benefit from the huge historical trading data available within a reasonable time frame and adopting an efficient trading strategy. In this paper, a novel multi-stock trading model is presented, based on free-model synchronous multi-agent deep reinforcement learning, which is able to interact with the trading market and to capture the financial market dynamics. The model can be executed on a standard personal computer with multiple core CPU or a GPU in a convenient time frame. The superiority of the proposed model is verified on datasets of different characteristics from the American stock market with huge historical trading data.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
4.
Zurück zum Zitat Gupta, P., Mehlawat, M.K., Mittal, G.: Asset portfolio optimization using support vector machines and real-coded genetic algorithm. J. Global Optim. 53(2), 297–315 (2012)MathSciNetCrossRef Gupta, P., Mehlawat, M.K., Mittal, G.: Asset portfolio optimization using support vector machines and real-coded genetic algorithm. J. Global Optim. 53(2), 297–315 (2012)MathSciNetCrossRef
7.
Zurück zum Zitat Chonga, E., Han, C., Parka, F.C.: Deep learning networks for stock market analysis and prediction: Methodology, data representations, and case studies. Exp. Syst. Appl. 83, 187–205 (2017)CrossRef Chonga, E., Han, C., Parka, F.C.: Deep learning networks for stock market analysis and prediction: Methodology, data representations, and case studies. Exp. Syst. Appl. 83, 187–205 (2017)CrossRef
10.
Zurück zum Zitat Jiang, W.: Applications of deep learning in stock market prediction: recent progress, arXiv:2003.01859 (2020), Preprint submitted to Elsevier Journal Jiang, W.: Applications of deep learning in stock market prediction: recent progress, arXiv:​2003.​01859 (2020), Preprint submitted to Elsevier Journal
12.
Zurück zum Zitat Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Baker, L., Lai, M., Bolton, A., Chen, Y., Lillicrap, T., Hui, F., Sifre, L., Driessche, G., Graepel, T., Hassabis, D.: Mastering the game of go without human knowledge. Int. J. Sci. Nat. 550, 354–359 (2017) Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Baker, L., Lai, M., Bolton, A., Chen, Y., Lillicrap, T., Hui, F., Sifre, L., Driessche, G., Graepel, T., Hassabis, D.: Mastering the game of go without human knowledge. Int. J. Sci. Nat. 550, 354–359 (2017)
14.
Zurück zum Zitat Watkins, C.J., Dayan, P.: Q-learning, Machine Learning, vol. 8, pp. 279–292. Springer, Berlin (1992)MATH Watkins, C.J., Dayan, P.: Q-learning, Machine Learning, vol. 8, pp. 279–292. Springer, Berlin (1992)MATH
15.
Zurück zum Zitat Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A.A., Veness, J., Bellemare, M.G., Graves, A., Riedmiller, M., Fidjeland, A.K., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., Hassabis, D.: Human-level control through deep reinforcement learning. Nature 518, 529–533 (2015)CrossRef Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A.A., Veness, J., Bellemare, M.G., Graves, A., Riedmiller, M., Fidjeland, A.K., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., Hassabis, D.: Human-level control through deep reinforcement learning. Nature 518, 529–533 (2015)CrossRef
16.
Zurück zum Zitat Schmid, Huber J.: Deep learning in neural networks: An overview. Neural Netw. V 61, 85–117 (2015)CrossRef Schmid, Huber J.: Deep learning in neural networks: An overview. Neural Netw. V 61, 85–117 (2015)CrossRef
17.
Zurück zum Zitat Hinton, G.E., Osindero, S., Teh, Y.W.: A fast learning algorithm for deep belief nets. Neural Comput. V18(7), 1527–1554 (2006)MathSciNetCrossRef Hinton, G.E., Osindero, S., Teh, Y.W.: A fast learning algorithm for deep belief nets. Neural Comput. V18(7), 1527–1554 (2006)MathSciNetCrossRef
18.
Zurück zum Zitat Bengio, Y., Lamblin, P., Popovici, D., Larochelle, H.: Greedy layer-wise training of deep networks. Adv. Neural Inf. Process. Syst. V(19), 153–160 (2006) Bengio, Y., Lamblin, P., Popovici, D., Larochelle, H.: Greedy layer-wise training of deep networks. Adv. Neural Inf. Process. Syst. V(19), 153–160 (2006)
19.
Zurück zum Zitat Hasselt, H.V.: Double Q-learning. Adv. Neural Inf. Process. Syst. 23, 2613–2621 (2010) Hasselt, H.V.: Double Q-learning. Adv. Neural Inf. Process. Syst. 23, 2613–2621 (2010)
20.
Zurück zum Zitat Wang, Z., Freitas, N., de., Lanctot, M.: Dueling network architectures for deep reinforcement learning, In the International Conference on Machine Learning (ICML), (2015). arXiv preprint arXiv:1511.06581 Wang, Z., Freitas, N., de., Lanctot, M.: Dueling network architectures for deep reinforcement learning, In the International Conference on Machine Learning (ICML), (2015). arXiv preprint arXiv:​1511.​06581
21.
Zurück zum Zitat Hessel, M., Modayil, J., van Hasselt, H., Schaul, T., Ostrovski, G., Dabney, W., Horgan, D., Piot, B., Azar, M., Silver, D.: Rainbow: Combining Improvements in Deep Reinforcement Learning, Thirty-Second AAAI Conference on Artificial Intelligence (2017). arXiv preprint arXiv:1710.02298 Hessel, M., Modayil, J., van Hasselt, H., Schaul, T., Ostrovski, G., Dabney, W., Horgan, D., Piot, B., Azar, M., Silver, D.: Rainbow: Combining Improvements in Deep Reinforcement Learning, Thirty-Second AAAI Conference on Artificial Intelligence (2017). arXiv preprint arXiv:​1710.​02298
22.
Zurück zum Zitat Sutton, R.S., McAllester, D.A., Singh, S.P., Mansour, Y.: Policy gradient methods for reinforcement learning with function approximation, Advances in Neural Information Processing Systems, Vol. 12, pp. 1057–1063. (NIPS 1999) MIT Press, Cambridge, MA (2000) Sutton, R.S., McAllester, D.A., Singh, S.P., Mansour, Y.: Policy gradient methods for reinforcement learning with function approximation, Advances in Neural Information Processing Systems, Vol. 12, pp. 1057–1063. (NIPS 1999) MIT Press, Cambridge, MA (2000)
23.
Zurück zum Zitat Lillicrap, T.P., Hunt, J.J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., Wierstra, D.: Continuous control with deep reinforcement learning, In: International Conference Learning Representations (2016). arXiv preprint arXiv:1509.02971 Lillicrap, T.P., Hunt, J.J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., Wierstra, D.: Continuous control with deep reinforcement learning, In: International Conference Learning Representations (2016). arXiv preprint arXiv:​1509.​02971
25.
Zurück zum Zitat Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms, Computing Research Repository (CoRR), 1707.06347 (2017). arXiv preprint arXiv:1707.06347 Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms, Computing Research Repository (CoRR), 1707.06347 (2017). arXiv preprint arXiv:​1707.​06347
26.
Zurück zum Zitat Wang, Z., Bapst, V., Heess, N., Mnih, V., Munos, R., Kavukcuoglu, K., Freitas, N.: Sample efficient actor-critic with experience replay,ICLR (2016). arXiv preprint arXiv:1611.01224 Wang, Z., Bapst, V., Heess, N., Mnih, V., Munos, R., Kavukcuoglu, K., Freitas, N.: Sample efficient actor-critic with experience replay,ICLR (2016). arXiv preprint arXiv:​1611.​01224
29.
Zurück zum Zitat Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmil-ler, M.: Playing atari with deep reinforcement learning, In NIPS Deep Learning Work-shop (2013) Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmil-ler, M.: Playing atari with deep reinforcement learning, In NIPS Deep Learning Work-shop (2013)
30.
Zurück zum Zitat Mnih, V., Badia, A.P., Mirza, M., Graves, A., Lillicrap, T.P., Harley, T., Silver, D., Kavukcuoglu, K.: Asynchronous methods for deep reinforcement learning, In: 33rd International Conference on Machine Learning, Vol. 48, pp. 1928-1937, PMLR (2016) Mnih, V., Badia, A.P., Mirza, M., Graves, A., Lillicrap, T.P., Harley, T., Silver, D., Kavukcuoglu, K.: Asynchronous methods for deep reinforcement learning, In: 33rd International Conference on Machine Learning, Vol. 48, pp. 1928-1937, PMLR (2016)
32.
Zurück zum Zitat Deng, Y., Bao, F., Youyong, K., Zhiquan, R., Qionghai, D.: Deep direct reinforcement learning for financial signal representation and trading. IEEE Trans. Neural Netw. Learn. Syst. 28(3), 653–664 (2017)CrossRef Deng, Y., Bao, F., Youyong, K., Zhiquan, R., Qionghai, D.: Deep direct reinforcement learning for financial signal representation and trading. IEEE Trans. Neural Netw. Learn. Syst. 28(3), 653–664 (2017)CrossRef
34.
Zurück zum Zitat Jiang, Z., Xu, D., Liang, J.: A deep reinforcement learning framework for the financial portfolio management problem, arXiv:1706.10059 (2017) Jiang, Z., Xu, D., Liang, J.: A deep reinforcement learning framework for the financial portfolio management problem, arXiv:​1706.​10059 (2017)
36.
Zurück zum Zitat Li, B., Sahoo, D., S. CH. Hoi.: Olps: A toolbox for online portfolio selection., J. Mach. Learn. Res. (JMLR), (2015) Li, B., Sahoo, D., S. CH. Hoi.: Olps: A toolbox for online portfolio selection., J. Mach. Learn. Res. (JMLR), (2015)
37.
Zurück zum Zitat Li, B., Hoi, S.C.H.: Online portfolio selection: A survey. ACM Comput. Surv. (CSUR) V46(3), 35 (2014)MATH Li, B., Hoi, S.C.H.: Online portfolio selection: A survey. ACM Comput. Surv. (CSUR) V46(3), 35 (2014)MATH
39.
Zurück zum Zitat Jiang, Z., Liang, J.: Cryptocurrency portfolio management with deep reinforcement learning., Intelligent Systems Conference., SAI Conferences,2017. Preprint: arXiv:1612.01277 Jiang, Z., Liang, J.: Cryptocurrency portfolio management with deep reinforcement learning., Intelligent Systems Conference., SAI Conferences,2017. Preprint: arXiv:​1612.​01277
40.
Zurück zum Zitat Liang, Z., Chen, H., Zhu, J., Jiang, K., Li, Y.: Adversarial Deep Reinforcement Learning in Portfolio Management, arXiv:1808.09940 (2018) Liang, Z., Chen, H., Zhu, J., Jiang, K., Li, Y.: Adversarial Deep Reinforcement Learning in Portfolio Management, arXiv:​1808.​09940 (2018)
41.
Zurück zum Zitat Hegde, S., Kumar, V., Singh, A.: Risk aware portfolio construction using deep deterministic policy gradients, IEEE Symposium Series on Computational Intelligence (2018) Hegde, S., Kumar, V., Singh, A.: Risk aware portfolio construction using deep deterministic policy gradients, IEEE Symposium Series on Computational Intelligence (2018)
42.
Zurück zum Zitat Wang, J., Zhang, Y., Tang, K., Wu, J., Xiong, Z.: AlphaStock: A Buying Winners-and-Selling-Losers Investment Strategy using Interpretable Deep Reinforcement Attention Networks, 25th ACM SIGKDD, pp.1900-1908 (2019) Wang, J., Zhang, Y., Tang, K., Wu, J., Xiong, Z.: AlphaStock: A Buying Winners-and-Selling-Losers Investment Strategy using Interpretable Deep Reinforcement Attention Networks, 25th ACM SIGKDD, pp.1900-1908 (2019)
43.
Zurück zum Zitat Li, Y., Zheng, W., Zheng, Z.: Deep Robust Reinforcement Learning for Practical Algorithmic Trading, IEEE Access, pp.108014–108022 (2019) Li, Y., Zheng, W., Zheng, Z.: Deep Robust Reinforcement Learning for Practical Algorithmic Trading, IEEE Access, pp.108014–108022 (2019)
44.
Zurück zum Zitat Soleymani, F., Elodie, P.: Financial portfolio optimization with online deep reinforcement learning and restricted stacked autoencoder - DeepBreath”. Exp. Syst. Appl. 156, 113456 (2020)CrossRef Soleymani, F., Elodie, P.: Financial portfolio optimization with online deep reinforcement learning and restricted stacked autoencoder - DeepBreath”. Exp. Syst. Appl. 156, 113456 (2020)CrossRef
Metadaten
Titel
A synchronous deep reinforcement learning model for automated multi-stock trading
verfasst von
Rasha AbdelKawy
Walid M. Abdelmoez
Amin Shoukry
Publikationsdatum
05.01.2021
Verlag
Springer Berlin Heidelberg
Erschienen in
Progress in Artificial Intelligence / Ausgabe 1/2021
Print ISSN: 2192-6352
Elektronische ISSN: 2192-6360
DOI
https://doi.org/10.1007/s13748-020-00225-z

Weitere Artikel der Ausgabe 1/2021

Progress in Artificial Intelligence 1/2021 Zur Ausgabe