Skip to main content
Top
Published in: Multimedia Systems 5/2023

13-04-2022 | Special Issue Paper

A MADDPG-based multi-agent antagonistic algorithm for sea battlefield confrontation

Authors: Wei Chen, Jing Nie

Published in: Multimedia Systems | Issue 5/2023

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

There is a concerted effort to build intelligent sea and numerous artificial intelligence technologies have been explored. At present, more and more people are engaged in the research of deep reinforcement learning algorithm, and its mainstream application is in the field of games. Reinforcement learning has conquered chess belonging to complete information game, and Texas poker belonging to incomplete information games. And it reached or even surpassed the highest player level of mankind in E-sports games with huge state space and complex action space. However, reinforcement learning algorithm still has great challenges in fields such as automatic driving. The main reason is that the training of reinforcement learning needs to build an environment for interacting with agents. However, it is very difficult to construct realistic simulation scenes, and there is no guarantee that we will not encounter the state that the agent has not seen. Therefore, it is necessary to explore the simulation scene first. Based on this, this paper mainly studies reinforcement learning in simulation scenario. There are huge challenges in migrating them to real scenario applications, especially in sea missions. Aiming at the heterogeneous multi-agent game confrontation scenario, this paper proposes a sea battlefield game confrontation decision algorithm based on multi-agent deep deterministic policy gradient. The algorithm combines long short-term memory and actor-critic, which not only realizes the convergence of the algorithm in huge state space and action space, but also solves the problem of sparse real rewards. At the same time, imitation learning is integrated into the decision algorithm, which not only improves the convergence speed of the algorithm, but also greatly improves the effectiveness of the algorithm. The results show that the algorithm can deal with a variety of different tactical sea battlefield scenarios, make flexible decisions according to the changes of the enemy, and the average winning rate is close to 90%.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Chu, T., Wang, J., Codecà, L., et al.: Multi-agent deep reinforcement learning for large-scale traffic signal control. IEEE Trans. Intell. Transp. Syst. 21(3), 1086–1095 (2019)CrossRef Chu, T., Wang, J., Codecà, L., et al.: Multi-agent deep reinforcement learning for large-scale traffic signal control. IEEE Trans. Intell. Transp. Syst. 21(3), 1086–1095 (2019)CrossRef
2.
go back to reference Yao, F., Jia, L.: A collaborative multi-agent reinforcement learning anti-jamming algorithm in wireless networks. IEEE Wirel. Commun. Lett. 8(4), 1024–1027 (2019)MathSciNetCrossRef Yao, F., Jia, L.: A collaborative multi-agent reinforcement learning anti-jamming algorithm in wireless networks. IEEE Wirel. Commun. Lett. 8(4), 1024–1027 (2019)MathSciNetCrossRef
3.
go back to reference Silver, D., Huang, A., Maddison, C.J., et al.: Mastering the game of go with deep neural networks and tree search. Nature 529(7587), 484–489 (2016)CrossRef Silver, D., Huang, A., Maddison, C.J., et al.: Mastering the game of go with deep neural networks and tree search. Nature 529(7587), 484–489 (2016)CrossRef
4.
go back to reference Silver, D., Schrittwieser, J., Simonyan, K., et al.: Mastering the game of go without human knowledge. Nature 550(7676), 354–359 (2017)CrossRef Silver, D., Schrittwieser, J., Simonyan, K., et al.: Mastering the game of go without human knowledge. Nature 550(7676), 354–359 (2017)CrossRef
5.
go back to reference Li, X., Lv, Z., Wang, S., et al.: A reinforcement learning model based on temporal difference algorithm. IEEE Access 7, 121922–121930 (2019)CrossRef Li, X., Lv, Z., Wang, S., et al.: A reinforcement learning model based on temporal difference algorithm. IEEE Access 7, 121922–121930 (2019)CrossRef
6.
go back to reference Silver, D., Hubert, T., Schrittwieser, J., et al.: A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play. Science 362(6419), 1140–1144 (2018)MathSciNetCrossRefMATH Silver, D., Hubert, T., Schrittwieser, J., et al.: A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play. Science 362(6419), 1140–1144 (2018)MathSciNetCrossRefMATH
8.
go back to reference Mnih, V., Kavukcuoglu, K., Silver, D., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)CrossRef Mnih, V., Kavukcuoglu, K., Silver, D., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)CrossRef
9.
10.
go back to reference Zhang, M., Zhang, Y., Gao, Z., et al.: An improved DDPG and its application based on the double-layer BP neural network. IEEE Access 8, 177734–177744 (2020)CrossRef Zhang, M., Zhang, Y., Gao, Z., et al.: An improved DDPG and its application based on the double-layer BP neural network. IEEE Access 8, 177734–177744 (2020)CrossRef
11.
go back to reference Font, J.M., Mahlmann, T.: Dota 2 bot competition. IEEE Trans. Games 11(3), 285–289 (2018)CrossRef Font, J.M., Mahlmann, T.: Dota 2 bot competition. IEEE Trans. Games 11(3), 285–289 (2018)CrossRef
12.
go back to reference Vinyals, O., Babuschkin, I., Czarnecki, W.M., et al.: Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature 575(7782), 350–354 (2019)CrossRef Vinyals, O., Babuschkin, I., Czarnecki, W.M., et al.: Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature 575(7782), 350–354 (2019)CrossRef
13.
go back to reference Wu, B., et al.: Hierarchical Macro Strategy Model for MOBA Game AI. Proceed. AAAI Conf. Artif. Intell. 33, 1206–1213 (2019) Wu, B., et al.: Hierarchical Macro Strategy Model for MOBA Game AI. Proceed. AAAI Conf. Artif. Intell. 33, 1206–1213 (2019)
14.
go back to reference Liu, H., Zhang, Z., Wang, D.: WRFMR: a multi-agent reinforcement learning method for cooperative tasks. IEEE Access 8, 216320–216331 (2020)CrossRef Liu, H., Zhang, Z., Wang, D.: WRFMR: a multi-agent reinforcement learning method for cooperative tasks. IEEE Access 8, 216320–216331 (2020)CrossRef
15.
go back to reference Cui, H., Zhang, Z.: A cooperative multi-agent reinforcement learning method based on coordination degree. IEEE Access 9, 123805–123814 (2021)CrossRef Cui, H., Zhang, Z.: A cooperative multi-agent reinforcement learning method based on coordination degree. IEEE Access 9, 123805–123814 (2021)CrossRef
16.
go back to reference Wen, J., Yang, J., Wang, T.: Path planning for autonomous underwater vehicles under the influence of ocean currents based on a fusion heuristic algorithm. IEEE Trans. Veh. Technol. 70(9), 8529–8544 (2021)CrossRef Wen, J., Yang, J., Wang, T.: Path planning for autonomous underwater vehicles under the influence of ocean currents based on a fusion heuristic algorithm. IEEE Trans. Veh. Technol. 70(9), 8529–8544 (2021)CrossRef
17.
go back to reference Yang, J., Wen, J., Wang, Y., et al.: Fog-based marine environmental information monitoring toward ocean of things. IEEE Internet Things J. 7(5), 4238–4247 (2019)CrossRef Yang, J., Wen, J., Wang, Y., et al.: Fog-based marine environmental information monitoring toward ocean of things. IEEE Internet Things J. 7(5), 4238–4247 (2019)CrossRef
18.
go back to reference Yang, J., Wen, J., Jiang, B., et al.: Blockchain-based sharing and tamper-proof framework of big data networking. IEEE Netw. 34(4), 62–67 (2020)CrossRef Yang, J., Wen, J., Jiang, B., et al.: Blockchain-based sharing and tamper-proof framework of big data networking. IEEE Netw. 34(4), 62–67 (2020)CrossRef
19.
go back to reference Yang, J., Guo, X., Li, Y., et al.: A survey of few-shot learning in smart agriculture: developments, applications, and challenges. Plant Methods 18(1), 1–12 (2022)CrossRef Yang, J., Guo, X., Li, Y., et al.: A survey of few-shot learning in smart agriculture: developments, applications, and challenges. Plant Methods 18(1), 1–12 (2022)CrossRef
22.
go back to reference Li, Y., Chao, X.: Toward sustainability: trade-off between data quality and quantity in crop pest recognition. Front. Plant Sci. 12, 811241 (2021)CrossRef Li, Y., Chao, X.: Toward sustainability: trade-off between data quality and quantity in crop pest recognition. Front. Plant Sci. 12, 811241 (2021)CrossRef
23.
go back to reference Li, Y., Chao, X.: Distance-entropy: an effective indicator for selecting informative data. Front. Plant Sci. 1, 8195 (2022) Li, Y., Chao, X.: Distance-entropy: an effective indicator for selecting informative data. Front. Plant Sci. 1, 8195 (2022)
Metadata
Title
A MADDPG-based multi-agent antagonistic algorithm for sea battlefield confrontation
Authors
Wei Chen
Jing Nie
Publication date
13-04-2022
Publisher
Springer Berlin Heidelberg
Published in
Multimedia Systems / Issue 5/2023
Print ISSN: 0942-4962
Electronic ISSN: 1432-1882
DOI
https://doi.org/10.1007/s00530-022-00922-w

Other articles of this Issue 5/2023

Multimedia Systems 5/2023 Go to the issue