Skip to main content

2024 | OriginalPaper | Buchkapitel

Vision Transformers for Computer Go

verfasst von : Amani Sagri, Tristan Cazenave, Jérôme Arjonilla, Abdallah Saffidine

Erschienen in: Applications of Evolutionary Computation

Verlag: Springer Nature Switzerland

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Motivated by transformers’ success in diverse fields like language understanding and image analysis, our investigation explores their potential in the game of Go. Specifically, we focus on analyzing Transformers in Vision. Through a comprehensive examination of factors like prediction accuracy, win rates, memory, speed, size, and learning rate, we underscore the significant impact transformers can make in the game of Go. Notably, our findings reveal that transformers outperform the previous state-of-the-art models, demonstrating superior performance metrics. This comparative study was conducted against conventional Residual Networks.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Browne, C., et al.: A survey of Monte Carlo tree search methods. IEEE Trans. Comput. Intell. AI in Games 4(1), 1–43 (2012)CrossRef Browne, C., et al.: A survey of Monte Carlo tree search methods. IEEE Trans. Comput. Intell. AI in Games 4(1), 1–43 (2012)CrossRef
2.
Zurück zum Zitat Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) Computer Vision – ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part I, pp. 213–229. Springer International Publishing, Cham (2020). https://doi.org/10.1007/978-3-030-58452-8_13CrossRef Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) Computer Vision – ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part I, pp. 213–229. Springer International Publishing, Cham (2020). https://​doi.​org/​10.​1007/​978-3-030-58452-8_​13CrossRef
3.
Zurück zum Zitat Cazenave, T.: Residual networks for computer go. IEEE Trans. Games 10(1), 107–110 (2018)CrossRef Cazenave, T.: Residual networks for computer go. IEEE Trans. Games 10(1), 107–110 (2018)CrossRef
4.
Zurück zum Zitat Cazenave, T.: Spatial average pooling for computer go. In: Computer Games: 7th Workshop, CGW 2018, IJCAI 2018, Stockholm, Sweden, July 13, 2018, Revised Selected Papers 7, pp. 119–126. Springer (2019) Cazenave, T.: Spatial average pooling for computer go. In: Computer Games: 7th Workshop, CGW 2018, IJCAI 2018, Stockholm, Sweden, July 13, 2018, Revised Selected Papers 7, pp. 119–126. Springer (2019)
5.
Zurück zum Zitat Cazenave, T.: Improving model and search for computer go. In: 2021 IEEE Conference on Games (CoG), pp. 1–8. IEEE (2021) Cazenave, T.: Improving model and search for computer go. In: 2021 IEEE Conference on Games (CoG), pp. 1–8. IEEE (2021)
6.
7.
Zurück zum Zitat Cazenave, T.: Mobile networks for computer go. IEEE Trans. Games 14(1), 76–84 (2022)CrossRef Cazenave, T.: Mobile networks for computer go. IEEE Trans. Games 14(1), 76–84 (2022)CrossRef
8.
Zurück zum Zitat Cazenave, T., et al.: Polygames: improved zero learning. J. Int. Comput. Games Assoc. 42(4), 244–256 (2020) Cazenave, T., et al.: Polygames: improved zero learning. J. Int. Comput. Games Assoc. 42(4), 244–256 (2020)
9.
Zurück zum Zitat Cazenave, T., Sentuc, J., Videau, M.: Cosine annealing, mixnet and swish activation for computer go. In: Browne, C., Kishimoto, A., Schaeffer, J. (eds.) Advances in Computer Games: 17th International Conference, ACG 2021, Virtual Event, November 23–25, 2021, Revised Selected Papers, pp. 53–60. Springer International Publishing, Cham (2022). https://doi.org/10.1007/978-3-031-11488-5_5CrossRef Cazenave, T., Sentuc, J., Videau, M.: Cosine annealing, mixnet and swish activation for computer go. In: Browne, C., Kishimoto, A., Schaeffer, J. (eds.) Advances in Computer Games: 17th International Conference, ACG 2021, Virtual Event, November 23–25, 2021, Revised Selected Papers, pp. 53–60. Springer International Publishing, Cham (2022). https://​doi.​org/​10.​1007/​978-3-031-11488-5_​5CrossRef
10.
Zurück zum Zitat Chen, X., Wu, Y., Wang, Z., Liu, S., Li, J.: Developing real-time streaming transformer transducer for speech recognition on large-scale dataset. In: ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5904–5908. IEEE (2021) Chen, X., Wu, Y., Wang, Z., Liu, S., Li, J.: Developing real-time streaming transformer transducer for speech recognition on large-scale dataset. In: ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5904–5908. IEEE (2021)
12.
Zurück zum Zitat Coulom, R.: Computing elo ratings of move patterns in the game of Go. ICGA J. 30(4), 198–208 (2007)CrossRef Coulom, R.: Computing elo ratings of move patterns in the game of Go. ICGA J. 30(4), 198–208 (2007)CrossRef
13.
Zurück zum Zitat Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4171–4186 (2019) Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4171–4186 (2019)
14.
Zurück zum Zitat Dong, L., Xu, S., Xu, B.: Speech-transformer: a no-recurrence sequence-to-sequence model for speech recognition. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (icassp), pp. 5884–5888. IEEE (2018) Dong, L., Xu, S., Xu, B.: Speech-transformer: a no-recurrence sequence-to-sequence model for speech recognition. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (icassp), pp. 5884–5888. IEEE (2018)
15.
Zurück zum Zitat Dosovitskiy, A.,et al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020) Dosovitskiy, A.,et al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:​2010.​11929 (2020)
16.
Zurück zum Zitat Gelly, S., Silver, D.: Monte-Carlo tree search and rapid action value estimation in computer Go. Artif. Intell. 175(11), 1856–1875 (2011)MathSciNetCrossRef Gelly, S., Silver, D.: Monte-Carlo tree search and rapid action value estimation in computer Go. Artif. Intell. 175(11), 1856–1875 (2011)MathSciNetCrossRef
17.
18.
Zurück zum Zitat He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016) He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
19.
Zurück zum Zitat Howard, A.G., et al.: Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017) Howard, A.G., et al.: Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:​1704.​04861 (2017)
21.
Zurück zum Zitat Li, Y., et al.: Efficientformer: vision transformers at mobilenet speed. Adv. Neural. Inf. Process. Syst. 35, 12934–12949 (2022) Li, Y., et al.: Efficientformer: vision transformers at mobilenet speed. Adv. Neural. Inf. Process. Syst. 35, 12934–12949 (2022)
22.
Zurück zum Zitat Parmar, N., et al.: Image transformer. In: International Conference on Machine Learning, pp. 4055–4064. PMLR (2018) Parmar, N., et al.: Image transformer. In: International Conference on Machine Learning, pp. 4055–4064. PMLR (2018)
23.
Zurück zum Zitat Qiu, X., Sun, T., Xu, Y., Shao, Y., Dai, N., Huang, X.: Pre-trained models for natural language processing: a survey. SCIENCE CHINA Technol. Sci. 63(10), 1872–1897 (2020)CrossRef Qiu, X., Sun, T., Xu, Y., Shao, Y., Dai, N., Huang, X.: Pre-trained models for natural language processing: a survey. SCIENCE CHINA Technol. Sci. 63(10), 1872–1897 (2020)CrossRef
24.
Zurück zum Zitat Rives, A., et al.: Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. Proc. Natl. Acad. Sci. 118(15), e2016239118 (2021)CrossRef Rives, A., et al.: Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. Proc. Natl. Acad. Sci. 118(15), e2016239118 (2021)CrossRef
25.
Zurück zum Zitat Schwaller, P., et al.: Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction. ACS Cent. Sci. 5(9), 1572–1583 (2019)CrossRef Schwaller, P., et al.: Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction. ACS Cent. Sci. 5(9), 1572–1583 (2019)CrossRef
26.
Zurück zum Zitat Silver, D.: Mastering the game of go with deep neural networks and tree search. Nature 529, 484–489 (2016)CrossRef Silver, D.: Mastering the game of go with deep neural networks and tree search. Nature 529, 484–489 (2016)CrossRef
28.
Zurück zum Zitat Silver, D., et al.: A general reinforcement learning algorithm that masters chess, shogi, and go through self-play. Science 362(6419), 1140–1144 (2018)MathSciNetCrossRef Silver, D., et al.: A general reinforcement learning algorithm that masters chess, shogi, and go through self-play. Science 362(6419), 1140–1144 (2018)MathSciNetCrossRef
29.
Zurück zum Zitat Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems, vol. 27 (2014) Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems, vol. 27 (2014)
30.
Zurück zum Zitat Vaswani, A., et al.: Attention is all you need. In: 31st International Conference on Neural Information Processing Systems (NIPS 2017), pp. 6000–6010 (2017) Vaswani, A., et al.: Attention is all you need. In: 31st International Conference on Neural Information Processing Systems (NIPS 2017), pp. 6000–6010 (2017)
32.
Zurück zum Zitat Wu, I.C., Wu, T.R., Liu, A.J., Guei, H., Wei, T.: On strength adjustment for MCTS-based programs. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 33, pp. 1222–1229 (2019) Wu, I.C., Wu, T.R., Liu, A.J., Guei, H., Wei, T.: On strength adjustment for MCTS-based programs. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 33, pp. 1222–1229 (2019)
33.
Zurück zum Zitat Wu, T.R., Wei, T.H., Wu, I.C.: Accelerating and improving AlphaZero using population based training. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 34, pp. 1046–1053 (2020) Wu, T.R., Wei, T.H., Wu, I.C.: Accelerating and improving AlphaZero using population based training. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 34, pp. 1046–1053 (2020)
34.
Zurück zum Zitat Wu, T.R., et al.: Multilabeled value networks for computer go. IEEE Trans. Games 10(4), 378–389 (2018)CrossRef Wu, T.R., et al.: Multilabeled value networks for computer go. IEEE Trans. Games 10(4), 378–389 (2018)CrossRef
Metadaten
Titel
Vision Transformers for Computer Go
verfasst von
Amani Sagri
Tristan Cazenave
Jérôme Arjonilla
Abdallah Saffidine
Copyright-Jahr
2024
DOI
https://doi.org/10.1007/978-3-031-56855-8_23

Premium Partner