Skip to main content
Top
Published in: Progress in Artificial Intelligence 1/2019

11-12-2018 | Regular Paper

Path planning of a mobile robot in a free-space environment using Q-learning

Authors: Jianxun Jiang, Jianbin Xin

Published in: Progress in Artificial Intelligence | Issue 1/2019

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

This paper proposes an improved Q-learning algorithm for the path planning of a mobile robot in a free-space environment. Existing Q-learning methods for path planning focus on the mesh routing environment; therefore, new methods must be developed for free-space environments in which robots move continuously. For the free-space environment, we construct fuzzified state variables for dividing the continuous space to avoid the curse of dimensionality. The state variables include the distances to the target point and obstacles and the heading of the robot. Based on the defined state variables, we propose an integrated learning strategy on the basis of the space allocation to accelerate the convergence during the learning process. Simulation experiments show that the path planning of mobile robots can be realized quickly, and the probability of obstacle collisions can be reduced. The results of the experiments also demonstrate the considerable advantages of the proposed learning algorithm compared to two commonly used methods.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Raja, P., Pugazhenthi, S.: Optimal path planning of mobile robots: a review. Int. J. Phys. Sci. 7(9), 1314–1320 (2012)CrossRef Raja, P., Pugazhenthi, S.: Optimal path planning of mobile robots: a review. Int. J. Phys. Sci. 7(9), 1314–1320 (2012)CrossRef
2.
go back to reference Rezaee, H., Abdollahi, F.: A decentralized cooperative control scheme with obstacle avoidance for a team of mobile robots. IEEE Trans. Ind. Electron. 61(1), 347–354 (2013)CrossRef Rezaee, H., Abdollahi, F.: A decentralized cooperative control scheme with obstacle avoidance for a team of mobile robots. IEEE Trans. Ind. Electron. 61(1), 347–354 (2013)CrossRef
3.
go back to reference Parasuraman, S., Ganapathy, V., Shirinzadeh, B.: Multiple sensors data integration using MFAM for mobile robot navigation. In: IEEE Congress on Evolutionary Computation, pp. 2421–2427 (2007) Parasuraman, S., Ganapathy, V., Shirinzadeh, B.: Multiple sensors data integration using MFAM for mobile robot navigation. In: IEEE Congress on Evolutionary Computation, pp. 2421–2427 (2007)
4.
go back to reference Cai, C., Ferrari, S.: Information-driven sensor path planning by approximate cell decomposition. IEEE Trans. Syst. Man Cybern. Part B Cybern. 39(3), 672–689 (2009)CrossRef Cai, C., Ferrari, S.: Information-driven sensor path planning by approximate cell decomposition. IEEE Trans. Syst. Man Cybern. Part B Cybern. 39(3), 672–689 (2009)CrossRef
5.
go back to reference Bhattacharya, S., Likhachev, M., Kumar, V.: Topological constraints in search-based robot path planning. Auton. Robots 33(3), 273–290 (2012)CrossRef Bhattacharya, S., Likhachev, M., Kumar, V.: Topological constraints in search-based robot path planning. Auton. Robots 33(3), 273–290 (2012)CrossRef
6.
go back to reference Zhang, G., Ferrari, S., Qian, M.: An information roadmap method for robotic sensor path planning. J. Intell. Robot. Syst. 56(1–2), 69–98 (2009)CrossRefMATH Zhang, G., Ferrari, S., Qian, M.: An information roadmap method for robotic sensor path planning. J. Intell. Robot. Syst. 56(1–2), 69–98 (2009)CrossRefMATH
7.
go back to reference Contreras-Cruz, M.A., Ayala-Ramirez, V., Hernandez-Belmonte, U.H.: Mobile robot path planning using artificial bee colony and evolutionary programming. Appl. Soft Comput. 30, 319–328 (2015)CrossRef Contreras-Cruz, M.A., Ayala-Ramirez, V., Hernandez-Belmonte, U.H.: Mobile robot path planning using artificial bee colony and evolutionary programming. Appl. Soft Comput. 30, 319–328 (2015)CrossRef
8.
go back to reference Tsai, C.C., Huang, H.C., Chan, C.K.: Parallel elite genetic algorithm and its application to global path planning for autonomous robot navigation. IEEE Trans. Ind. Electron. 58(10), 4813–4821 (2011)CrossRef Tsai, C.C., Huang, H.C., Chan, C.K.: Parallel elite genetic algorithm and its application to global path planning for autonomous robot navigation. IEEE Trans. Ind. Electron. 58(10), 4813–4821 (2011)CrossRef
9.
go back to reference Miao, H., Tian, Y.C.: Dynamic robot path planning using an enhanced simulated annealing approach. Appl. Math. Comput. 222(5), 420–437 (2013)MATH Miao, H., Tian, Y.C.: Dynamic robot path planning using an enhanced simulated annealing approach. Appl. Math. Comput. 222(5), 420–437 (2013)MATH
11.
go back to reference Sun, B., et al.: A novel fuzzy control algorithm for three-dimensional AUV path planning based on sonar model. J. Intell. Fuzzy Syst. 26(6), 2913–2926 (2014)MATH Sun, B., et al.: A novel fuzzy control algorithm for three-dimensional AUV path planning based on sonar model. J. Intell. Fuzzy Syst. 26(6), 2913–2926 (2014)MATH
12.
go back to reference Qu, H., et al.: Real-time robot path planning based on a modified pulse-coupled neural network model. IEEE Trans. Neural Netw. 20(11), 1724–39 (2009)CrossRef Qu, H., et al.: Real-time robot path planning based on a modified pulse-coupled neural network model. IEEE Trans. Neural Netw. 20(11), 1724–39 (2009)CrossRef
13.
go back to reference Zhu, Y., et al.: Target-driven visual navigation in indoor scenes using deep reinforcement learning. IEEE Int. Conf. Robot. Autom. 2017, 3357–3364 (2017) Zhu, Y., et al.: Target-driven visual navigation in indoor scenes using deep reinforcement learning. IEEE Int. Conf. Robot. Autom. 2017, 3357–3364 (2017)
14.
go back to reference Watkins, C.J., Dayan, P.: Q-learning. Mach. Learn. 8(3–4), 279–292 (1992)MATH Watkins, C.J., Dayan, P.: Q-learning. Mach. Learn. 8(3–4), 279–292 (1992)MATH
15.
go back to reference Gomes, E.R., Kowalczyk, R.: Modelling the dynamics of multiagent Q-learning with \(\epsilon \)-greedy exploration. In: International Conference on Autonomous Agents and Multiagent Systems, pp. 1181–1182 (2009) Gomes, E.R., Kowalczyk, R.: Modelling the dynamics of multiagent Q-learning with \(\epsilon \)-greedy exploration. In: International Conference on Autonomous Agents and Multiagent Systems, pp. 1181–1182 (2009)
16.
go back to reference Kim, I., et al.: Obstacle avoidance path planning for UAV using reinforcement learning under simulated environment. In: IASER 3rd International Conference on Electronics, Electrical Engineering, Computer Science, Okinawa, pp. 34–36 (2017) Kim, I., et al.: Obstacle avoidance path planning for UAV using reinforcement learning under simulated environment. In: IASER 3rd International Conference on Electronics, Electrical Engineering, Computer Science, Okinawa, pp. 34–36 (2017)
17.
go back to reference Konar, A., Indrani, G., Sapam, J.S., Lakhmi, C.J., Atulya, K.N.: A deterministic improved Q-learning for path planning of a mobile robot. IEEE Trans. Syst. Man Cybern. Syst. 43(5), 1141–1153 (2013)CrossRef Konar, A., Indrani, G., Sapam, J.S., Lakhmi, C.J., Atulya, K.N.: A deterministic improved Q-learning for path planning of a mobile robot. IEEE Trans. Syst. Man Cybern. Syst. 43(5), 1141–1153 (2013)CrossRef
18.
go back to reference Nirmalya, R., et al.: Implementation of image processing and reinforcement learning in path planning of mobile robots. Int. J. Eng. Sci. 7, 15211 (2017) Nirmalya, R., et al.: Implementation of image processing and reinforcement learning in path planning of mobile robots. Int. J. Eng. Sci. 7, 15211 (2017)
19.
go back to reference Wei, Q., Frank, L.L., Sun, Q., Yan, P., Song, R.: Discrete-time deterministic Q-learning: a novel convergence analysis. IEEE Trans. Cybern. 47(5), 1224–1237 (2017)CrossRef Wei, Q., Frank, L.L., Sun, Q., Yan, P., Song, R.: Discrete-time deterministic Q-learning: a novel convergence analysis. IEEE Trans. Cybern. 47(5), 1224–1237 (2017)CrossRef
20.
go back to reference Liu, J., Wei, Q., Xu, L.: Multi-step reinforcement learning algorithm of mobile robot path planning based on virtual potential field. In: International Conference of Pioneering Computer Scientists, Engineers and Educators, Singapore, pp. 528–538 (2017) Liu, J., Wei, Q., Xu, L.: Multi-step reinforcement learning algorithm of mobile robot path planning based on virtual potential field. In: International Conference of Pioneering Computer Scientists, Engineers and Educators, Singapore, pp. 528–538 (2017)
21.
go back to reference Li, S., Xin, X., Lei, Z.: Dynamic path planning of a mobile robot with improved Q-learning algorithm. In: IEEE International Conference on Information and Automation, pp. 409–411 (2015) Li, S., Xin, X., Lei, Z.: Dynamic path planning of a mobile robot with improved Q-learning algorithm. In: IEEE International Conference on Information and Automation, pp. 409–411 (2015)
22.
go back to reference Lee, K., et al.: Deep reinforcement learning in continuous action spaces: a case study in the game of simulated curling. In: International Conference on Machine Learning, pp. 2943–2952 (2018) Lee, K., et al.: Deep reinforcement learning in continuous action spaces: a case study in the game of simulated curling. In: International Conference on Machine Learning, pp. 2943–2952 (2018)
23.
go back to reference Luuk, A., Heskes, T., de Vries, A. P.: Comparing discretization methods for applying Q-learning in continuous state-action space (2017) Luuk, A., Heskes, T., de Vries, A. P.: Comparing discretization methods for applying Q-learning in continuous state-action space (2017)
24.
go back to reference Keogh, E., Mueen, A.: Curse of dimensionality. Ind. Eng. Chem 29(1), 48–53 (2017) Keogh, E., Mueen, A.: Curse of dimensionality. Ind. Eng. Chem 29(1), 48–53 (2017)
25.
go back to reference Cai, J., Yu, R., Cheng, L.: Autonomous navigation research for mobile robot. In: The World Congress on Intelligent Control and Automation, pp. 331–335 (2012) Cai, J., Yu, R., Cheng, L.: Autonomous navigation research for mobile robot. In: The World Congress on Intelligent Control and Automation, pp. 331–335 (2012)
26.
go back to reference Seising, R.: The Fuzzification of Systems: The Genesis of Fuzzy Set Theory and Its Initial Applications. Springer, Berlin (2007)MATH Seising, R.: The Fuzzification of Systems: The Genesis of Fuzzy Set Theory and Its Initial Applications. Springer, Berlin (2007)MATH
27.
28.
go back to reference Dearden, R., Friedman, N., Russell, S.: Bayesian Q-learning. In: AAAI/IAAI, pp. 761–768 (1998) Dearden, R., Friedman, N., Russell, S.: Bayesian Q-learning. In: AAAI/IAAI, pp. 761–768 (1998)
29.
go back to reference Kianercy, A., Galstyan, A.: Dynamics of Boltzmann Q-learning in two-player two-action games. Phys. Rev. E 85(4), 041145 (2012)CrossRef Kianercy, A., Galstyan, A.: Dynamics of Boltzmann Q-learning in two-player two-action games. Phys. Rev. E 85(4), 041145 (2012)CrossRef
30.
go back to reference Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (2018)MATH Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (2018)MATH
Metadata
Title
Path planning of a mobile robot in a free-space environment using Q-learning
Authors
Jianxun Jiang
Jianbin Xin
Publication date
11-12-2018
Publisher
Springer Berlin Heidelberg
Published in
Progress in Artificial Intelligence / Issue 1/2019
Print ISSN: 2192-6352
Electronic ISSN: 2192-6360
DOI
https://doi.org/10.1007/s13748-018-00168-6

Other articles of this Issue 1/2019

Progress in Artificial Intelligence 1/2019 Go to the issue

Premium Partner