Skip to main content
Erschienen in: Cognitive Computation 3/2020

02.01.2020

Indoor Topological Localization Based on a Novel Deep Learning Technique

verfasst von: Qiang Liu, Ruihao Li, Huosheng Hu, Dongbing Gu

Erschienen in: Cognitive Computation | Ausgabe 3/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Millions of people in the world suffer from vision impairment or vision loss. Traditionally, they rely on guide sticks or dogs to move around and avoid potential obstacles. However, both guide sticks and dogs are passive. They are unable to provide conceptual knowledge or semantic contents of an environment. To address this issue, this paper presents a vision-based cognitive system to support the independence of visually impaired people. More specifically, a 3D indoor semantic map is firstly constructed with a hand-held RGB-D sensor. The constructed map is then deployed for indoor topological localization. Convolutional neural networks are used for both semantic information extraction and location inference. Semantic information is used to further verify localization results and eliminate errors. The topological localization performance can be effectively improved despite significant appearance changes within an environment. Experiments have been conducted to demonstrate that the proposed method can increase both localization accuracy and recall rates. The proposed system can be potentially deployed by visually impaired people to move around safely and have independent life.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
2.
Zurück zum Zitat Schönberger JL, Pollefeys M, Geiger A, Sattler T. Semantic visual localization. IEEE Conference on computer vision and pattern recognition (CVPR); 2018. p. 6896–06. Schönberger JL, Pollefeys M, Geiger A, Sattler T. Semantic visual localization. IEEE Conference on computer vision and pattern recognition (CVPR); 2018. p. 6896–06.
3.
Zurück zum Zitat Liu Q, Li R, Hu H, Gu D. Extracting semantic information from visual data: a survey. Robotics 2016; 5(1):8.CrossRef Liu Q, Li R, Hu H, Gu D. Extracting semantic information from visual data: a survey. Robotics 2016; 5(1):8.CrossRef
4.
Zurück zum Zitat Endres F, Hess J, Sturm J, Cremers D, Burgard W. 3-D mapping with an RGB-D camera. IEEE Trans Robot 2014;30(1):177–87.CrossRef Endres F, Hess J, Sturm J, Cremers D, Burgard W. 3-D mapping with an RGB-D camera. IEEE Trans Robot 2014;30(1):177–87.CrossRef
5.
Zurück zum Zitat Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, et al. ImageNet large scale visual recognition challenge. Int J Comput Vis (IJCV) 2015;115(3):211–52.CrossRef Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, et al. ImageNet large scale visual recognition challenge. Int J Comput Vis (IJCV) 2015;115(3):211–52.CrossRef
6.
Zurück zum Zitat Lowry S, Sünderhauf N, Newman P, Leonard JJ, Cox D, Corke P, Milford M. Visual place recognition: a survey. IEEE Trans Robot 2016;32(1):1–9.CrossRef Lowry S, Sünderhauf N, Newman P, Leonard JJ, Cox D, Corke P, Milford M. Visual place recognition: a survey. IEEE Trans Robot 2016;32(1):1–9.CrossRef
7.
Zurück zum Zitat Biswas R, Limketkai B, Sanner S, Thrun S. Towards object mapping in non-stationary environments with mobile robots. 2002 IEEE/RSJ International C on Intelligent Robots and Systems (IROS). IEEE; 2002. p. 1014–9. Biswas R, Limketkai B, Sanner S, Thrun S. Towards object mapping in non-stationary environments with mobile robots. 2002 IEEE/RSJ International C on Intelligent Robots and Systems (IROS). IEEE; 2002. p. 1014–9.
8.
Zurück zum Zitat Brucker M, Durner M, Ambruş R, Márton ZC, Wendt A, Jensfelt P, Arras KO, Triebel R. Semantic labeling of indoor environments from 3D RGB maps. IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2018. p. 1871–78. Brucker M, Durner M, Ambruş R, Márton ZC, Wendt A, Jensfelt P, Arras KO, Triebel R. Semantic labeling of indoor environments from 3D RGB maps. IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2018. p. 1871–78.
9.
Zurück zum Zitat Garg S, Suenderhauf N, Milford M. Don’t look back: robustifying place categorization for viewpoint-and condition-invariant place recognition. IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2018. p. 3645–52. Garg S, Suenderhauf N, Milford M. Don’t look back: robustifying place categorization for viewpoint-and condition-invariant place recognition. IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2018. p. 3645–52.
10.
Zurück zum Zitat Lee T-j, Kim C-h, Cho D-iD. A monocular vision sensor-based efficient slam method for indoor service robots. IEEE Trans Ind Electron 2019;66(1):318–28.CrossRef Lee T-j, Kim C-h, Cho D-iD. A monocular vision sensor-based efficient slam method for indoor service robots. IEEE Trans Ind Electron 2019;66(1):318–28.CrossRef
11.
Zurück zum Zitat Bay H, Ess A, Tuytelaars T, Van Gool L. Speeded-up robust features (SURF). Comput Vis Image Understand 2008;110(3):346–59.CrossRef Bay H, Ess A, Tuytelaars T, Van Gool L. Speeded-up robust features (SURF). Comput Vis Image Understand 2008;110(3):346–59.CrossRef
12.
Zurück zum Zitat Bloesch M, Czarnowski J, Clark R, Leutenegger S, Davison AJ. CodeSLAM—learning a compact, optimisable representation for dense visual SLAM. IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2018. p. 2560–68. Bloesch M, Czarnowski J, Clark R, Leutenegger S, Davison AJ. CodeSLAM—learning a compact, optimisable representation for dense visual SLAM. IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2018. p. 2560–68.
13.
Zurück zum Zitat Li R, Wang S, Long Z, Gu D. Undeepvo: monocular visual odometry through unsupervised deep learning. IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2018. p. 7286–91. Li R, Wang S, Long Z, Gu D. Undeepvo: monocular visual odometry through unsupervised deep learning. IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2018. p. 7286–91.
14.
Zurück zum Zitat Liu Q, Li R, Hu H, Gu D. Using unsupervised deep learning technique for monocular visual odometry. IEEE Access 2019;7:18076–88.CrossRef Liu Q, Li R, Hu H, Gu D. Using unsupervised deep learning technique for monocular visual odometry. IEEE Access 2019;7:18076–88.CrossRef
15.
Zurück zum Zitat Li R, Wang S, Gu D. Ongoing evolution of visual slam from geometry to deep learning: challenges and opportunities. Cogn Comput 2018;10(6):875–89.CrossRef Li R, Wang S, Gu D. Ongoing evolution of visual slam from geometry to deep learning: challenges and opportunities. Cogn Comput 2018;10(6):875–89.CrossRef
16.
Zurück zum Zitat Grimmett H, Buerki M, Paz L, Pinies P, Furgale P, Posner I, Newman P. Integrating metric and semantic maps for vision-only automated parking. 2015 IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2015. p. 2159–66. Grimmett H, Buerki M, Paz L, Pinies P, Furgale P, Posner I, Newman P. Integrating metric and semantic maps for vision-only automated parking. 2015 IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2015. p. 2159–66.
17.
Zurück zum Zitat Hart JW, Shah R, Kirmani S, Walker N, Baldauf K, John N, Stone P. PRISM: pose registration for integrated semantic mapping. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE; 2018. p. 896–902. Hart JW, Shah R, Kirmani S, Walker N, Baldauf K, John N, Stone P. PRISM: pose registration for integrated semantic mapping. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE; 2018. p. 896–902.
18.
Zurück zum Zitat Maturana D, Chou P. -W., Uenoyama M, Scherer S. Real-time semantic mapping for autonomous off-road navigation. Field and service robotics. Springer; 2018. p. 335–50. Maturana D, Chou P. -W., Uenoyama M, Scherer S. Real-time semantic mapping for autonomous off-road navigation. Field and service robotics. Springer; 2018. p. 335–50.
19.
Zurück zum Zitat Arroyo R, Alcantarilla PF, Bergasa LM, Romera E. Are you able to perform a life-long visual topological localization? Auton Robot 2018;42(3):665–85.CrossRef Arroyo R, Alcantarilla PF, Bergasa LM, Romera E. Are you able to perform a life-long visual topological localization? Auton Robot 2018;42(3):665–85.CrossRef
20.
Zurück zum Zitat Li Y, Hu Z, Hu Y, Chu D. Integration of vision and topological self-localization for intelligent vehicles. Mechatronics 2018;51:46–58.CrossRef Li Y, Hu Z, Hu Y, Chu D. Integration of vision and topological self-localization for intelligent vehicles. Mechatronics 2018;51:46–58.CrossRef
21.
Zurück zum Zitat Huang AS, Bachrach A, Henry P, Krainin M, Maturana D, Fox D, Roy N. Visual odometry and mapping for autonomous flight using an RGB-D camera. Robotics research. Springer; 2017 . p. 235–52. Huang AS, Bachrach A, Henry P, Krainin M, Maturana D, Fox D, Roy N. Visual odometry and mapping for autonomous flight using an RGB-D camera. Robotics research. Springer; 2017 . p. 235–52.
22.
Zurück zum Zitat Sharif Razavian A, Azizpour H, Sullivan J, Carlsson S. CNN features off-the-shelf: an astounding baseline for recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops; 2014. p. 806–13. Sharif Razavian A, Azizpour H, Sullivan J, Carlsson S. CNN features off-the-shelf: an astounding baseline for recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops; 2014. p. 806–13.
23.
Zurück zum Zitat Wen G, Hou Z, Li H, Li D, Jiang L, Xun E. 2017. Ensemble of deep neural networks with probability-based fusion for facial expression recognition. Cogn Comput, 1–4. Wen G, Hou Z, Li H, Li D, Jiang L, Xun E. 2017. Ensemble of deep neural networks with probability-based fusion for facial expression recognition. Cogn Comput, 1–4.
24.
Zurück zum Zitat Yue Z, Gao F, Xiong Q, Wang J, Huang T, Yang E, Zhou H. 2019. A novel semi-supervised convolutional neural network method for synthetic aperture radar image recognition. Cogn Comput, 1–2. Yue Z, Gao F, Xiong Q, Wang J, Huang T, Yang E, Zhou H. 2019. A novel semi-supervised convolutional neural network method for synthetic aperture radar image recognition. Cogn Comput, 1–2.
25.
Zurück zum Zitat Ren P, Sun W, Luo C, Hussain A. 2017. Clustering-oriented multiple convolutional neural networks for single image super-resolution. Cogn Comput, 1–4. Ren P, Sun W, Luo C, Hussain A. 2017. Clustering-oriented multiple convolutional neural networks for single image super-resolution. Cogn Comput, 1–4.
26.
Zurück zum Zitat Li R, Gu D, Liu Q, Long Z, Hu H. 2017. Semantic scene mapping with spatio-temporal deep neural network for robotic applications. Cogn Comput, 1–2. Li R, Gu D, Liu Q, Long Z, Hu H. 2017. Semantic scene mapping with spatio-temporal deep neural network for robotic applications. Cogn Comput, 1–2.
27.
Zurück zum Zitat Chen X, Kundu K, Zhu Y, Ma H, Fidler S, Urtasun R. 3D object proposals using stereo imagery for accurate object class detection. IEEE Trans Pattern Anal Mach Intell 2018;40(5):1259–72.CrossRef Chen X, Kundu K, Zhu Y, Ma H, Fidler S, Urtasun R. 3D object proposals using stereo imagery for accurate object class detection. IEEE Trans Pattern Anal Mach Intell 2018;40(5):1259–72.CrossRef
28.
Zurück zum Zitat Sunderhauf N, Shirazi S, Dayoub F, Upcroft B, Milford M. On the performance of ConvNet features for place recognition. 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE; 2015. p. 4297–304. Sunderhauf N, Shirazi S, Dayoub F, Upcroft B, Milford M. On the performance of ConvNet features for place recognition. 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE; 2015. p. 4297–304.
30.
Zurück zum Zitat Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A. Going deeper with convolutions. IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2015. p. 1–9. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A. Going deeper with convolutions. IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2015. p. 1–9.
31.
Zurück zum Zitat Zeng Z, Li Z, Cheng D, Zhang H, Zhan K, Yang Y. Two-stream multirate recurrent neural network for video-based pedestrian reidentification. IEEE Trans Indus Inform 2018;14(7):3179–86.CrossRef Zeng Z, Li Z, Cheng D, Zhang H, Zhan K, Yang Y. Two-stream multirate recurrent neural network for video-based pedestrian reidentification. IEEE Trans Indus Inform 2018;14(7):3179–86.CrossRef
32.
Zurück zum Zitat Kümmerle R, Grisetti G, Strasdat H, Konolige K, Burgard W. g2o: a general framework for graph optimization. 2011 IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2011. p. 3607–13. Kümmerle R, Grisetti G, Strasdat H, Konolige K, Burgard W. g2o: a general framework for graph optimization. 2011 IEEE International Conference on Robotics and Automation (ICRA). IEEE; 2011. p. 3607–13.
33.
Zurück zum Zitat Fischler MA, Bolles RC. Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun ACM 1981;24(6):381–95.CrossRef Fischler MA, Bolles RC. Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun ACM 1981;24(6):381–95.CrossRef
34.
Zurück zum Zitat Zender H, Mozos OM, Jensfelt P, Kruijff G-J, Burgard W. Conceptual spatial representations for indoor mobile robots. Robot Auton Syst 2008;56(6):493–502.CrossRef Zender H, Mozos OM, Jensfelt P, Kruijff G-J, Burgard W. Conceptual spatial representations for indoor mobile robots. Robot Auton Syst 2008;56(6):493–502.CrossRef
38.
Zurück zum Zitat Zhou B, Lapedriza A, Khosla A, Oliva A, Torralba A. Places: a 10 million image database for scene recognition. IEEE Trans Pattern Anal Mach Intell 2018;40(6):1452–64.CrossRef Zhou B, Lapedriza A, Khosla A, Oliva A, Torralba A. Places: a 10 million image database for scene recognition. IEEE Trans Pattern Anal Mach Intell 2018;40(6):1452–64.CrossRef
Metadaten
Titel
Indoor Topological Localization Based on a Novel Deep Learning Technique
verfasst von
Qiang Liu
Ruihao Li
Huosheng Hu
Dongbing Gu
Publikationsdatum
02.01.2020
Verlag
Springer US
Erschienen in
Cognitive Computation / Ausgabe 3/2020
Print ISSN: 1866-9956
Elektronische ISSN: 1866-9964
DOI
https://doi.org/10.1007/s12559-019-09693-5

Weitere Artikel der Ausgabe 3/2020

Cognitive Computation 3/2020 Zur Ausgabe