Skip to main content
Top

2021 | OriginalPaper | Chapter

Place Inference via Graph-Based Decisions on Deep Embeddings and Blur Detections

Authors : Piotr Wozniak, Bogdan Kwolek

Published in: Computational Science – ICCS 2021

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Current approaches to visual place recognition for loop closure do not provide information about confidence of decisions. In this work we present an algorithm for place recognition on the basis of graph-based decisions on deep embeddings and blur detections. The graph constructed in advance permits together with information about the room category an inference on usefulness of place recognition, and in particular, it enables the evaluation the confidence of final decision. We demonstrate experimentally that thanks to proposed blur detection the accuracy of scene recognition is much higher. We evaluate performance of place recognition on the basis of manually selected places for recognition with corresponding sets of relevant and irrelevant images. The algorithm has been evaluated on large dataset for visual place recognition that contains both images with severe (unknown) blurs and sharp images. Images with 6-DOF viewpoint variations were recorded using a humanoid robot.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Cadena, C., Carlone, L., Carrillo, H., Latif, Y., Scaramuzza, D., Neira, J., Reid, I., Leonard, J.: Past, present, and future of simultaneous localization and mapping: towards the robust-perception age. IEEE Trans. Robot. 32(6), 1309–1332 (2016)CrossRef Cadena, C., Carlone, L., Carrillo, H., Latif, Y., Scaramuzza, D., Neira, J., Reid, I., Leonard, J.: Past, present, and future of simultaneous localization and mapping: towards the robust-perception age. IEEE Trans. Robot. 32(6), 1309–1332 (2016)CrossRef
2.
go back to reference Cebollada, S., Paya, L., Flores, M., Peidro, A., Reinoso, O.:A state-of-the-art review on mobile robotics tasks using artificial intelligence and visual data. Expert Syst. Appl. 167, 114–195 (2020) Cebollada, S., Paya, L., Flores, M., Peidro, A., Reinoso, O.:A state-of-the-art review on mobile robotics tasks using artificial intelligence and visual data. Expert Syst. Appl. 167, 114–195 (2020)
3.
go back to reference Lowry, S., Sünderhauf, N., Newman, P., Leonard, J., Cox, D., Corke, P., Milford, M.J.: Visual place recognition: a survey. IEEE Trans. Robot 32, 1–19 (2016)CrossRef Lowry, S., Sünderhauf, N., Newman, P., Leonard, J., Cox, D., Corke, P., Milford, M.J.: Visual place recognition: a survey. IEEE Trans. Robot 32, 1–19 (2016)CrossRef
4.
go back to reference Odo, A., McKenna, S., Flynn, D., Vorstius, J.: Towards the automatic visual monitoring of electricity pylons from aerial images. In: 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISAPP) (2020) Odo, A., McKenna, S., Flynn, D., Vorstius, J.: Towards the automatic visual monitoring of electricity pylons from aerial images. In: 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISAPP) (2020)
5.
go back to reference Zhao, J., et al. J.:Place recognition with deep superpixel features for brain-inspire dnavigation. Rev. Sci. Instrum. 91(12), 125110 (2020) Zhao, J., et al. J.:Place recognition with deep superpixel features for brain-inspire dnavigation. Rev. Sci. Instrum. 91(12), 125110 (2020)
6.
go back to reference Tolias, G., Avrithis, Y., Jégou, H.: Image search with selective match kernels: aggregation across single and multiple images. Int. J. Comput. Vision 116(3), 247–261 (2015)MathSciNetCrossRef Tolias, G., Avrithis, Y., Jégou, H.: Image search with selective match kernels: aggregation across single and multiple images. Int. J. Comput. Vision 116(3), 247–261 (2015)MathSciNetCrossRef
7.
go back to reference Ovalle-Magallanes, E., Aldana-Murillo, N.G., Avina-Cervantes, J.G., Ruiz-Pinales, J., Cepeda-Negrete, J., Ledesma, S.: Transfer learning for humanoid robot appearance-based localization in a visual map. IEEE Access 9, 6868–6877 (2021)CrossRef Ovalle-Magallanes, E., Aldana-Murillo, N.G., Avina-Cervantes, J.G., Ruiz-Pinales, J., Cepeda-Negrete, J., Ledesma, S.: Transfer learning for humanoid robot appearance-based localization in a visual map. IEEE Access 9, 6868–6877 (2021)CrossRef
8.
go back to reference Pretto, A., Menegatti, E., Bennewitz, M., Burgard, W., Pagello, E.: A visual odometry framework robust to motion blur. In: IEEE International Conference on Robotics and Automation, pp. 2250–2257(2009) Pretto, A., Menegatti, E., Bennewitz, M., Burgard, W., Pagello, E.: A visual odometry framework robust to motion blur. In: IEEE International Conference on Robotics and Automation, pp. 2250–2257(2009)
9.
go back to reference Torii, A., Arandjelović, R., Sivic, J., Okutomi, M., Pajdla, T.: 24/7 place recognition by view synthesis. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1808–1817 (2015) Torii, A., Arandjelović, R., Sivic, J., Okutomi, M., Pajdla, T.: 24/7 place recognition by view synthesis. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1808–1817 (2015)
10.
go back to reference Maffra, F., Chen, Z., Chli, M.: Viewpoint-tolerant place recognition combining 2D and 3D information for UAV navigation. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 2542–2549(2018) Maffra, F., Chen, Z., Chli, M.: Viewpoint-tolerant place recognition combining 2D and 3D information for UAV navigation. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 2542–2549(2018)
11.
go back to reference Garg, S., Milford, M.: Straightening sequence-search for appearance-invariant place recognition using robust motion estimation. In: Proceedings of Australasian Conference on Robotics and Automation (ACRA), pp. 203–212 (2017) Garg, S., Milford, M.: Straightening sequence-search for appearance-invariant place recognition using robust motion estimation. In: Proceedings of Australasian Conference on Robotics and Automation (ACRA), pp. 203–212 (2017)
12.
go back to reference Chen, Z., Lam, O., Adam, J., Milford, M.: Convolutional neural network-based place recognition. In: Proceedings of Australasian Conference on Robotics and Automation, pp. 1–8 (2014) Chen, Z., Lam, O., Adam, J., Milford, M.: Convolutional neural network-based place recognition. In: Proceedings of Australasian Conference on Robotics and Automation, pp. 1–8 (2014)
13.
go back to reference Suenderhauf, N., Shirazi, S., Dayoub, F., Upcroft, B., Milford, M.: On the performance of ConvNet features for place recognition. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4297–4304 (2015) Suenderhauf, N., Shirazi, S., Dayoub, F., Upcroft, B., Milford, M.: On the performance of ConvNet features for place recognition. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4297–4304 (2015)
14.
go back to reference Arandjelovic, R., Gronat, P., Torii, A., Pajdla, T., Sivic, J.: NetVLAD: CNN architecture for weakly supervised place recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40(6), 1437–1451 (2018) Arandjelovic, R., Gronat, P., Torii, A., Pajdla, T., Sivic, J.: NetVLAD: CNN architecture for weakly supervised place recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40(6), 1437–1451 (2018)
15.
go back to reference Arandjelovic, R., Zisserman, A.: All About VLAD. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1578–1585. IEEE Computer Society (2013) Arandjelovic, R., Zisserman, A.: All About VLAD. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1578–1585. IEEE Computer Society (2013)
16.
go back to reference Zaffar, M., Khaliq, A., Ehsan, S., Milford, M., McDonald-Maier, K.: Levelling the playing field: a comprehensive comparison of visual place recognition approaches under changing conditions. CoRR abs/1207.0016 (2019) Zaffar, M., Khaliq, A., Ehsan, S., Milford, M., McDonald-Maier, K.: Levelling the playing field: a comprehensive comparison of visual place recognition approaches under changing conditions. CoRR abs/1207.0016 (2019)
17.
go back to reference Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., Torralba, A.: Places: A 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40(6), 1452–1464 (2018) Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., Torralba, A.: Places: A 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40(6), 1452–1464 (2018)
18.
go back to reference López-Cifuentes, A., Escudero-Vin̄olo, M., Bescós, J., Álvaro García-Martín: Semantic-aware scene recognition. Pattern Recogn. 102, 107256 (2020) López-Cifuentes, A., Escudero-Vin̄olo, M., Bescós, J., Álvaro García-Martín: Semantic-aware scene recognition. Pattern Recogn. 102, 107256 (2020)
19.
go back to reference Oliva, A., Torralba, A.: Modeling the shape of the scene: a holistic representation of the spatial envelope. Int J. Comput. Vision 42(3), 145–175 (2001)CrossRef Oliva, A., Torralba, A.: Modeling the shape of the scene: a holistic representation of the spatial envelope. Int J. Comput. Vision 42(3), 145–175 (2001)CrossRef
20.
go back to reference Yandex, A.B., Lempitsky, V.: Aggregating local deep features for image retrieval. In: IEEE International Conference on Computer Vision (ICCV), pp. 1269–1277 (2015) Yandex, A.B., Lempitsky, V.: Aggregating local deep features for image retrieval. In: IEEE International Conference on Computer Vision (ICCV), pp. 1269–1277 (2015)
21.
go back to reference Ma, J., Jiang, X., Fan, A., Jiang, J., Yan, J.: Image matching from handcrafted to deep features: a survey. Int. J. Comput. Vision 129(1), 23–79 (2020)MathSciNetCrossRef Ma, J., Jiang, X., Fan, A., Jiang, J., Yan, J.: Image matching from handcrafted to deep features: a survey. Int. J. Comput. Vision 129(1), 23–79 (2020)MathSciNetCrossRef
22.
go back to reference Kwolek, B.: Visual odometry based on gabor filters and sparse bundle adjustment. In: Proceedings IEEE International Conference on Robotics and Automation, pp. 3573–3578 (2007) Kwolek, B.: Visual odometry based on gabor filters and sparse bundle adjustment. In: Proceedings IEEE International Conference on Robotics and Automation, pp. 3573–3578 (2007)
23.
go back to reference Arth, C., Pirchheim, C., Ventura, J., Schmalstieg, D., Lepetit, V.: Instant outdoor localization and SLAM initialization from 2.5d maps. IEEE Trans. Visual. Comput. Graph. 21(11), 1309–1318 (2015) Arth, C., Pirchheim, C., Ventura, J., Schmalstieg, D., Lepetit, V.: Instant outdoor localization and SLAM initialization from 2.5d maps. IEEE Trans. Visual. Comput. Graph. 21(11), 1309–1318 (2015)
24.
go back to reference Chen, Z., Maffra, F., Sa, I., Chli, M.: Only look once, mining distinctive landmarks from ConvNet for visual place recognition. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). pp. 9–16 (2017) Chen, Z., Maffra, F., Sa, I., Chli, M.: Only look once, mining distinctive landmarks from ConvNet for visual place recognition. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). pp. 9–16 (2017)
25.
go back to reference Hou, Y., Zhang, H., Zhou, S.: Evaluation of object proposals and ConvNet features for landmark-based visual place recognition. J. Intell. Robot. Syst. 92(3–4), 505–520 (2017) Hou, Y., Zhang, H., Zhou, S.: Evaluation of object proposals and ConvNet features for landmark-based visual place recognition. J. Intell. Robot. Syst. 92(3–4), 505–520 (2017)
27.
go back to reference Tolias, G., Avrithis, Y., Jégou, H.: To aggregate or not to aggregate: selective match kernels for image search. In: IEEE International Conference on Computer Vision, pp. 1401–1408 (2013) Tolias, G., Avrithis, Y., Jégou, H.: To aggregate or not to aggregate: selective match kernels for image search. In: IEEE International Conference on Computer Vision, pp. 1401–1408 (2013)
28.
go back to reference Mao, J., Hu, X., He, X., Zhang, L., Wu, L., Milford, M.J.: Learning to fuse multiscale features for visual place recognition. IEEE Access 7, 5723–5735 (2019)CrossRef Mao, J., Hu, X., He, X., Zhang, L., Wu, L., Milford, M.J.: Learning to fuse multiscale features for visual place recognition. IEEE Access 7, 5723–5735 (2019)CrossRef
29.
go back to reference Camara, L.G., Pr̆euc̆il, L.: Visual place recognition by spatial matching of high-level CNN features. Robot. Auton. Syst. 133, 103625 (2020) Camara, L.G., Pr̆euc̆il, L.: Visual place recognition by spatial matching of high-level CNN features. Robot. Auton. Syst. 133, 103625 (2020)
30.
go back to reference Wozniak, P., Afrisal, H., Esparza, R.G., Kwolek, B.: Scene recognition for indoor localization of mobile robots using deep CNN. In: Chmielewski, L.J., Kozera, R., Orłowski, A., Wojciechowski, K., Bruckstein, A.M., Petkov, N. (eds.) ICCVG 2018. LNCS, vol. 11114, pp. 137–147. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00692-1_13CrossRef Wozniak, P., Afrisal, H., Esparza, R.G., Kwolek, B.: Scene recognition for indoor localization of mobile robots using deep CNN. In: Chmielewski, L.J., Kozera, R., Orłowski, A., Wojciechowski, K., Bruckstein, A.M., Petkov, N. (eds.) ICCVG 2018. LNCS, vol. 11114, pp. 137–147. Springer, Cham (2018). https://​doi.​org/​10.​1007/​978-3-030-00692-1_​13CrossRef
31.
go back to reference Quattoni, A., Torralba, A.: Recognizing indoor scenes. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 413–420 (2009) Quattoni, A., Torralba, A.: Recognizing indoor scenes. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 413–420 (2009)
32.
go back to reference Tolias, G., Sicre, R., Jégou, H.: Particular object retrieval with integral max-pooling of CNN activations. In: International Conference Learning Representations (ICLR 2016) (2016) Tolias, G., Sicre, R., Jégou, H.: Particular object retrieval with integral max-pooling of CNN activations. In: International Conference Learning Representations (ICLR 2016) (2016)
33.
go back to reference Sivic, J., Zisserman, A.: Video Google: a text retrieval approach to object matching in videos. In: IEEE International Conference on Computer Vision, pp. 1470–1477(2003) Sivic, J., Zisserman, A.: Video Google: a text retrieval approach to object matching in videos. In: IEEE International Conference on Computer Vision, pp. 1470–1477(2003)
34.
go back to reference Zhong, C., Malinen, M., Miao, D., Fränti, P.: A fast minimum spanning tree algorithm based on k-means. Inf. Sci. 295(C), 1–17 (2015) Zhong, C., Malinen, M., Miao, D., Fränti, P.: A fast minimum spanning tree algorithm based on k-means. Inf. Sci. 295(C), 1–17 (2015)
36.
go back to reference Narvekar, N., Karam, L.: A no-reference image blur metric based on the cumulative probability of blur detection (CPBD). IEEE Trans. Image Process. 20(9), 2678–2683 (2011)MathSciNetCrossRef Narvekar, N., Karam, L.: A no-reference image blur metric based on the cumulative probability of blur detection (CPBD). IEEE Trans. Image Process. 20(9), 2678–2683 (2011)MathSciNetCrossRef
37.
go back to reference Pech-Pacheco, J.L., Cristobal, G., Chamorro-Martinez, J., Fernandez-Valdivia, J.: Diatom autofocusing in brightfield microscopy: a comparative study. In: Proceedings of the 15th International coneference on Pattern Recognition, vol. 3, pp. 314–317 (2000) Pech-Pacheco, J.L., Cristobal, G., Chamorro-Martinez, J., Fernandez-Valdivia, J.: Diatom autofocusing in brightfield microscopy: a comparative study. In: Proceedings of the 15th International coneference on Pattern Recognition, vol. 3, pp. 314–317 (2000)
38.
go back to reference Sun, J., Wenfei Cao, Zongben Xu, Ponce, J.: Learning a convolutional neural network for non-uniform motion blur removal. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 769–777 (2015) Sun, J., Wenfei Cao, Zongben Xu, Ponce, J.: Learning a convolutional neural network for non-uniform motion blur removal. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 769–777 (2015)
39.
go back to reference Cun, X., Pun, C.M.: Defocus blur detection via depth distillation. In: European Conference on Computer Vision (ECCV), pp. 747–763. Springer (2020) Cun, X., Pun, C.M.: Defocus blur detection via depth distillation. In: European Conference on Computer Vision (ECCV), pp. 747–763. Springer (2020)
40.
go back to reference Tao, X., Gao, H., Shen, X., Wang, J., Jia, J.: Scale-recurrent network for deep image deblurring. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 8174–8182 (2018) Tao, X., Gao, H., Shen, X., Wang, J., Jia, J.: Scale-recurrent network for deep image deblurring. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 8174–8182 (2018)
Metadata
Title
Place Inference via Graph-Based Decisions on Deep Embeddings and Blur Detections
Authors
Piotr Wozniak
Bogdan Kwolek
Copyright Year
2021
DOI
https://doi.org/10.1007/978-3-030-77977-1_14

Premium Partner