Skip to main content
Erschienen in: 3D Research 2/2017

01.06.2017 | 3DR Express

A Hierarchical Optimization Algorithm Based on GPU for Real-Time 3D Reconstruction

verfasst von: Jin-hua Lin, Lu Wang, Yan-jie Wang

Erschienen in: 3D Research | Ausgabe 2/2017

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In machine vision sensing system, it is important to realize high-quality real-time 3D reconstruction in large-scale scene. The recent online approach performed well, but scaling up the reconstruction, it causes pose estimation drift, resulting in the cumulative error, usually requiring a large number of off-line operation to completely correct the error, reducing the reconstruction performance. In order to optimize the traditional volume fusion method and improve the old frame-to-frame pose estimation strategy, this paper presents a real-time CPU to Graphic Processing Unit reconstruction system. Based on a robust camera pose estimation strategy, the algorithm fuses all the RGB-D input values into an effective hierarchical optimization framework, and optimizes each frame according to the global camera attitude, eliminating the serious dependence on the tracking timeliness and continuously tracking globally optimized frames. The system estimates the global optimization of gestures (bundling) in real-time, supports for robust tracking recovery (re-positioning), and re-estimation of large-scale 3D scenes to ensure global consistency. It uses a set of sparse corresponding features, geometric and ray matching functions in one of the parallel optimization systems. The experimental results show that the average reconstruction time is 415 ms per frame, the ICP pose is estimated 20 times in 100.0 ms. For large scale 3D reconstruction scene, the system performs well in online reconstruction area, keeping the reconstruction accuracy at the same time.

Graphical Abstract

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Weise, T., Wismer, T., & Leibe, B., et al. (2009). In-hand scanning with online loop closure. In IEEE international conference on computer vision workshops (pp. 1630–1637). Weise, T., Wismer, T., & Leibe, B., et al. (2009). In-hand scanning with online loop closure. In IEEE international conference on computer vision workshops (pp. 1630–1637).
2.
Zurück zum Zitat Henry, P., Krainin, M., Herbst, E., et al. (2012). RGB-D mapping: Using depth cameras for dense 3d modeling of indoor environments. International Journal of Robotics Research, 31(5), 647–663.CrossRef Henry, P., Krainin, M., Herbst, E., et al. (2012). RGB-D mapping: Using depth cameras for dense 3d modeling of indoor environments. International Journal of Robotics Research, 31(5), 647–663.CrossRef
3.
Zurück zum Zitat Keller, M., Lefloch, D., & Lambers, M., et al. (2013). Real-time 3D reconstruction in dynamic scenes using point-based fusion. In International conference on 3D Vision-3DV (Vol. 8768, Issue 2, pp. 1–8). Keller, M., Lefloch, D., & Lambers, M., et al. (2013). Real-time 3D reconstruction in dynamic scenes using point-based fusion. In International conference on 3D Vision-3DV (Vol. 8768, Issue 2, pp. 1–8).
4.
Zurück zum Zitat Whelan, T., Leutenegger, S., & Salas-Moreno, R. F., et al. (2015). ElasticFusion: Dense SLAM without a pose graph. Robotics: Science and Systems (RSS). Whelan, T., Leutenegger, S., & Salas-Moreno, R. F., et al. (2015). ElasticFusion: Dense SLAM without a pose graph. Robotics: Science and Systems (RSS).
5.
Zurück zum Zitat Merrell, P., Akbarzadeh, A., & Wang, L., et al. (2007). Real-time visibilitybased fusion of depth maps. In IEEE international conference on computer vision (Vol. 8, pp. 1–8). Merrell, P., Akbarzadeh, A., & Wang, L., et al. (2007). Real-time visibilitybased fusion of depth maps. In IEEE international conference on computer vision (Vol. 8, pp. 1–8).
6.
Zurück zum Zitat Meilland, M., & Comport, A. (2013). On unifying key-frame and voxel-based dense visual slam at large scales. IEEE/RSJ International Conference on Intelligent Robots & Systems, 8215(2), 3677–3683. Meilland, M., & Comport, A. (2013). On unifying key-frame and voxel-based dense visual slam at large scales. IEEE/RSJ International Conference on Intelligent Robots & Systems, 8215(2), 3677–3683.
7.
Zurück zum Zitat Gallup, D., Pollefeys, M., & Frahm, J. M. (2010). 3D reconstruction using an n-layer heightmap. Pattern Recognition, 6376, 1–10. Gallup, D., Pollefeys, M., & Frahm, J. M. (2010). 3D reconstruction using an n-layer heightmap. Pattern Recognition, 6376, 1–10.
8.
Zurück zum Zitat Wurm, K. M., Hornung, A., & Bennewitz, M., et al. (2010). OctoMap: A probabilistic, flexible, and compact 3D map representation for robotic systems. In IEEE international conference on robotics and automation. Wurm, K. M., Hornung, A., & Bennewitz, M., et al. (2010). OctoMap: A probabilistic, flexible, and compact 3D map representation for robotic systems. In IEEE international conference on robotics and automation.
9.
Zurück zum Zitat Curless, B., & Levoy, M. (1996). A volumetric method for building complex models from range images. In In Proceedings of SIGGRAPH. ACM (pp. 303–312). Curless, B., & Levoy, M. (1996). A volumetric method for building complex models from range images. In In Proceedings of SIGGRAPH. ACM (pp. 303–312).
10.
Zurück zum Zitat Newcombe, R. A., Izadi, S., & Hilliges, O., et al. (1996). KinectFusion: Real-time dense surface mapping and tracking. In Conference on computer graphics and interactive techniques (Vol. 3, pp. 303–312). Newcombe, R. A., Izadi, S., & Hilliges, O., et al. (1996). KinectFusion: Real-time dense surface mapping and tracking. In Conference on computer graphics and interactive techniques (Vol. 3, pp. 303–312).
11.
Zurück zum Zitat Steinbruecker, F., Sturm, J., & Cremers, D. (2014). Volumetric 3D mapping in real-time on a CPU. In IEEE international conference on robotics and automation (pp. 2021–2028). Steinbruecker, F., Sturm, J., & Cremers, D. (2014). Volumetric 3D mapping in real-time on a CPU. In IEEE international conference on robotics and automation (pp. 2021–2028).
12.
Zurück zum Zitat Zollhöfer, M., Thies, J., Colaianni, M., et al. (2014). Interactive model-based reconstruction of the human head using an RGB-D sensor. Computer Animation and Virtual Worlds, 25(25), 213–222.CrossRef Zollhöfer, M., Thies, J., Colaianni, M., et al. (2014). Interactive model-based reconstruction of the human head using an RGB-D sensor. Computer Animation and Virtual Worlds, 25(25), 213–222.CrossRef
13.
Zurück zum Zitat Zhou, Q. Y., & Koltun, V. (2014). Color map optimization for 3d reconstruction with consumer depth cameras. ACM Transactions on Graphics, 33(4), 1–10. Zhou, Q. Y., & Koltun, V. (2014). Color map optimization for 3d reconstruction with consumer depth cameras. ACM Transactions on Graphics, 33(4), 1–10.
14.
Zurück zum Zitat Choi, S., Zhou, Q. Y., & Koltun, V. (2015). Robust reconstruction of indoor scenes. In: Computer vision and pattern recognition (pp. 5556–5565). Choi, S., Zhou, Q. Y., & Koltun, V. (2015). Robust reconstruction of indoor scenes. In: Computer vision and pattern recognition (pp. 5556–5565).
15.
Zurück zum Zitat Wikowski, A., Kornuta, T., Stefańczyk, M., et al. (2016). Efficient generation of 3D surfel maps using RGB-D sensors. International Journal of Applied Mathematics and Computer Science, 1, 99–122.MathSciNetMATH Wikowski, A., Kornuta, T., Stefańczyk, M., et al. (2016). Efficient generation of 3D surfel maps using RGB-D sensors. International Journal of Applied Mathematics and Computer Science, 1, 99–122.MathSciNetMATH
16.
Zurück zum Zitat Kornuta, T., & Laszkowski, M. (2016). Perception subsystem for object recognition and pose estimation in RGB-D images. Automation, Springer International Publishing, 44(10), 995–1003. Kornuta, T., & Laszkowski, M. (2016). Perception subsystem for object recognition and pose estimation in RGB-D images. Automation, Springer International Publishing, 44(10), 995–1003.
17.
Zurück zum Zitat Whelan, T., Johannsson, H., & Kaess, M., et al. (2013). Robust real-time visual odometry for dense RGB-D mapping. In IEEE international conference on robotics and automation. Whelan, T., Johannsson, H., & Kaess, M., et al. (2013). Robust real-time visual odometry for dense RGB-D mapping. In IEEE international conference on robotics and automation.
18.
19.
Zurück zum Zitat Qu, Y., Liu, Z., Jiang, Y., et al. (2017). Self-adaptative variable-metric feature point extraction method. Editorial Office of Optics and Precision Engineering, 25(1), 188–197. (In Chinese).CrossRef Qu, Y., Liu, Z., Jiang, Y., et al. (2017). Self-adaptative variable-metric feature point extraction method. Editorial Office of Optics and Precision Engineering, 25(1), 188–197. (In Chinese).CrossRef
20.
Zurück zum Zitat Liu, Y., Wang, C., Gao, N., et al. (2017). Point cloud adaptive simplification of feature extraction. Editorial Office of Optics and Precision Engineering, 25(1), 245–254. (In Chinese).CrossRef Liu, Y., Wang, C., Gao, N., et al. (2017). Point cloud adaptive simplification of feature extraction. Editorial Office of Optics and Precision Engineering, 25(1), 245–254. (In Chinese).CrossRef
21.
Zurück zum Zitat Maier, R., Sturm, J., & Cremers, D. (2014). Submap-based bundle adjustment for 3D reconstruction from RGB-D Data. In Pattern Recognition (pp. 54–65). Maier, R., Sturm, J., & Cremers, D. (2014). Submap-based bundle adjustment for 3D reconstruction from RGB-D Data. In Pattern Recognition (pp. 54–65).
22.
Zurück zum Zitat Engel, J., Schöps, T., & Cremers, D. (2014). LSD-SLAM: Large-scale direct monocular SLAM (pp. 834–849). Zurich: Springer. Engel, J., Schöps, T., & Cremers, D. (2014). LSD-SLAM: Large-scale direct monocular SLAM (pp. 834–849). Zurich: Springer.
23.
Zurück zum Zitat Stückler, J., & Behnke, S. (2014). Multi-resolution surfel maps for efficient dense 3D modeling and tracking. Journal of Visual Communication and Image Representation, 25(1), 137–147.CrossRef Stückler, J., & Behnke, S. (2014). Multi-resolution surfel maps for efficient dense 3D modeling and tracking. Journal of Visual Communication and Image Representation, 25(1), 137–147.CrossRef
24.
Zurück zum Zitat Nießner, M., Dai, A., & Fisher, M. (2014). Combining inertial navigation and ICP for real-time 3D Surface Reconstruction. Nießner, M., Dai, A., & Fisher, M. (2014). Combining inertial navigation and ICP for real-time 3D Surface Reconstruction.
25.
Zurück zum Zitat Kerl, C., Sturm, J., & Cremers, D. (2013). Dense visual SLAM for RGB-D cameras. In IEEE international conference on intelligent robots and systems (pp. 2100–2106). Kerl, C., Sturm, J., & Cremers, D. (2013). Dense visual SLAM for RGB-D cameras. In IEEE international conference on intelligent robots and systems (pp. 2100–2106).
26.
Zurück zum Zitat Zhang, L., Wang, Y., Sun, H., et al. (2016). Adaptive scale object tracking with kernelized correlation filters. Editorial Office of Optics and Precision Engineering, 24(2), 448–459. (In Chinese).CrossRef Zhang, L., Wang, Y., Sun, H., et al. (2016). Adaptive scale object tracking with kernelized correlation filters. Editorial Office of Optics and Precision Engineering, 24(2), 448–459. (In Chinese).CrossRef
27.
Zurück zum Zitat Wang, Y., Zhang, Q., & Zhou, Y. (2015). Dense 3D mapping for indoor environment based on Kinect-style depth cameras (Vol. 345, pp. 317–330). Cham: Springer. Wang, Y., Zhang, Q., & Zhou, Y. (2015). Dense 3D mapping for indoor environment based on Kinect-style depth cameras (Vol. 345, pp. 317–330). Cham: Springer.
28.
Zurück zum Zitat Sturm, J., Engelhard, N., Endres, F., et al. (2012). A benchmark for the evaluation of RGB-D SLAM systems. In IEEE/RSJ international conference on intelligent robots and systems (pp. 573–580). IEEE. Sturm, J., Engelhard, N., Endres, F., et al. (2012). A benchmark for the evaluation of RGB-D SLAM systems. In IEEE/RSJ international conference on intelligent robots and systems (pp. 573–580). IEEE.
Metadaten
Titel
A Hierarchical Optimization Algorithm Based on GPU for Real-Time 3D Reconstruction
verfasst von
Jin-hua Lin
Lu Wang
Yan-jie Wang
Publikationsdatum
01.06.2017
Verlag
3D Display Research Center
Erschienen in
3D Research / Ausgabe 2/2017
Elektronische ISSN: 2092-6731
DOI
https://doi.org/10.1007/s13319-017-0127-x

Weitere Artikel der Ausgabe 2/2017

3D Research 2/2017 Zur Ausgabe