Skip to main content
Erschienen in: Neural Computing and Applications 19/2022

04.08.2022 | Review

Monocular depth map estimation based on a multi-scale deep architecture and curvilinear saliency feature boosting

verfasst von: Saddam Abdulwahab, Hatem A. Rashwan, Miguel Angel Garcia, Armin Masoumian, Domenec Puig

Erschienen in: Neural Computing and Applications | Ausgabe 19/2022

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Estimating depth from a monocular camera is a must for many applications, including scene understanding and reconstruction, robot vision, and self-driving cars. However, generating depth maps from single RGB images is still a challenge as object shapes are to be inferred from intensity images strongly affected by viewpoint changes, texture content and light conditions. Therefore, most current solutions produce blurry approximations of low-resolution depth maps. We propose a novel depth map estimation technique based on an autoencoder network. This network is endowed with a multi-scale architecture and a multi-level depth estimator that preserve high-level information extracted from coarse feature maps as well as detailed local information present in fine feature maps. Curvilinear saliency, which is related to curvature estimation, is exploited as a loss function to boost the depth accuracy at object boundaries and raise the performance of the estimated high-resolution depth maps. We evaluate our model on the public NYU Depth v2 and Make3D datasets. The proposed model yields superior performance on both datasets compared to the state-of-the-art, achieving an accuracy of \(~86\%\) and showing exceptional performance at the preservation of object boundaries and small 3D structures. The code of the proposed model is publicly available at https://​github.​com/​SaddamAbdulrhman​/​MDACSFB.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Andhare P, Rawat S (2016 Aug) Pick and place industrial robot controller with computer vision. In: 2016 International Conference on Computing Communication Control and automation (ICCUBEA) vol 12, pp. 1-4 Andhare P, Rawat S (2016 Aug) Pick and place industrial robot controller with computer vision. In: 2016 International Conference on Computing Communication Control and automation (ICCUBEA) vol 12, pp. 1-4
2.
Zurück zum Zitat Agarwal N, Chiang CW, Sharma A (2018) A study on computer vision techniques for self-driving cars. InInternational Conference on Frontier Computing, Springer, Singapore, vol 3, pp. 629-634 Agarwal N, Chiang CW, Sharma A (2018) A study on computer vision techniques for self-driving cars. InInternational Conference on Frontier Computing, Springer, Singapore, vol 3, pp. 629-634
3.
Zurück zum Zitat Kanbara M, Okuma T, Takemura H, Yokoya N (2000) A stereoscopic video see-through augmented reality system based on real-time vision-based registration. In: Proceedings IEEE Virtual Reality 2000 (Cat. No. 00CB37048), vol 18, pp. 255–262 Kanbara M, Okuma T, Takemura H, Yokoya N (2000) A stereoscopic video see-through augmented reality system based on real-time vision-based registration. In: Proceedings IEEE Virtual Reality 2000 (Cat. No. 00CB37048), vol 18, pp. 255–262
4.
Zurück zum Zitat Ding Y et al (2020) Digging into the multi-scale structure for a more refined depth map and 3D reconstruction. Neural Comput Appl 32(15):11217–11228CrossRef Ding Y et al (2020) Digging into the multi-scale structure for a more refined depth map and 3D reconstruction. Neural Comput Appl 32(15):11217–11228CrossRef
5.
Zurück zum Zitat Trelinski J, Kwolek B (2021) CNN-based and DTW features for human activity recognition on depth maps. Neural Comput Appl 33(21):14551–14563CrossRef Trelinski J, Kwolek B (2021) CNN-based and DTW features for human activity recognition on depth maps. Neural Comput Appl 33(21):14551–14563CrossRef
6.
Zurück zum Zitat Saxena A, Chung S, Andrew N (2005) Learning depth from single monocular images. Adv Neural Inf Process Syst 18 Saxena A, Chung S, Andrew N (2005) Learning depth from single monocular images. Adv Neural Inf Process Syst 18
7.
Zurück zum Zitat Saxena A, Schulte J, Andrew NY (2007) Depth estimation using monocular and stereo cues. IJCAI. 7:2197 Saxena A, Schulte J, Andrew NY (2007) Depth estimation using monocular and stereo cues. IJCAI. 7:2197
8.
Zurück zum Zitat Choi Y et al (2018) Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE conference on computer vision and pattern recognition Choi Y et al (2018) Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE conference on computer vision and pattern recognition
9.
Zurück zum Zitat Wang N, Zhang Y, Li Z, Fu Y, Liu W, Jiang YG (2018) Pixel2mesh: generating 3d mesh models from single rgb images. In: Proceedings of the European conference on computer vision (ECCV). pp 52-67 Wang N, Zhang Y, Li Z, Fu Y, Liu W, Jiang YG (2018) Pixel2mesh: generating 3d mesh models from single rgb images. In: Proceedings of the European conference on computer vision (ECCV). pp 52-67
10.
Zurück zum Zitat Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition
11.
Zurück zum Zitat Ronneberger O, Philipp F, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. Springer, Cham Ronneberger O, Philipp F, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. Springer, Cham
12.
Zurück zum Zitat Xu Shuzhen, Zhu Qing, Wang Jin (2020) Generative image completion with image-to-image translation. Neural Comput Appl 32(11):7333–7345CrossRef Xu Shuzhen, Zhu Qing, Wang Jin (2020) Generative image completion with image-to-image translation. Neural Comput Appl 32(11):7333–7345CrossRef
13.
Zurück zum Zitat Sun H et al (2021) Scale-free heterogeneous cycleGAN for defogging from a single image for autonomous driving in fog. Neural Comput Appl pp 1-15 Sun H et al (2021) Scale-free heterogeneous cycleGAN for defogging from a single image for autonomous driving in fog. Neural Comput Appl pp 1-15
14.
Zurück zum Zitat Eigen D, Puhrsch C, Fergus R (2014) Depth map prediction from a single image using a multi-scale deep network. Adv Neural Inf Process Syst 27 Eigen D, Puhrsch C, Fergus R (2014) Depth map prediction from a single image using a multi-scale deep network. Adv Neural Inf Process Syst 27
15.
Zurück zum Zitat Ge L, Liang H, Yuan J, Thalmann D (2017) 3d convolutional neural networks for efficient and robust hand pose estimation from single depth images. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp 1991-2000 Ge L, Liang H, Yuan J, Thalmann D (2017) 3d convolutional neural networks for efficient and robust hand pose estimation from single depth images. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp 1991-2000
16.
Zurück zum Zitat Wiles O, Gkioxari G, Szeliski R, Johnson J (2020) Synsin: End-to-end view synthesis from a single image. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 7467-7477 Wiles O, Gkioxari G, Szeliski R, Johnson J (2020) Synsin: End-to-end view synthesis from a single image. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 7467-7477
17.
Zurück zum Zitat Wu J et al (2022) Fast monocular depth estimation via side prediction aggregation with continuous spatial refinement. In: IEEE Transactions on Multimedia Wu J et al (2022) Fast monocular depth estimation via side prediction aggregation with continuous spatial refinement. In: IEEE Transactions on Multimedia
18.
Zurück zum Zitat Liu J et al (2016) Retrieval compensated group structured sparsity for image super-resolution. IEEE Trans Multimed 19(2):302–316CrossRef Liu J et al (2016) Retrieval compensated group structured sparsity for image super-resolution. IEEE Trans Multimed 19(2):302–316CrossRef
19.
Zurück zum Zitat Jun J et al (2021) Monocular human depth estimation via pose estimation. In: IEEE Access 9: 151444-151457 Jun J et al (2021) Monocular human depth estimation via pose estimation. In: IEEE Access 9: 151444-151457
20.
Zurück zum Zitat Alhashim I, Wonka P (2018) High quality monocular depth estimation via transfer learning Alhashim I, Wonka P (2018) High quality monocular depth estimation via transfer learning
21.
Zurück zum Zitat Lin L, Huang G, Chen Y, Zhang L, He B (2020) Efficient and high-quality monocular depth estimation via gated multi-scale network. IEEE Access 7(8):7709–18CrossRef Lin L, Huang G, Chen Y, Zhang L, He B (2020) Efficient and high-quality monocular depth estimation via gated multi-scale network. IEEE Access 7(8):7709–18CrossRef
22.
Zurück zum Zitat Rashwan HA, Chambon S, Gurdjos P, Morin G, Charvillat V (2019) Using curvilinear features in focus for registering a single image to a 3D object. IEEE Trans Image Process 28(9):4429–43MathSciNetMATHCrossRef Rashwan HA, Chambon S, Gurdjos P, Morin G, Charvillat V (2019) Using curvilinear features in focus for registering a single image to a 3D object. IEEE Trans Image Process 28(9):4429–43MathSciNetMATHCrossRef
23.
Zurück zum Zitat Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition
24.
Zurück zum Zitat Xie S, Girshick R, Dollár P, Tu Z, He K (2017) Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 1492-1500 Xie S, Girshick R, Dollár P, Tu Z, He K (2017) Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 1492-1500
25.
Zurück zum Zitat Pirvu M, Robu V, Licaret V, Costea D, Marcu A, Slusanschi E, Sukthankar R, Leordeanu M (2021) Depth distillation: unsupervised metric depth estimation for UAVs by finding consensus between kinematics, optical flow and deep learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 3215–3223 Pirvu M, Robu V, Licaret V, Costea D, Marcu A, Slusanschi E, Sukthankar R, Leordeanu M (2021) Depth distillation: unsupervised metric depth estimation for UAVs by finding consensus between kinematics, optical flow and deep learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 3215–3223
26.
Zurück zum Zitat Schonberger JL, Frahm JM (2016) Structure-from-motion revisited. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4104–4113 Schonberger JL, Frahm JM (2016) Structure-from-motion revisited. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4104–4113
27.
Zurück zum Zitat Lowe David G (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60(2):91–110CrossRef Lowe David G (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60(2):91–110CrossRef
28.
Zurück zum Zitat Li B et al (2015) Depth and surface normal estimation from monocular images using regression on deep features and hierarchical crfs. In: Proceedings of the IEEE conference on computer vision and pattern recognition Li B et al (2015) Depth and surface normal estimation from monocular images using regression on deep features and hierarchical crfs. In: Proceedings of the IEEE conference on computer vision and pattern recognition
29.
Zurück zum Zitat Achanta R et al (2012) SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans Pattern Anal Mach Intell 34(11):2274–2282CrossRef Achanta R et al (2012) SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans Pattern Anal Mach Intell 34(11):2274–2282CrossRef
30.
Zurück zum Zitat Liu F, Shen C, Lin G (2015) Deep convolutional neural fields for depth estimation from a single image. In: Proceedings of the IEEE conference on computer vision and pattern recognition Liu F, Shen C, Lin G (2015) Deep convolutional neural fields for depth estimation from a single image. In: Proceedings of the IEEE conference on computer vision and pattern recognition
31.
Zurück zum Zitat Ling, Chuanwu, Zhang Xiaogang, Chen Hua (2021) ‘Unsupervised monocular depth estimation using attention and multi-warp reconstruction.‘ IEEE Transactions on Multimedia Ling, Chuanwu, Zhang Xiaogang, Chen Hua (2021) ‘Unsupervised monocular depth estimation using attention and multi-warp reconstruction.‘ IEEE Transactions on Multimedia
32.
Zurück zum Zitat Ji R et al (2019) Semi-supervised adversarial monocular depth estimation. IEEE Trans Pattern Anal Mach Intell 42(10):2410–2422CrossRef Ji R et al (2019) Semi-supervised adversarial monocular depth estimation. IEEE Trans Pattern Anal Mach Intell 42(10):2410–2422CrossRef
33.
Zurück zum Zitat Shen G, Zhang Y, Li J, Wei M, Wang Q, Chen G, Heng PA (2021) Learning regularizer for monocular depth estimation with adversarial guidance. In: Proceedings of the 29th ACM International Conference on Multimedia, vol 17, pp 5222–5230 Shen G, Zhang Y, Li J, Wei M, Wang Q, Chen G, Heng PA (2021) Learning regularizer for monocular depth estimation with adversarial guidance. In: Proceedings of the 29th ACM International Conference on Multimedia, vol 17, pp 5222–5230
34.
Zurück zum Zitat Abdulwahab S et al (2020) Adversarial learning for depth and viewpoint estimation from a single image. IEEE Trans Circuits Syst Video Technol 30(9):2947–2958CrossRef Abdulwahab S et al (2020) Adversarial learning for depth and viewpoint estimation from a single image. IEEE Trans Circuits Syst Video Technol 30(9):2947–2958CrossRef
35.
Zurück zum Zitat Fu H, Gong M, Wang C, Batmanghelich K, Tao D (2018) Deep ordinal regression network for monocular depth estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2002–2011 Fu H, Gong M, Wang C, Batmanghelich K, Tao D (2018) Deep ordinal regression network for monocular depth estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2002–2011
36.
Zurück zum Zitat Hao Z, Li Y, You S, Lu F (2018 Sep) Detail preserving depth estimation from a single image using attention guided networks. In: 2018 International Conference on 3D Vision (3DV), pp 304–313 Hao Z, Li Y, You S, Lu F (2018 Sep) Detail preserving depth estimation from a single image using attention guided networks. In: 2018 International Conference on 3D Vision (3DV), pp 304–313
37.
Zurück zum Zitat Laina I et al (2016) Deeper depth prediction with fully convolutional residual networks. In: 2016 Fourth international conference on 3D vision (3DV). IEEE Laina I et al (2016) Deeper depth prediction with fully convolutional residual networks. In: 2016 Fourth international conference on 3D vision (3DV). IEEE
38.
Zurück zum Zitat Zheng Jin, Peng Lihui (2018) An autoencoder-based image reconstruction for electrical capacitance tomography. IEEE Sens J 18(13):5464–5474CrossRef Zheng Jin, Peng Lihui (2018) An autoencoder-based image reconstruction for electrical capacitance tomography. IEEE Sens J 18(13):5464–5474CrossRef
39.
Zurück zum Zitat Blendowski Max, Bouteldja Nassim, Heinrich Mattias P (2020) Multimodal 3D medical image registration guided by shape encoder-decoder networks. Int J Comput Assist Radiol Surg 15(2):269–276CrossRef Blendowski Max, Bouteldja Nassim, Heinrich Mattias P (2020) Multimodal 3D medical image registration guided by shape encoder-decoder networks. Int J Comput Assist Radiol Surg 15(2):269–276CrossRef
40.
Zurück zum Zitat Abdallah BM et al (2018) Noise-estimation-based anisotropic diffusion approach for retinal blood vessel segmentation. Neural Comput Appl 29(8):159–180CrossRef Abdallah BM et al (2018) Noise-estimation-based anisotropic diffusion approach for retinal blood vessel segmentation. Neural Comput Appl 29(8):159–180CrossRef
41.
Zurück zum Zitat Luo B et al (2020) Decomposition algorithm for depth image of human health posture based on brain health. Neural Comput Appl 32(10):6327–6342CrossRef Luo B et al (2020) Decomposition algorithm for depth image of human health posture based on brain health. Neural Comput Appl 32(10):6327–6342CrossRef
42.
Zurück zum Zitat Garg R, Bg VK, Carneiro G, Reid I (2016) Unsupervised CNN for single view depth estimation: geometry to the rescue. Eur Conf Comput Vis 8:740–756 Garg R, Bg VK, Carneiro G, Reid I (2016) Unsupervised CNN for single view depth estimation: geometry to the rescue. Eur Conf Comput Vis 8:740–756
43.
Zurück zum Zitat Wofk D, Ma F, Yang TJ, Karaman S, Sze V (2019) Fastdepth: fast monocular depth estimation on embedded systems. In: 2019 International Conference on Robotics and Automation (ICRA), vol 20, pp 6101–6108 Wofk D, Ma F, Yang TJ, Karaman S, Sze V (2019) Fastdepth: fast monocular depth estimation on embedded systems. In: 2019 International Conference on Robotics and Automation (ICRA), vol 20, pp 6101–6108
44.
Zurück zum Zitat PUIG Domenec (2019) Mgnet: depth map prediction from a single photograph using a multi-generative network. In: Artificial Intelligence Research and Development: Proceedings of the 22nd International Conference of the Catalan Association for Artificial Intelligence. Vol. 319. IOS Press PUIG Domenec (2019) Mgnet: depth map prediction from a single photograph using a multi-generative network. In: Artificial Intelligence Research and Development: Proceedings of the 22nd International Conference of the Catalan Association for Artificial Intelligence. Vol. 319. IOS Press
45.
Zurück zum Zitat Kostadinov D, Ivanovski Z (2012) Single image depth estimation using local gradient-based features. In: 2012 19th International Conference on Systems, Signals and Image Processing (IWSSIP) vol 11, pp 596–599 Kostadinov D, Ivanovski Z (2012) Single image depth estimation using local gradient-based features. In: 2012 19th International Conference on Systems, Signals and Image Processing (IWSSIP) vol 11, pp 596–599
46.
Zurück zum Zitat Godard C, Mac Aodha O, Brostow GJ (2017) Unsupervised monocular depth estimation with left-right consistency. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 270–279 Godard C, Mac Aodha O, Brostow GJ (2017) Unsupervised monocular depth estimation with left-right consistency. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 270–279
47.
Zurück zum Zitat Rashwan HA, Chambon S, Gurdjos P, Morin G, Charvillat V (2016) Towards multi-scale feature detection repeatable over intensity and depth images. IEEE Int Conf Image Process (ICIP) 25:36–40 Rashwan HA, Chambon S, Gurdjos P, Morin G, Charvillat V (2016) Towards multi-scale feature detection repeatable over intensity and depth images. IEEE Int Conf Image Process (ICIP) 25:36–40
48.
Zurück zum Zitat Rashwan Hatem A et al (2019) Using curvilinear features in focus for registering a single image to a 3D object. IEEE Trans Image Process 28(9):4429–4443MathSciNetMATHCrossRef Rashwan Hatem A et al (2019) Using curvilinear features in focus for registering a single image to a 3D object. IEEE Trans Image Process 28(9):4429–4443MathSciNetMATHCrossRef
49.
Zurück zum Zitat Abdulwahab S, Rashwan HA, Cristiano J, Chambon S, Puig D (2019) Effective 2D/3D registration using curvilinear saliency features and multi-class SVM. VISIGRAPP 5:354–361 Abdulwahab S, Rashwan HA, Cristiano J, Chambon S, Puig D (2019) Effective 2D/3D registration using curvilinear saliency features and multi-class SVM. VISIGRAPP 5:354–361
50.
Zurück zum Zitat Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, vol 20, pp 248–255 Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, vol 20, pp 248–255
51.
Zurück zum Zitat He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770-778 He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770-778
52.
Zurück zum Zitat Lehtinen J, Munkberg J, Hasselgren J, Laine S, Karras T, Aittala M, Aila T (2018) Noise2Noise: Learning Image Restoration without Clean Data. Int Conf Mach Learn 3:2965–2974 Lehtinen J, Munkberg J, Hasselgren J, Laine S, Karras T, Aittala M, Aila T (2018) Noise2Noise: Learning Image Restoration without Clean Data. Int Conf Mach Learn 3:2965–2974
53.
Zurück zum Zitat Maas AL, Hannun AY, Andrew NY (2013) Rectifier nonlinearities improve neural network acoustic models. Proc icml 30:1 Maas AL, Hannun AY, Andrew NY (2013) Rectifier nonlinearities improve neural network acoustic models. Proc icml 30:1
54.
Zurück zum Zitat Silberman N, Hoiem D, Kohli P, Fergus R (2012) Indoor segmentation and support inference from rgbd images. In: European conference on computer vision. Springer, Berlin, vol 7, pp 746–760 Silberman N, Hoiem D, Kohli P, Fergus R (2012) Indoor segmentation and support inference from rgbd images. In: European conference on computer vision. Springer, Berlin, vol 7, pp 746–760
55.
Zurück zum Zitat Saxena A, Sun M, Ng AY (2008) Make3D: depth perception from a single still image. AAAI 3:1571–1576 Saxena A, Sun M, Ng AY (2008) Make3D: depth perception from a single still image. AAAI 3:1571–1576
56.
Zurück zum Zitat Kingma DP, and Jimmy LB ADAM: AMETHOD FOR STOCHASTIC OPTIMIZATION.‘ Kingma DP, and Jimmy LB ADAM: AMETHOD FOR STOCHASTIC OPTIMIZATION.‘
57.
Zurück zum Zitat Paszke A, Gross S, Chintala S, Chanan G (2017) Pytorch: Tensors and dynamic neural networks in python with strong gpu acceleration. PyTorch Tensors Dyn Neural Netw Python Strong GPU Accel 6(3):67 Paszke A, Gross S, Chintala S, Chanan G (2017) Pytorch: Tensors and dynamic neural networks in python with strong gpu acceleration. PyTorch Tensors Dyn Neural Netw Python Strong GPU Accel 6(3):67
58.
Zurück zum Zitat Ramamonjisoa M, Firman M, Watson J, Lepetit V, Turmukhambetov D (2021) Single Image Depth Estimation using Wavelet Decomposition.‘ Ramamonjisoa M, Firman M, Watson J, Lepetit V, Turmukhambetov D (2021) Single Image Depth Estimation using Wavelet Decomposition.‘
59.
Zurück zum Zitat Tang M et al (2021) Encoder-decoder structure with the feature pyramid for depth estimation from a single image. IEEE Access 9:22640–22650CrossRef Tang M et al (2021) Encoder-decoder structure with the feature pyramid for depth estimation from a single image. IEEE Access 9:22640–22650CrossRef
60.
Zurück zum Zitat Karsch K, Liu C, Kang SB (2012) Depth extraction from video using non-parametric sampling. In: European conference on computer vision. Springer, Berlin Karsch K, Liu C, Kang SB (2012) Depth extraction from video using non-parametric sampling. In: European conference on computer vision. Springer, Berlin
61.
Zurück zum Zitat Kuznietsov Y, Stuckler J, Leibe B (2017) Semi-supervised deep learning for monocular depth map prediction. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 6647–6655 Kuznietsov Y, Stuckler J, Leibe B (2017) Semi-supervised deep learning for monocular depth map prediction. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 6647–6655
62.
Zurück zum Zitat Karsch K, Liu C, Kang SB (2014) Depth transfer: depth extraction from video using non-parametric sampling. In: IEEE transactions on pattern analysis and machine intelligence 36 11 : 2144–2158 Karsch K, Liu C, Kang SB (2014) Depth transfer: depth extraction from video using non-parametric sampling. In: IEEE transactions on pattern analysis and machine intelligence 36 11 : 2144–2158
Metadaten
Titel
Monocular depth map estimation based on a multi-scale deep architecture and curvilinear saliency feature boosting
verfasst von
Saddam Abdulwahab
Hatem A. Rashwan
Miguel Angel Garcia
Armin Masoumian
Domenec Puig
Publikationsdatum
04.08.2022
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 19/2022
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-022-07663-x

Weitere Artikel der Ausgabe 19/2022

Neural Computing and Applications 19/2022 Zur Ausgabe

Premium Partner