Skip to main content
Top
Published in: Machine Vision and Applications 1/2021

01-02-2021 | Original Paper

Deblur and deep depth from single defocus image

Authors: Saeed Anwar, Zeeshan Hayder, Fatih Porikli

Published in: Machine Vision and Applications | Issue 1/2021

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

In this paper, we tackle depth estimation and blur removal from a single out-of-focus image. Previously, depth is estimated, and blurred is removed using multiple images; for example, from multiview or stereo scenes, but doing so with a single image is challenging. Earlier works of monocular images for depth estimated and deblurring either exploited geometric characteristics or priors using hand-crafted features. Lately, there is enough evidence that deep convolutional neural networks (CNN) significantly improved numerous vision applications; hence, in this article, we present a depth estimation method that leverages rich representations learned from cascaded convolutional and fully connected neural networks operating on a patch-pooled set of feature maps. Furthermore, from this depth, we computationally reconstruct an all-focus image, i.e., removing the blur and achieve synthetic re-focusing, all from a single image. Our method is fast, and it substantially improves depth accuracy over the state-of-the-art alternatives. Our proposed depth estimation approach can be utilized for everyday scenes without any geometric priors or extra information. Furthermore, our experiments on two benchmark datasets consist images of indoor and outdoor scenes, i.e., Make3D and NYU-v2 demonstrate superior performance in comparison with other available depth estimation state-of-the-art methods by reducing the root-mean-squared error by 57% and 46%, and state-of-the-art blur removal methods by 0.36 dB and 0.72 dB in PSNR, respectively. This improvement in-depth estimation and deblurring is further demonstrated by the superior performance using real defocus images against images captured with a prototype lens.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Footnotes
1
Taken from literature for fair comparison.
 
Literature
1.
go back to reference Zhu, X., Ramanan, D.: Face detection, pose estimation, and landmark localization in the wild. in CVPR (2012) Zhu, X., Ramanan, D.: Face detection, pose estimation, and landmark localization in the wild. in CVPR (2012)
2.
go back to reference Eigen, D., Fergus, R.: Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. in ICCV (2015) Eigen, D., Fergus, R.: Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. in ICCV (2015)
3.
go back to reference Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. in CVPR (2001) Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. in CVPR (2001)
4.
go back to reference He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. in CVPR (2016) He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. in CVPR (2016)
5.
go back to reference Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. in CVPR (2015) Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. in CVPR (2015)
6.
go back to reference Fergus, R., Singh, B., Hertzmann, A., Roweis, S. T., Freeman, W. T.: Removing camera shake from a single photograph (2006) Fergus, R., Singh, B., Hertzmann, A., Roweis, S. T., Freeman, W. T.: Removing camera shake from a single photograph (2006)
7.
go back to reference Levin, A.: Blind motion deblurring using image statistics. in NIPS (2006) Levin, A.: Blind motion deblurring using image statistics. in NIPS (2006)
8.
go back to reference Shan, Q., Jia, J., Agarwala, A.: High-quality motion deblurring from a single image. ACM Trans. Graph (2008) Shan, Q., Jia, J., Agarwala, A.: High-quality motion deblurring from a single image. ACM Trans. Graph (2008)
9.
go back to reference Cho, S., Lee, S.: Fast motion deblurring. ACM Trans. Gr. (TOG) (2009) Cho, S., Lee, S.: Fast motion deblurring. ACM Trans. Gr. (TOG) (2009)
10.
go back to reference Levin, A., Weiss, Y., Durand, F., Freeman, W. T.: Understanding blind deconvolution algorithms. TPAMI (2011) Levin, A., Weiss, Y., Durand, F., Freeman, W. T.: Understanding blind deconvolution algorithms. TPAMI (2011)
11.
go back to reference Nayar, S. K., Ben-Ezra, M.: Motion-based motion deblurring. TPAMI (2004) Nayar, S. K., Ben-Ezra, M.: Motion-based motion deblurring. TPAMI (2004)
12.
go back to reference Li, F., Yu, J., Chai, J.: A hybrid camera for motion deblurring and depth map super-resolution. in CVPR (2008) Li, F., Yu, J., Chai, J.: A hybrid camera for motion deblurring and depth map super-resolution. in CVPR (2008)
13.
go back to reference Tai, Y.-W., Du, H., Brown, M. S., Lin, S.: Image/video deblurring using a hybrid camera. in CVPR (2008) Tai, Y.-W., Du, H., Brown, M. S., Lin, S.: Image/video deblurring using a hybrid camera. in CVPR (2008)
14.
go back to reference Yuan, L., Sun, J., Quan, L., Shum, H.-Y.: Image deblurring with blurred/noisy image pairs. Ser. SIGGRAPH (2007) Yuan, L., Sun, J., Quan, L., Shum, H.-Y.: Image deblurring with blurred/noisy image pairs. Ser. SIGGRAPH (2007)
15.
go back to reference Nathan Silberman, P. K., Hoiem, D., Fergus, R.: Indoor segmentation and support inference from RGBD images. in ECCV (2012) Nathan Silberman, P. K., Hoiem, D., Fergus, R.: Indoor segmentation and support inference from RGBD images. in ECCV (2012)
16.
go back to reference Anwar, S., Hayder, Z., Porikli, F.: Depth estimation and blur removal from a single out-of-focus image. BMVC 1, 2 (2017) Anwar, S., Hayder, Z., Porikli, F.: Depth estimation and blur removal from a single out-of-focus image. BMVC 1, 2 (2017)
17.
go back to reference Levin, A., Fergus, R., Durand, F., Freeman, W.T.: Image and depth from a conventional camera with a coded aperture. ACM Trans. Gr. (2007) Levin, A., Fergus, R., Durand, F., Freeman, W.T.: Image and depth from a conventional camera with a coded aperture. ACM Trans. Gr. (2007)
18.
go back to reference Veeraraghavan, A., Raskar, R., Agrawal, A., Mohan, A., Tumblin, J.: Dappled photography: Mask enhanced cameras for heterodyned light fields and coded aperture refocusing. ACM Trans. Gr. (2007) Veeraraghavan, A., Raskar, R., Agrawal, A., Mohan, A., Tumblin, J.: Dappled photography: Mask enhanced cameras for heterodyned light fields and coded aperture refocusing. ACM Trans. Gr. (2007)
19.
go back to reference Moreno-Noguer, F., Belhumeur, P. N., Nayar, S. K.: Active refocusing of images and videos. ACM Trans. Gr. (2007) Moreno-Noguer, F., Belhumeur, P. N., Nayar, S. K.: Active refocusing of images and videos. ACM Trans. Gr. (2007)
20.
go back to reference Zhou, C., Cossairt, O., Nayar, S.: Depth from diffusion. in CVPR (2010) Zhou, C., Cossairt, O., Nayar, S.: Depth from diffusion. in CVPR (2010)
21.
go back to reference Zhou, C., Nayar, S.: What are good apertures for defocus deblurring? in ICCP (2009) Zhou, C., Nayar, S.: What are good apertures for defocus deblurring? in ICCP (2009)
22.
go back to reference Zhou, C., Lin, S., Nayar, S. K.: Coded aperture pairs for depth from defocus and defocus deblurring. IJCV (2011) Zhou, C., Lin, S., Nayar, S. K.: Coded aperture pairs for depth from defocus and defocus deblurring. IJCV (2011)
23.
go back to reference Levin, A.: Analyzing depth from coded aperture sets. in ECCV (2010) Levin, A.: Analyzing depth from coded aperture sets. in ECCV (2010)
24.
go back to reference Pertuz, S., Puig, D., Garcia, M. A.: Analysis of focus measure operators for shape-from-focus. PR (2013) Pertuz, S., Puig, D., Garcia, M. A.: Analysis of focus measure operators for shape-from-focus. PR (2013)
25.
go back to reference Mahmood, M., Choi, T. S.: Nonlinear approach for enhancement of image focus volume in shape from focus. TIP (2012) Mahmood, M., Choi, T. S.: Nonlinear approach for enhancement of image focus volume in shape from focus. TIP (2012)
26.
go back to reference Shim, S. O., Choi, T. S.: A fast and robust depth estimation method for 3D cameras. in ICCE (2012) Shim, S. O., Choi, T. S.: A fast and robust depth estimation method for 3D cameras. in ICCE (2012)
27.
go back to reference Subbarao, M., Choi, T.: Accurate recovery of three-dimensional shape from image focus. TPAMI (1995) Subbarao, M., Choi, T.: Accurate recovery of three-dimensional shape from image focus. TPAMI (1995)
28.
go back to reference Bae, S., Durand, F.: Defocus magnification. CG Forum (2007) Bae, S., Durand, F.: Defocus magnification. CG Forum (2007)
29.
go back to reference Calderero, F., Caselles, V.: Recovering relative depth from low-level features without explicit t-junction detection and interpretation. IJCV (2013) Calderero, F., Caselles, V.: Recovering relative depth from low-level features without explicit t-junction detection and interpretation. IJCV (2013)
30.
go back to reference Cao, Y., Fang, S., Wang, F.: Single image multi-focusing based on local blur estimation. in ICIG (2011) Cao, Y., Fang, S., Wang, F.: Single image multi-focusing based on local blur estimation. in ICIG (2011)
31.
go back to reference Zhuo, S., Sim, T.: Defocus map estimation from a single image. PR (2011) Zhuo, S., Sim, T.: Defocus map estimation from a single image. PR (2011)
32.
go back to reference Namboodiri, V. P., Chaudhuri, S.: Recovery of relative depth from a single observation using an uncalibrated (real-aperture) camera. in CVPR (2008) Namboodiri, V. P., Chaudhuri, S.: Recovery of relative depth from a single observation using an uncalibrated (real-aperture) camera. in CVPR (2008)
33.
go back to reference Liu, M., Salzmann, M., He, X.: Discrete-continuous depth estimation from a single image. in CVPR (2014) Liu, M., Salzmann, M., He, X.: Discrete-continuous depth estimation from a single image. in CVPR (2014)
34.
go back to reference Liu, F., Shen, C., Lin, G.: Deep convolutional neural fields for depth estimation from a single image. in CVPR (2015) Liu, F., Shen, C., Lin, G.: Deep convolutional neural fields for depth estimation from a single image. in CVPR (2015)
35.
go back to reference Watanabe, M., Nayar, S. K.: Rational filters for passive depth from defocus. IJCV (1998) Watanabe, M., Nayar, S. K.: Rational filters for passive depth from defocus. IJCV (1998)
36.
go back to reference Paramanand, C., Rajagopalan, A. N.: Non-uniform motion deblurring for bilayer scenes. in CVPR (2013) Paramanand, C., Rajagopalan, A. N.: Non-uniform motion deblurring for bilayer scenes. in CVPR (2013)
37.
go back to reference Xu, L., Jia, J.: Depth-aware motion deblurring. in ICCP (2012) Xu, L., Jia, J.: Depth-aware motion deblurring. in ICCP (2012)
38.
go back to reference Li, C., Su, S., Matsushita, Y., Zhou, K., Lin, S.: Bayesian depth-from-defocus with shading constraints. in CVPR (2013) Li, C., Su, S., Matsushita, Y., Zhou, K., Lin, S.: Bayesian depth-from-defocus with shading constraints. in CVPR (2013)
39.
go back to reference Farid, M.S., Mahmood, A., Al-Maadeed, S.A.: Multi-focus image fusion using content adaptive blurring. Inf Fus 45, 96–112 (2019)CrossRef Farid, M.S., Mahmood, A., Al-Maadeed, S.A.: Multi-focus image fusion using content adaptive blurring. Inf Fus 45, 96–112 (2019)CrossRef
40.
go back to reference Krizhevsky, A., Sutskever, I., Hinton, G. E.: Imagenet classification with deep convolutional neural networks. in NIPS (2012) Krizhevsky, A., Sutskever, I., Hinton, G. E.: Imagenet classification with deep convolutional neural networks. in NIPS (2012)
41.
go back to reference Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014) Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:​1409.​1556 (2014)
42.
go back to reference Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. in CVPR (2014) Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. in CVPR (2014)
43.
44.
go back to reference Razavian, A., Azizpour, H., Sullivan, J., Carlsson, S.: CNN features off-the-shelf: an astounding baseline for recognition. in CVPR Workshops (2014) Razavian, A., Azizpour, H., Sullivan, J., Carlsson, S.: CNN features off-the-shelf: an astounding baseline for recognition. in CVPR Workshops (2014)
45.
go back to reference Su, H., Huang, Q., Mitra, N. J., Li, Y., Guibas, L.: Estimating image depth using shape collections. TG (2014) Su, H., Huang, Q., Mitra, N. J., Li, Y., Guibas, L.: Estimating image depth using shape collections. TG (2014)
46.
go back to reference Kar, A., Tulsiani, S., Carreira, J., Malik, J.: Category-specific object reconstruction from a single image. in CVPR (2015) Kar, A., Tulsiani, S., Carreira, J., Malik, J.: Category-specific object reconstruction from a single image. in CVPR (2015)
47.
go back to reference Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. in NIPS (2014) Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. in NIPS (2014)
48.
go back to reference Li, J., Guo, X., Lu, G., Zhang, B., Xu, Y., Wu, F., Zhang, D.: Drpl: Deep regression pair learning for multi-focus image fusion. IEEE Trans Image Process 29, 4816–4831 (2020)CrossRef Li, J., Guo, X., Lu, G., Zhang, B., Xu, Y., Wu, F., Zhang, D.: Drpl: Deep regression pair learning for multi-focus image fusion. IEEE Trans Image Process 29, 4816–4831 (2020)CrossRef
49.
go back to reference Krishnan, D., Tay, T., Fergus, R.: Blind deconvolution using a normalized sparsity measure. in CVPR (2011) Krishnan, D., Tay, T., Fergus, R.: Blind deconvolution using a normalized sparsity measure. in CVPR (2011)
50.
go back to reference Xu, L., Jia, J.: Two-phase kernel estimation for robust motion deblurring. in ECCV (2010) Xu, L., Jia, J.: Two-phase kernel estimation for robust motion deblurring. in ECCV (2010)
51.
go back to reference Whyte, O., Sivic, J., Zisserman, A., Ponce, J.: Non-uniform deblurring for shaken images. IJCV (2012) Whyte, O., Sivic, J., Zisserman, A., Ponce, J.: Non-uniform deblurring for shaken images. IJCV (2012)
52.
go back to reference Joshi, N., Szeliski, R., Kriegman, D.: PSF estimation using sharp edge prediction. in CVPR (2008) Joshi, N., Szeliski, R., Kriegman, D.: PSF estimation using sharp edge prediction. in CVPR (2008)
53.
go back to reference Cho, T. S., Paris, S., Horn, B. K., Freeman, W. T.: Blur kernel estimation using the radon transform. in CVPR (2011) Cho, T. S., Paris, S., Horn, B. K., Freeman, W. T.: Blur kernel estimation using the radon transform. in CVPR (2011)
54.
go back to reference Krishnan, D., Fergus, R.: Fast image deconvolution using hyper-Laplacian priors. in NIPS (2009) Krishnan, D., Fergus, R.: Fast image deconvolution using hyper-Laplacian priors. in NIPS (2009)
55.
go back to reference Whyte, O., Sivic, J., and Zisserman A.: Deblurring shaken and partially saturated images. IJCV (2014) Whyte, O., Sivic, J., and Zisserman A.: Deblurring shaken and partially saturated images. IJCV (2014)
56.
go back to reference Pan, J., Hu, Z., Su, Z., Yang, M. H.: Deblurring text images via L0 regularized intensity and gradient prior. in CVPR (2014) Pan, J., Hu, Z., Su, Z., Yang, M. H.: Deblurring text images via L0 regularized intensity and gradient prior. in CVPR (2014)
57.
go back to reference Zoran, D., Weiss, Y.: From learning models of natural image patches to whole image restoration. in ICCV (2011) Zoran, D., Weiss, Y.: From learning models of natural image patches to whole image restoration. in ICCV (2011)
58.
go back to reference Sun, L., Cho, S., Wang, J., Hays, J.: Edge-based blur Kernel estimation using patch priors. in ICCP (2013) Sun, L., Cho, S., Wang, J., Hays, J.: Edge-based blur Kernel estimation using patch priors. in ICCP (2013)
59.
go back to reference Michaeli, T., Irani, M.: Blind deblurring using internal patch recurrence. in ECCV (2014) Michaeli, T., Irani, M.: Blind deblurring using internal patch recurrence. in ECCV (2014)
60.
go back to reference Schuler, C. J., Hirsch, M., Harmeling, S., Schölkopf, B.: Learning to deblur. TPAMI (2016) Schuler, C. J., Hirsch, M., Harmeling, S., Schölkopf, B.: Learning to deblur. TPAMI (2016)
61.
go back to reference Chakrabarti, A.: A neural approach to blind motion deblurring. in ECCV (2016) Chakrabarti, A.: A neural approach to blind motion deblurring. in ECCV (2016)
62.
go back to reference Anwar, S., Phuoc Huynh, C., Porikli, F.: Class-specific image deblurring. in ICCV (2015) Anwar, S., Phuoc Huynh, C., Porikli, F.: Class-specific image deblurring. in ICCV (2015)
63.
go back to reference Anwar, S., Huynh, C. P., Porikli, F.: Image deblurring with a class-specific prior. TPAMI (2017) Anwar, S., Huynh, C. P., Porikli, F.: Image deblurring with a class-specific prior. TPAMI (2017)
64.
go back to reference Joshi, N., Matusik, W., Adelson, E.H., Kriegman, D.J.: Personal photo enhancement using example images. ACM Trans. Gr. (2010) Joshi, N., Matusik, W., Adelson, E.H., Kriegman, D.J.: Personal photo enhancement using example images. ACM Trans. Gr. (2010)
65.
go back to reference Hacohen, Y., Shechtman, E., Lischinski, D.: Deblurring by example using dense correspondence. in ICCV. (2013) Hacohen, Y., Shechtman, E., Lischinski, D.: Deblurring by example using dense correspondence. in ICCV. (2013)
66.
go back to reference Sun, L., Cho, S., Wang, J., Hays, J.: Good image priors for non-blind deconvolution—generic versus specific. in ECCV (2014) Sun, L., Cho, S., Wang, J., Hays, J.: Good image priors for non-blind deconvolution—generic versus specific. in ECCV (2014)
67.
go back to reference Pan, J., Hu, Z., Su, Z., Yang, M.: Deblurring face images with exemplars. in ECCV (2014) Pan, J., Hu, Z., Su, Z., Yang, M.: Deblurring face images with exemplars. in ECCV (2014)
68.
go back to reference Saxena, A., Sun, M., Ng, A. Y.: Make3d: Learning 3d scene structure from a single still image. TPAMI (2009) Saxena, A., Sun, M., Ng, A. Y.: Make3d: Learning 3d scene structure from a single still image. TPAMI (2009)
69.
go back to reference He, K., Zhang, X., Ren, S., Sun, J.: Spatial pyramid pooling in deep convolutional networks for visual recognition. in ECCV (2014) He, K., Zhang, X., Ren, S., Sun, J.: Spatial pyramid pooling in deep convolutional networks for visual recognition. in ECCV (2014)
70.
go back to reference Levin, A., Zomet, A., Weiss, Y.: Learning how to inpaint from global image statistics. in ICCV (2003) Levin, A., Zomet, A., Weiss, Y.: Learning how to inpaint from global image statistics. in ICCV (2003)
71.
go back to reference Karsch, K., Liu, C., Kang, S. B.: Depth transfer: depth extraction from video using non-parametric sampling. TPAMI. (2014) Karsch, K., Liu, C., Kang, S. B.: Depth transfer: depth extraction from video using non-parametric sampling. TPAMI. (2014)
72.
go back to reference Chakrabarti, A., Zickler, T.: Depth and deblurring from a spectrally-varying depth-of-field. in ECCV (2012) Chakrabarti, A., Zickler, T.: Depth and deblurring from a spectrally-varying depth-of-field. in ECCV (2012)
73.
go back to reference Levin, A., Weiss, Y., Durand, F., Freeman, W. T.: Efficient marginal likelihood optimization in blind deconvolution. in CVPR (2011) Levin, A., Weiss, Y., Durand, F., Freeman, W. T.: Efficient marginal likelihood optimization in blind deconvolution. in CVPR (2011)
74.
go back to reference Cho, S., Lee, S.: Fast motion deblurring. Ser SIGGRAPH Asia (2009) Cho, S., Lee, S.: Fast motion deblurring. Ser SIGGRAPH Asia (2009)
Metadata
Title
Deblur and deep depth from single defocus image
Authors
Saeed Anwar
Zeeshan Hayder
Fatih Porikli
Publication date
01-02-2021
Publisher
Springer Berlin Heidelberg
Published in
Machine Vision and Applications / Issue 1/2021
Print ISSN: 0932-8092
Electronic ISSN: 1432-1769
DOI
https://doi.org/10.1007/s00138-020-01162-6

Other articles of this Issue 1/2021

Machine Vision and Applications 1/2021 Go to the issue

Premium Partner