Skip to main content

2023 | OriginalPaper | Buchkapitel

Cascaded Network-Based Single-View Bird 3D Reconstruction

verfasst von : Pei Su, Qijun Zhao, Fan Pan, Fei Gao

Erschienen in: Artificial Neural Networks and Machine Learning – ICANN 2023

Verlag: Springer Nature Switzerland

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Existing single-view bird 3D reconstruction methods mostly cannot well recover the local geometry such as feet and wing tips, and the resulting 3D models often appear to have poor appearance when viewed from a new perspective. We thus propose a new method that requires only images and their silhouettes to accurately predict the shape of birds, as well as to obtain reasonable appearance in new perspectives. The key to the method lies in the introduction of a cascaded structure in the shape reconstruction network. This allows for the gradual generation of the 3D shape of birds from coarse to fine, enabling better capturing of local geometric features. Meanwhile, we recover the texture, lighting and camera pose with attention-enhanced encoders. To further improve the plausibility of the reconstructed 3D bird in novel views, we introduce the Multi-view Cycle Consistency loss to train the proposed method. We compare our method with state-of-the-art methods and demonstrate its superiority both qualitatively and quantitatively.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Kato, H., Ushiku, Y., Harada, T.: Neural 3d mesh renderer. In: CVPR (2018) Kato, H., Ushiku, Y., Harada, T.: Neural 3d mesh renderer. In: CVPR (2018)
2.
Zurück zum Zitat Chen, W., et al.: Learning to predict 3d objects with an interpolation-based differentiable renderer. In: NeurIPS (2019) Chen, W., et al.: Learning to predict 3d objects with an interpolation-based differentiable renderer. In: NeurIPS (2019)
3.
Zurück zum Zitat Liu, S., Li, T., Chen, W., Li, H.: Soft rasterizer: A differentiable renderer for image-based 3d reasoning. In: ICCV (2019) Liu, S., Li, T., Chen, W., Li, H.: Soft rasterizer: A differentiable renderer for image-based 3d reasoning. In: ICCV (2019)
4.
Zurück zum Zitat Pavllo, D., Spinks, G., Hofmann, T., Moens, M.-F., Lucchi, A.: Convolutional generation of textured 3d meshes. In: NeurIPS (2020) Pavllo, D., Spinks, G., Hofmann, T., Moens, M.-F., Lucchi, A.: Convolutional generation of textured 3d meshes. In: NeurIPS (2020)
5.
Zurück zum Zitat Kanazawa, A., Tulsiani, S., Efros, A.A., Malik, J.: Learning category-specific mesh reconstruction from image collections. In: ECCV (2018) Kanazawa, A., Tulsiani, S., Efros, A.A., Malik, J.: Learning category-specific mesh reconstruction from image collections. In: ECCV (2018)
6.
Zurück zum Zitat Sun, S., Zhu, Z., Dai, X., Zhao, Q., Li, J.: Weakly-supervised reconstruction of 3d objects with large shape variation from single in-the-wild images. In: ACCV (2020) Sun, S., Zhu, Z., Dai, X., Zhao, Q., Li, J.: Weakly-supervised reconstruction of 3d objects with large shape variation from single in-the-wild images. In: ACCV (2020)
7.
Zurück zum Zitat Li, X., et al.: Self-supervised single-view 3d reconstruction via semantic consistency. In: ECCV (2020) Li, X., et al.: Self-supervised single-view 3d reconstruction via semantic consistency. In: ECCV (2020)
8.
Zurück zum Zitat Monnier, T., Fisher, M., Efros, A.A., Aubry, M.: Share with thy neighbors: single-view reconstruction by cross-instance consistency. In: ECCV (2022) Monnier, T., Fisher, M., Efros, A.A., Aubry, M.: Share with thy neighbors: single-view reconstruction by cross-instance consistency. In: ECCV (2022)
9.
Zurück zum Zitat Goel, S., Kanazawa, A., Malik, J.: Shape and viewpoint without keypoints. In: ECCV (2020) Goel, S., Kanazawa, A., Malik, J.: Shape and viewpoint without keypoints. In: ECCV (2020)
10.
Zurück zum Zitat Hung, W.-C., et al.: Scops: self-supervised co-part segmentation. In: CVPR (2019) Hung, W.-C., et al.: Scops: self-supervised co-part segmentation. In: CVPR (2019)
11.
Zurück zum Zitat Kokkinos, F., Kokkinos, I.: To the point: Correspondence-driven monocular 3d category reconstruction. In: NeurIPS (2021) Kokkinos, F., Kokkinos, I.: To the point: Correspondence-driven monocular 3d category reconstruction. In: NeurIPS (2021)
12.
Zurück zum Zitat Hu, T., Wang, L., Xu, X., Liu, S., Jia, J.: Self-supervised 3d mesh reconstruction from single images. In: CVPR (2021) Hu, T., Wang, L., Xu, X., Liu, S., Jia, J.: Self-supervised 3d mesh reconstruction from single images. In: CVPR (2021)
13.
Zurück zum Zitat Zhang, J., et al.: Monocular 3d object reconstruction with GAN inversion. In: ECCV (2022) Zhang, J., et al.: Monocular 3d object reconstruction with GAN inversion. In: ECCV (2022)
14.
Zurück zum Zitat Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: convolutional block attention module. In: ECCV (2018) Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: convolutional block attention module. In: ECCV (2018)
15.
Zurück zum Zitat Ramamoorthi, R., Hanrahan, P.: An efficient representation for irradiance environment maps. In: CGIT (2001) Ramamoorthi, R., Hanrahan, P.: An efficient representation for irradiance environment maps. In: CGIT (2001)
16.
Zurück zum Zitat Wang, T., Ma, C., Su, H., Wang, W.: CSPN: multi-scale cascade spatial pyramid network for object detection. In: ICASSP (2021) Wang, T., Ma, C., Su, H., Wang, W.: CSPN: multi-scale cascade spatial pyramid network for object detection. In: ICASSP (2021)
17.
Zurück zum Zitat Li, X., Zhang, G., Pan, H., Wang, Z.: CPGNET: cascade point-grid fusion network for real-time lidar semantic segmentation. In: ICRA (2022) Li, X., Zhang, G., Pan, H., Wang, Z.: CPGNET: cascade point-grid fusion network for real-time lidar semantic segmentation. In: ICRA (2022)
18.
Zurück zum Zitat Shim, D., Kim, H.J.: Swindepth: unsupervised depth estimation using monocular sequences via swin transformer and densely cascaded network, arXiv (2023) Shim, D., Kim, H.J.: Swindepth: unsupervised depth estimation using monocular sequences via swin transformer and densely cascaded network, arXiv (2023)
19.
Zurück zum Zitat Chen, X., et al.: Salience-guided cascaded suppression network for person re-identification. In: CVPR (2020) Chen, X., et al.: Salience-guided cascaded suppression network for person re-identification. In: CVPR (2020)
20.
Zurück zum Zitat Li, Y., Zhao, Z., Fan, J., Li, W.: ADR-MVSNet: a cascade network for 3d point cloud reconstruction with pixel occlusion. In: PR (2022) Li, Y., Zhao, Z., Fan, J., Li, W.: ADR-MVSNet: a cascade network for 3d point cloud reconstruction with pixel occlusion. In: PR (2022)
21.
Zurück zum Zitat Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The caltech-UCSD birds-200-2011 dataset. Technical Report (2011) Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The caltech-UCSD birds-200-2011 dataset. Technical Report (2011)
22.
Zurück zum Zitat Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: NeurIPS (2017) Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: NeurIPS (2017)
Metadaten
Titel
Cascaded Network-Based Single-View Bird 3D Reconstruction
verfasst von
Pei Su
Qijun Zhao
Fan Pan
Fei Gao
Copyright-Jahr
2023
DOI
https://doi.org/10.1007/978-3-031-44210-0_10

Premium Partner