Skip to main content
Erschienen in: International Journal of Computer Vision 3/2020

21.06.2019

Convolutional Networks with Adaptive Inference Graphs

verfasst von: Andreas Veit, Serge Belongie

Erschienen in: International Journal of Computer Vision | Ausgabe 3/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Do convolutional networks really need a fixed feed-forward structure? What if, after identifying the high-level concept of an image, a network could move directly to a layer that can distinguish fine-grained differences? Currently, a network would first need to execute sometimes hundreds of intermediate layers that specialize in unrelated aspects. Ideally, the more a network already knows about an image, the better it should be at deciding which layer to compute next. In this work, we propose convolutional networks with adaptive inference graphs (ConvNet-AIG) that adaptively define their network topology conditioned on the input image. Following a high-level structure similar to residual networks (ResNets), ConvNet-AIG decides for each input image on the fly which layers are needed. In experiments on ImageNet we show that ConvNet-AIG learns distinct inference graphs for different categories. Both ConvNet-AIG with 50 and 101 layers outperform their ResNet counterpart, while using \(20\%\) and \(38\%\) less computations respectively. By grouping parameters into layers for related classes and only executing relevant layers, ConvNet-AIG improves both efficiency and overall classification quality. Lastly, we also study the effect of adaptive inference graphs on the susceptibility towards adversarial examples. We observe that ConvNet-AIG shows a higher robustness than ResNets, complementing other known defense mechanisms.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Andreas, J., Rohrbach, M., Darrell, T., Klein, D. (2016). Learning to compose neural networks for question answering. In: Proceedings of NAACL-HLT. Andreas, J., Rohrbach, M., Darrell, T., Klein, D. (2016). Learning to compose neural networks for question answering. In: Proceedings of NAACL-HLT.
Zurück zum Zitat Andreas, J., Rohrbach, M., Darrell, T., & Klein, D. (2016). Neural module networks. In: Conference on computer vision and pattern recognition (CVPR). Andreas, J., Rohrbach, M., Darrell, T., & Klein, D. (2016). Neural module networks. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat Bengio, E., Bacon, P. L., Pineau, J., & Precup, D. (2015). Conditional computation in neural networks for faster models. arXiv preprint arXiv:1511.06297. Bengio, E., Bacon, P. L., Pineau, J., & Precup, D. (2015). Conditional computation in neural networks for faster models. arXiv preprint arXiv:​1511.​06297.
Zurück zum Zitat Bengio, Y., Léonard, N., & Courville, A. (2013). Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432. Bengio, Y., Léonard, N., & Courville, A. (2013). Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:​1308.​3432.
Zurück zum Zitat Deng, J., Dong, W., Socher, R., Li, L. J., Li, K., & Fei-Fei, L. (2009). Imagenet: A large-scale hierarchical image database. In: Conference on computer vision and pattern recognition (CVPR). Deng, J., Dong, W., Socher, R., Li, L. J., Li, K., & Fei-Fei, L. (2009). Imagenet: A large-scale hierarchical image database. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat Figurnov, M., Collins, M. D., Zhu, Y., Zhang, L., Huang, J., Vetrov, D., & Salakhutdinov, R. (2017). Spatially adaptive computation time for residual networks. In: Conference on computer vision and pattern recognition (CVPR). Figurnov, M., Collins, M. D., Zhu, Y., Zhang, L., Huang, J., Vetrov, D., & Salakhutdinov, R. (2017). Spatially adaptive computation time for residual networks. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat Glorot, X., Bordes, A., & Bengio, Y. (2011). Deep sparse rectifier neural networks. In: International conference on artificial intelligence and statistics (AISTATS). Glorot, X., Bordes, A., & Bengio, Y. (2011). Deep sparse rectifier neural networks. In: International conference on artificial intelligence and statistics (AISTATS).
Zurück zum Zitat Goodfellow, I. J., Shlens, J., & Szegedy, C. (2014). Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572. Goodfellow, I. J., Shlens, J., & Szegedy, C. (2014). Explaining and harnessing adversarial examples. arXiv preprint arXiv:​1412.​6572.
Zurück zum Zitat Gumbel, E. J. (1954). Statistical theory of extreme values and some practical applications: A series of lectures. 33. US Govt. Print. Office. Gumbel, E. J. (1954). Statistical theory of extreme values and some practical applications: A series of lectures. 33. US Govt. Print. Office.
Zurück zum Zitat Guo, C., Rana, M., Cisse, M., & van der Maaten, L. (2017). Countering adversarial images using input transformations. arXiv preprint arXiv:1711.00117. Guo, C., Rana, M., Cisse, M., & van der Maaten, L. (2017). Countering adversarial images using input transformations. arXiv preprint arXiv:​1711.​00117.
Zurück zum Zitat He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In: Conference on computer vision and pattern recognition (CVPR). He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat He, K., Zhang, X., Ren, S., & Sun, J. (2016). Identity mappings in deep residual networks. In: European conference on computer vision (ECCV). He, K., Zhang, X., Ren, S., & Sun, J. (2016). Identity mappings in deep residual networks. In: European conference on computer vision (ECCV).
Zurück zum Zitat Huang, G., Chen, D., Li, T., Wu, F., van der Maaten, L., & Weinberger, K. Q. (2017). Multi-scale dense convolutional networks for efficient prediction. arXiv preprint arXiv:1703.09844. Huang, G., Chen, D., Li, T., Wu, F., van der Maaten, L., & Weinberger, K. Q. (2017). Multi-scale dense convolutional networks for efficient prediction. arXiv preprint arXiv:​1703.​09844.
Zurück zum Zitat Huang, G., Liu, Z., Weinberger, K. Q., & van der Maaten, L. (2017). Densely connected convolutional networks. In: Conference on computer vision and pattern recognition (CVPR). Huang, G., Liu, Z., Weinberger, K. Q., & van der Maaten, L. (2017). Densely connected convolutional networks. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat Huang, G., Sun, Y., Liu, Z., Sedra, D., & Weinberger, K. Q. (2016). Deep networks with stochastic depth. In: European conference on computer vision (ECCV) Huang, G., Sun, Y., Liu, Z., Sedra, D., & Weinberger, K. Q. (2016). Deep networks with stochastic depth. In: European conference on computer vision (ECCV)
Zurück zum Zitat Huang, X., & Belongie, S. (2017). Arbitrary style transfer in real-time with adaptive instance normalization. In: International conference on computer vision (ICCV). Huang, X., & Belongie, S. (2017). Arbitrary style transfer in real-time with adaptive instance normalization. In: International conference on computer vision (ICCV).
Zurück zum Zitat Ioffe, S., & Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, pp. 448–456. Ioffe, S., & Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, pp. 448–456.
Zurück zum Zitat Johnson, J., Hariharan, B., van der Maaten, L., Hoffman, J., Fei-Fei, L., Zitnick, C. L., et al. (2017). Inferring andexecuting programs for visual reasoning. In: International conference on computer vision (ICCV). Johnson, J., Hariharan, B., van der Maaten, L., Hoffman, J., Fei-Fei, L., Zitnick, C. L., et al. (2017). Inferring andexecuting programs for visual reasoning. In: International conference on computer vision (ICCV).
Zurück zum Zitat Krizhevsky, A., & Hinton, G. (2009). Learning multiple layers of features from tiny images. Krizhevsky, A., & Hinton, G. (2009). Learning multiple layers of features from tiny images.
Zurück zum Zitat Li, H., Lin, Z., Shen, X., Brandt, J., & Hua, G. (2015). A convolutional neural network cascade for face detection. In: Conference on computer vision and pattern recognition (CVPR). Li, H., Lin, Z., Shen, X., Brandt, J., & Hua, G. (2015). A convolutional neural network cascade for face detection. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat Maddison, C. J., Mnih, A., & Teh, Y. W. (2016). The concrete distribution: A continuous relaxation of discrete random variables. arXiv preprint arXiv:1611.00712. Maddison, C. J., Mnih, A., & Teh, Y. W. (2016). The concrete distribution: A continuous relaxation of discrete random variables. arXiv preprint arXiv:​1611.​00712.
Zurück zum Zitat Misra, I., Gupta, A., & Hebert, M. (2017). From red wine to red tomato: Composition with context. In: Conference on computer vision and pattern recognition (CVPR). Misra, I., Gupta, A., & Hebert, M. (2017). From red wine to red tomato: Composition with context. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le, Q., Hinton, G., & Dean, J. (2017). Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. arXiv preprint arXiv:1701.06538. Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le, Q., Hinton, G., & Dean, J. (2017). Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. arXiv preprint arXiv:​1701.​06538.
Zurück zum Zitat Srivastava, N., Hinton, G. E., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R. (2014). Dropout: A simple way to prevent neural networks from overfitting. Journal of machine learning research (JMLR), 15(1), 1929–1958.MathSciNetMATH Srivastava, N., Hinton, G. E., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R. (2014). Dropout: A simple way to prevent neural networks from overfitting. Journal of machine learning research (JMLR), 15(1), 1929–1958.MathSciNetMATH
Zurück zum Zitat Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., & Rabinovich, A. (2015). Going deeper with convolutions. In: Conference on computer vision and pattern recognition (CVPR). Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., & Rabinovich, A. (2015). Going deeper with convolutions. In: Conference on computer vision and pattern recognition (CVPR).
Zurück zum Zitat Teerapittayanon, S., McDanel, B., & Kung, H. (2016). Branchynet: Fast inference via early exiting from deep neural networks. In: Conference on pattern recognition (ICPR). Teerapittayanon, S., McDanel, B., & Kung, H. (2016). Branchynet: Fast inference via early exiting from deep neural networks. In: Conference on pattern recognition (ICPR).
Zurück zum Zitat Veit, A., Wilber, M. J., & Belongie, S. (2016). Residual networks behave like ensembles of relatively shallow networks. In: Advances in neural information processing systems (NIPS). Veit, A., Wilber, M. J., & Belongie, S. (2016). Residual networks behave like ensembles of relatively shallow networks. In: Advances in neural information processing systems (NIPS).
Zurück zum Zitat Viola, P., & Jones, M. J. (2004). Robust real-time face detection. International Journal of Computer Vision (IJCV), 57(2), 137–154.CrossRef Viola, P., & Jones, M. J. (2004). Robust real-time face detection. International Journal of Computer Vision (IJCV), 57(2), 137–154.CrossRef
Zurück zum Zitat Yang, F., Choi, W., & Lin, Y. (2016). Exploit all the layers: Fast and accurate cnn object detector with scale dependent pooling and cascaded rejection classifiers. In: Conference on computer vision and pattern recognition (CVPR). Yang, F., Choi, W., & Lin, Y. (2016). Exploit all the layers: Fast and accurate cnn object detector with scale dependent pooling and cascaded rejection classifiers. In: Conference on computer vision and pattern recognition (CVPR).
Metadaten
Titel
Convolutional Networks with Adaptive Inference Graphs
verfasst von
Andreas Veit
Serge Belongie
Publikationsdatum
21.06.2019
Verlag
Springer US
Erschienen in
International Journal of Computer Vision / Ausgabe 3/2020
Print ISSN: 0920-5691
Elektronische ISSN: 1573-1405
DOI
https://doi.org/10.1007/s11263-019-01190-4

Weitere Artikel der Ausgabe 3/2020

International Journal of Computer Vision 3/2020 Zur Ausgabe

Premium Partner