Skip to main content
Erschienen in: Artificial Intelligence Review 3/2023

22.06.2022

A review of convolutional neural network architectures and their optimizations

verfasst von: Shuang Cong, Yang Zhou

Erschienen in: Artificial Intelligence Review | Ausgabe 3/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

The research advances concerning the typical architectures of convolutional neural networks (CNNs) as well as their optimizations are analyzed and elaborated in detail in this paper. This paper proposes a typical approach to classifying CNNs architecture based on modules in order to accommodate more new network architectures with multiple characteristics that make them difficult to rely on the original classification method. Through the pros and cons analysis of diverse network architectures and their performance comparisons, six types of typical CNNs architectures are analyzed and explained in detail. The CNNs architectures intrinsic characteristics is also explored. Moreover, this paper provides a comprehensive classification of network compression and accelerated network architecture optimization algorithms based on the mathematical principle of various optimization algorithms. Finally, this paper analyses the strategy of NAS algorithms, discusses the applications of CNNs, and sheds light on the challenges and prospects of the current CNNs architecture and its optimizations. The explanation of the advantages brought by optimizing different network architecture types, the basis for constructively choosing appropriate CNNs in specific designs and applications are provided. This paper will help the readers to choose constructively appropriate CNNs in specific designs and applications.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Aghli N, Ribeiro E (2021) Combining weight pruning and knowledge distillation for CN compression. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3191–3198 Aghli N, Ribeiro E (2021) Combining weight pruning and knowledge distillation for CN compression. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3191–3198
Zurück zum Zitat Alzubaidi L, Al-Shamma O, Fadhel MA, Farhan L, Zhang J, Duan Y (2020) Optimizing the performance of breast cancer classification by employing the same domain transfer learning from hybrid deep convolutional neural network model. Electronics 9(3):445CrossRef Alzubaidi L, Al-Shamma O, Fadhel MA, Farhan L, Zhang J, Duan Y (2020) Optimizing the performance of breast cancer classification by employing the same domain transfer learning from hybrid deep convolutional neural network model. Electronics 9(3):445CrossRef
Zurück zum Zitat Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, Santamaría J, Fadhel MA, Al-Amidie M, Farhan L (2021) Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data 8(1):1–74CrossRef Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, Santamaría J, Fadhel MA, Al-Amidie M, Farhan L (2021) Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data 8(1):1–74CrossRef
Zurück zum Zitat Alzubaidi L, Al-Shamma O, Fadhel MA, Farhan L, Zhang J (2018) Classification of red blood cells in sickle cell anemia using deep convolutional neural network. In: International conference on intelligent systems design and applications. Springer, Cham, pp 550–559 Alzubaidi L, Al-Shamma O, Fadhel MA, Farhan L, Zhang J (2018) Classification of red blood cells in sickle cell anemia using deep convolutional neural network. In: International conference on intelligent systems design and applications. Springer, Cham, pp 550–559
Zurück zum Zitat Anwar S, Hwang K, Sung W (2017) Structured pruning of deep convolutional neural networks. ACM J Emerg Technol Comput Syst JETC) 13(3):1–18CrossRef Anwar S, Hwang K, Sung W (2017) Structured pruning of deep convolutional neural networks. ACM J Emerg Technol Comput Syst JETC) 13(3):1–18CrossRef
Zurück zum Zitat Astrid M, Lee S-I (2017) Cp-decomposition with tensor power method for convolutional neural networks compression. In: 2017 IEEE international conference on Big Data and Smart Computing (BigComp). IEEE, pp 115–118 Astrid M, Lee S-I (2017) Cp-decomposition with tensor power method for convolutional neural networks compression. In: 2017 IEEE international conference on Big Data and Smart Computing (BigComp). IEEE, pp 115–118
Zurück zum Zitat Ba LJ, Caruana R (2013) Do deep nets really need to be deep? Adv Neural Inf Process Syst 3:2654–2662 Ba LJ, Caruana R (2013) Do deep nets really need to be deep? Adv Neural Inf Process Syst 3:2654–2662
Zurück zum Zitat Bengio Y (2013) Deep learning of representations: Looking forward. In: International conference on statistical language and speech processing. Springer, Berlin, pp 1–37 Bengio Y (2013) Deep learning of representations: Looking forward. In: International conference on statistical language and speech processing. Springer, Berlin, pp 1–37
Zurück zum Zitat Bochkovskiy A, Wang C-Y, Liao H-YM (2020) Yolov4: optimal speed and accuracy of object detection. arXiv preprint. arXiv:2004.10934 Bochkovskiy A, Wang C-Y, Liao H-YM (2020) Yolov4: optimal speed and accuracy of object detection. arXiv preprint. arXiv:​2004.​10934
Zurück zum Zitat Bucilu C, Caruana R, Niculescu-Mizil A (2006) Model compression. In: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pp 535–541 Bucilu C, Caruana R, Niculescu-Mizil A (2006) Model compression. In: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pp 535–541
Zurück zum Zitat Chen X, Hsieh C-J (2020) Stabilizing differentiable architecture search via perturbation-based regularization. In: International conference on machine learning (PMLR), pp 1554–1565 Chen X, Hsieh C-J (2020) Stabilizing differentiable architecture search via perturbation-based regularization. In: International conference on machine learning (PMLR), pp 1554–1565
Zurück zum Zitat Chen T, Goodfellow I, Shlens J (2015a) Net2net: Accelerating learning via knowledge transfer. Computer Science Chen T, Goodfellow I, Shlens J (2015a) Net2net: Accelerating learning via knowledge transfer. Computer Science
Zurück zum Zitat Chen W, Wilson J, Tyree S, Weinberger K, Chen Y (2015b) Compressing neural networks with the hashing trick. In: International conference on machine learning (PMLR), pp 2285–2294 Chen W, Wilson J, Tyree S, Weinberger K, Chen Y (2015b) Compressing neural networks with the hashing trick. In: International conference on machine learning (PMLR), pp 2285–2294
Zurück zum Zitat Chen X, Li Z, Yuan Y, Yu G, Shen J, Qi D (2020) State-aware tracker for real-time video object segmentation. In: Proceedings of the IEEE/CVF international conference on computer vision and pattern recognition, pp 9384–9393 Chen X, Li Z, Yuan Y, Yu G, Shen J, Qi D (2020) State-aware tracker for real-time video object segmentation. In: Proceedings of the IEEE/CVF international conference on computer vision and pattern recognition, pp 9384–9393
Zurück zum Zitat Chen B, Li P, Li B, Lin C, Li C, Sun M, Yan J, Ouyang W (2021a) BN-NAS: neural architecture search with batch normalization. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 307–316 Chen B, Li P, Li B, Lin C, Li C, Sun M, Yan J, Ouyang W (2021a) BN-NAS: neural architecture search with batch normalization. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 307–316
Zurück zum Zitat Chen M, Peng H, Fu J, Ling H (2021b) Autoformer: searching transformers for visual recognition. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 12270–12280 Chen M, Peng H, Fu J, Ling H (2021b) Autoformer: searching transformers for visual recognition. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 12270–12280
Zurück zum Zitat Chen Y, Dai X, Chen D, Liu M, Dong X, Yuan L, Liu Z (2021c) Mobile-former: bridging mobilenet and transformer. arXiv preprint. arXiv:2108.05895 Chen Y, Dai X, Chen D, Liu M, Dong X, Yuan L, Liu Z (2021c) Mobile-former: bridging mobilenet and transformer. arXiv preprint. arXiv:​2108.​05895
Zurück zum Zitat Cheng A-C, Lin CH, Juan D-C, Wei W, Sun M (2020) InstaNAS: instance-aware neural architecture search. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 3577–3584 Cheng A-C, Lin CH, Juan D-C, Wei W, Sun M (2020) InstaNAS: instance-aware neural architecture search. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 3577–3584
Zurück zum Zitat Chollet F (2017) Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1251–1258 Chollet F (2017) Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1251–1258
Zurück zum Zitat Choudhary T, Mishra V, Goswami A, Sarangapani J (2020) A comprehensive survey on model compression and acceleration. Artif Intell Rev 53(7):5113–5155CrossRef Choudhary T, Mishra V, Goswami A, Sarangapani J (2020) A comprehensive survey on model compression and acceleration. Artif Intell Rev 53(7):5113–5155CrossRef
Zurück zum Zitat Chu X, Zhou T, Zhang B, Li J (2020a) Fair DARTS: eliminating unfair advantages in differentiable architecture search. In: European conference on computer vision. Springer, Munich, pp 465–480 Chu X, Zhou T, Zhang B, Li J (2020a) Fair DARTS: eliminating unfair advantages in differentiable architecture search. In: European conference on computer vision. Springer, Munich, pp 465–480
Zurück zum Zitat Chu X, Zhang B, Xu R (2020b) Multi-objective reinforced evolution in mobile neural architecture search. In: European European conference on computer vision. Springer, Munich, pp 99–113 Chu X, Zhang B, Xu R (2020b) Multi-objective reinforced evolution in mobile neural architecture search. In: European European conference on computer vision. Springer, Munich, pp 99–113
Zurück zum Zitat Clevert D-A, Unterthiner T, Hochreiter S (2015) Fast and accurate deep network learning by exponential linear units (ELUS). arXiv preprint. arXiv:1511.07289 Clevert D-A, Unterthiner T, Hochreiter S (2015) Fast and accurate deep network learning by exponential linear units (ELUS). arXiv preprint. arXiv:​1511.​07289
Zurück zum Zitat Costa-Pazo A, Bhattacharjee S, Vazquez-Fernandez E, Marcel S (2016) The replay-mobile face presentation-attack database. In: International conference of the Biometrics Special Interest Group (BIOSIG). IEEE, pp 1–7 Costa-Pazo A, Bhattacharjee S, Vazquez-Fernandez E, Marcel S (2016) The replay-mobile face presentation-attack database. In: International conference of the Biometrics Special Interest Group (BIOSIG). IEEE, pp 1–7
Zurück zum Zitat Courbariaux M, Bengio Y, David J-P (2015) BinaryConnect: training deep neural networks with binary weights during propagations. In: Advances in neural information processing systems, pp 3123–3131 Courbariaux M, Bengio Y, David J-P (2015) BinaryConnect: training deep neural networks with binary weights during propagations. In: Advances in neural information processing systems, pp 3123–3131
Zurück zum Zitat Courbariaux M, Hubara I, Soudry D, El-Yaniv R, Bengio Y (2016) Binarized neural networks: training deep neural networks with weights and activations constrained to +1 or −1. arXiv preprint. arXiv:1602.02830 Courbariaux M, Hubara I, Soudry D, El-Yaniv R, Bengio Y (2016) Binarized neural networks: training deep neural networks with weights and activations constrained to +1 or −1. arXiv preprint. arXiv:​1602.​02830
Zurück zum Zitat Csáji BC et al (2001) Approximation with artificial neural networks, vol 24(48). MSc thesis, Faculty of Sciences, Eötvös Loránd University, p 7 Csáji BC et al (2001) Approximation with artificial neural networks, vol 24(48). MSc thesis, Faculty of Sciences, Eötvös Loránd University, p 7
Zurück zum Zitat Dai Z, Liu H, Le QV, Tan M (2021) Coatnet: marrying convolution and attention for all data sizes. arXiv preprint. arXiv:2106.04803 Dai Z, Liu H, Le QV, Tan M (2021) Coatnet: marrying convolution and attention for all data sizes. arXiv preprint. arXiv:​2106.​04803
Zurück zum Zitat Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: 2005 IEEE computer society conference on computer vision and pattern recognition (CVPR’05), vol 1. IEEE, pp 886–893 Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: 2005 IEEE computer society conference on computer vision and pattern recognition (CVPR’05), vol 1. IEEE, pp 886–893
Zurück zum Zitat de Freitas Pereira T, Anjos A, De Martino JM, Marcel S (2013) Can face anti-spoofing countermeasures work in a real world scenario? In: Proceedings of the 2013 international conference on biometrics (ICB). IEEE, pp 1–8 de Freitas Pereira T, Anjos A, De Martino JM, Marcel S (2013) Can face anti-spoofing countermeasures work in a real world scenario? In: Proceedings of the 2013 international conference on biometrics (ICB). IEEE, pp 1–8
Zurück zum Zitat Denil M, Shakibi B, Dinh L, Ranzato M, Freitas ND (2013) Predicting parameters in deep learning. University of British Columbia, Vancouver Denil M, Shakibi B, Dinh L, Ranzato M, Freitas ND (2013) Predicting parameters in deep learning. University of British Columbia, Vancouver
Zurück zum Zitat Denton EL, Zaremba W, Bruna J, LeCun Y, Fergus R (2014) Exploiting linear structure within convolutional networks for efficient evaluation. In: Advances in neural information processing systems, pp 1269–1277 Denton EL, Zaremba W, Bruna J, LeCun Y, Fergus R (2014) Exploiting linear structure within convolutional networks for efficient evaluation. In: Advances in neural information processing systems, pp 1269–1277
Zurück zum Zitat Devlin J, Chang M-W, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint. arXiv:1810.04805 Devlin J, Chang M-W, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint. arXiv:​1810.​04805
Zurück zum Zitat Dey N, Ren M, Dalca AV, Gerig G (2021) Generative adversarial registration for improved conditional deformable templates. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 3929–3941 Dey N, Ren M, Dalca AV, Gerig G (2021) Generative adversarial registration for improved conditional deformable templates. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 3929–3941
Zurück zum Zitat Ding H, Chen K, Huo Q (2019a) Compressing CNN–DBLSTM models for ocr with teacher–student learning and Tucker decomposition. Pattern Recogn 96:106957CrossRef Ding H, Chen K, Huo Q (2019a) Compressing CNN–DBLSTM models for ocr with teacher–student learning and Tucker decomposition. Pattern Recogn 96:106957CrossRef
Zurück zum Zitat Ding R, Chin T-W, Liu Z, Marculescu D (2019b) Regularizing activation distribution for training binarized deep networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11408–11417 Ding R, Chin T-W, Liu Z, Marculescu D (2019b) Regularizing activation distribution for training binarized deep networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11408–11417
Zurück zum Zitat Ding X, Hao T, Tan J, Liu J, Han J, Guo Y, Ding G (2021) ResRep: lossless CNN pruning via decoupling remembering and forgetting. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 4510–4520 Ding X, Hao T, Tan J, Liu J, Han J, Guo Y, Ding G (2021) ResRep: lossless CNN pruning via decoupling remembering and forgetting. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 4510–4520
Zurück zum Zitat Dong J-D, Cheng A-C, Juan D-C, Wei W, Sun M (2018) DPP-Net: deviceevice-aware progressive search for pareto-optimal neural architectures. In: Proceedings of the European conference on computer vision (ECCV), pp 517–531 Dong J-D, Cheng A-C, Juan D-C, Wei W, Sun M (2018) DPP-Net: deviceevice-aware progressive search for pareto-optimal neural architectures. In: Proceedings of the European conference on computer vision (ECCV), pp 517–531
Zurück zum Zitat Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, et al (2020) An image is worth 16 × 16 words: transformers for image recognition at scale. arXiv preprint. arXiv:2010.11929 Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, et al (2020) An image is worth 16 × 16 words: transformers for image recognition at scale. arXiv preprint. arXiv:​2010.​11929
Zurück zum Zitat Elsken T, Metzen JH, Hutter F (2019) Neural architecture search: a survey. J Mach Learn Res 20(1):1997–2017MathSciNetMATH Elsken T, Metzen JH, Hutter F (2019) Neural architecture search: a survey. J Mach Learn Res 20(1):1997–2017MathSciNetMATH
Zurück zum Zitat Erdogmus N, Marcel S (2014) Spoofing face recognition with 3d masks. IEEE Trans Inf Forensics Security 9(7):1084–1097CrossRef Erdogmus N, Marcel S (2014) Spoofing face recognition with 3d masks. IEEE Trans Inf Forensics Security 9(7):1084–1097CrossRef
Zurück zum Zitat Fang J, Sun Y, Zhang Q, Li Y, Liu W, Wang X (2020) Densely connected search space for more flexible neural architecture search. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10628–10637 Fang J, Sun Y, Zhang Q, Li Y, Liu W, Wang X (2020) Densely connected search space for more flexible neural architecture search. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10628–10637
Zurück zum Zitat Fukushima K, Miyake S (1982) Neocognitron: a self-organizing neural network model for a mechanism of visual pattern recognition. In: Competition and cooperation in neural nets. Springer, Heidelberg, pp 267–285 Fukushima K, Miyake S (1982) Neocognitron: a self-organizing neural network model for a mechanism of visual pattern recognition. In: Competition and cooperation in neural nets. Springer, Heidelberg, pp 267–285
Zurück zum Zitat George A, Marcel S (2021) Cross modal focal loss for rgbd face anti-spoofing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7882–7891 George A, Marcel S (2021) Cross modal focal loss for rgbd face anti-spoofing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7882–7891
Zurück zum Zitat Fukushima K (1989) Neocognitron: a hierarchical neural network capable of visual pattern recognition. Neural Netw 1:119–130 Fukushima K (1989) Neocognitron: a hierarchical neural network capable of visual pattern recognition. Neural Netw 1:119–130
Zurück zum Zitat Girshick R (2015) Fast R-CNN. In: Proceedings of the IEEE international conference on computer vision, pp 1440–1448 Girshick R (2015) Fast R-CNN. In: Proceedings of the IEEE international conference on computer vision, pp 1440–1448
Zurück zum Zitat Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 580–587 Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 580–587
Zurück zum Zitat Gong Y, Liu L, Ming Y, Bourdev L (2014) Compressing deep convolutional networks using vector quantization. Comput Sci+++ Gong Y, Liu L, Ming Y, Bourdev L (2014) Compressing deep convolutional networks using vector quantization. Comput Sci+++
Zurück zum Zitat Graham B, El-Nouby A, Touvron H, Stock P, Joulin A, Jégou H, Douze M (2021) Levit: a vision transformer in convnet’s clothing for faster inference. arXiv preprint. arXiv:2104.01136 Graham B, El-Nouby A, Touvron H, Stock P, Joulin A, Jégou H, Douze M (2021) Levit: a vision transformer in convnet’s clothing for faster inference. arXiv preprint. arXiv:​2104.​01136
Zurück zum Zitat Gu J, Wang Z, Kuen J, Ma L, Shahroudy A, Shuai B, Liu T, Wang X, Wang G, Cai J et al (2018) Recent advances in convolutional neural networks. Pattern Recogn 77:354–377CrossRef Gu J, Wang Z, Kuen J, Ma L, Shahroudy A, Shuai B, Liu T, Wang X, Wang G, Cai J et al (2018) Recent advances in convolutional neural networks. Pattern Recogn 77:354–377CrossRef
Zurück zum Zitat Gulcehre C, Cho K, Pascanu R, Bengio Y (2014) Learned-norm pooling for deep feedforward and recurrent neural networks. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Springer, pp. 530–546 Gulcehre C, Cho K, Pascanu R, Bengio Y (2014) Learned-norm pooling for deep feedforward and recurrent neural networks. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Springer, pp. 530–546
Zurück zum Zitat Guo Z, Zhang X, Mu H, Heng W, Liu Z, Wei Y, Sun J (2020) Single path one-shot neural architecture search with uniform sampling. In: European conference on computer vision. Springer, Munich, pp 544–560 Guo Z, Zhang X, Mu H, Heng W, Liu Z, Wei Y, Sun J (2020) Single path one-shot neural architecture search with uniform sampling. In: European conference on computer vision. Springer, Munich, pp 544–560
Zurück zum Zitat Guo J, Han K, Wang Y, Wu H, Chen X, Xu C, Xu C (2021) Distilling object detectors via decoupled features. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2154–2164 Guo J, Han K, Wang Y, Wu H, Chen X, Xu C, Xu C (2021) Distilling object detectors via decoupled features. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2154–2164
Zurück zum Zitat Han S, Mao H, Dally WJ (2015a) Deep compression: compressing deep neural networks with pruning, trained quantization and Huffman coding. Fiber 56(4):3–7 Han S, Mao H, Dally WJ (2015a) Deep compression: compressing deep neural networks with pruning, trained quantization and Huffman coding. Fiber 56(4):3–7
Zurück zum Zitat Han S, Pool J, Tran J, Dally WJ (2015b) Learning both weights and connections for efficient neural networks. MIT, Cambridge Han S, Pool J, Tran J, Dally WJ (2015b) Learning both weights and connections for efficient neural networks. MIT, Cambridge
Zurück zum Zitat Han D, Kim J, Kim J (2017) Deep pyramidal residual networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5927–5935 Han D, Kim J, Kim J (2017) Deep pyramidal residual networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5927–5935
Zurück zum Zitat Hanson S, Pratt L (1988) Comparing biases for minimal network construction with back-propagation. Adv Neural Inf Process Syst 1:177–185 Hanson S, Pratt L (1988) Comparing biases for minimal network construction with back-propagation. Adv Neural Inf Process Syst 1:177–185
Zurück zum Zitat Hassibi B, Stork DG, Wolff G, Watanabe T (1994) Optimal brain surgeon: extensions and performance comparison. In: Cowan JD, Tesauro G, Alspector J (eds) Advances in neural information processing systems, vol 6. Morgan Kaufmann, San Mateo, pp 263–270 Hassibi B, Stork DG, Wolff G, Watanabe T (1994) Optimal brain surgeon: extensions and performance comparison. In: Cowan JD, Tesauro G, Alspector J (eds) Advances in neural information processing systems, vol 6. Morgan Kaufmann, San Mateo, pp 263–270
Zurück zum Zitat He K, Sun J (2015) Convolutional neural networks at constrained time cost. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5353–5360 He K, Sun J (2015) Convolutional neural networks at constrained time cost. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5353–5360
Zurück zum Zitat He K, Zhang X, Ren S, Sun J (2015a) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904–1916CrossRef He K, Zhang X, Ren S, Sun J (2015a) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904–1916CrossRef
Zurück zum Zitat He K, Zhang X, Ren S, Sun J (2015b) Delving deep into rectifiers: surpassing human-level performance on ImageNet classification. In: Proceedings of the IEEE international conference on computer vision, pp 1026–1034 He K, Zhang X, Ren S, Sun J (2015b) Delving deep into rectifiers: surpassing human-level performance on ImageNet classification. In: Proceedings of the IEEE international conference on computer vision, pp 1026–1034
Zurück zum Zitat He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778 He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Zurück zum Zitat He Y, Zhang X, Sun J (2017) Channel pruning for accelerating very deep neural networks. In: Proceedings of the IEEE international conference on computer vision, pp 1389–1397 He Y, Zhang X, Sun J (2017) Channel pruning for accelerating very deep neural networks. In: Proceedings of the IEEE international conference on computer vision, pp 1389–1397
Zurück zum Zitat Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. Comput Sci 14(7):38–39 Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. Comput Sci 14(7):38–39
Zurück zum Zitat Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint. arXiv:1704.04861 Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint. arXiv:​1704.​04861
Zurück zum Zitat Howard A, Sandler M, Chu G, Chen L-C, Chen B, Tan M, Wang W, Zhu Y, Pang R, Vasudevan V et al (2019) Searching for mobilenetv3. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1314–1324 Howard A, Sandler M, Chu G, Chen L-C, Chen B, Tan M, Wang W, Zhu Y, Pang R, Vasudevan V et al (2019) Searching for mobilenetv3. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1314–1324
Zurück zum Zitat Hua W, Zhou Y, De Sa C, Zhang Z, Suh GE (2019) Boosting the performance of cnn accelerators with dynamic fine-grained channel gating. In: Proceedings of the 52nd Annual IEEE/ACM international symposium on microarchitecture, pp 139–150 Hua W, Zhou Y, De Sa C, Zhang Z, Suh GE (2019) Boosting the performance of cnn accelerators with dynamic fine-grained channel gating. In: Proceedings of the 52nd Annual IEEE/ACM international symposium on microarchitecture, pp 139–150
Zurück zum Zitat Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700–4708 Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700–4708
Zurück zum Zitat Huang G, Liu S, Van der Maaten L, Weinberger KQ (2018) Condensenet: an efficient densenet using learned group convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2752–2761 Huang G, Liu S, Van der Maaten L, Weinberger KQ (2018) Condensenet: an efficient densenet using learned group convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2752–2761
Zurück zum Zitat Huang X, Xu J, Tai Y-W, Tang C-K (2020) Fast video object segmentation with temporal aggregation network and dynamic template matching. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8879–8889 Huang X, Xu J, Tai Y-W, Tang C-K (2020) Fast video object segmentation with temporal aggregation network and dynamic template matching. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8879–8889
Zurück zum Zitat Huang H, Zhang J, Shan H (2021) When age-invariant face recognition meets face age synthesis: a multi-task learning framework. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7282–7291 Huang H, Zhang J, Shan H (2021) When age-invariant face recognition meets face age synthesis: a multi-task learning framework. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7282–7291
Zurück zum Zitat Hubel DH, Wiesel TN (1962) Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J Physiol 160(1):106–154CrossRef Hubel DH, Wiesel TN (1962) Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J Physiol 160(1):106–154CrossRef
Zurück zum Zitat Hubel DH, Wiesel TN (2009) Republication of The Journal of Physiology (1959) 148, 574-591: Receptive fields of single neurones in the cat's striate cortex. 1959. J Physiol 587(Pt 12):2721–2732 Hubel DH, Wiesel TN (2009) Republication of The Journal of Physiology (1959) 148, 574-591: Receptive fields of single neurones in the cat's striate cortex. 1959. J Physiol 587(Pt 12):2721–2732
Zurück zum Zitat Hu H, Peng R, Tai YW, Tang CK (2016) Network trimming: a data-driven neuron pruning approach towards efficient deep architectures. arXiv preprint. arXiv:1607.03250 Hu H, Peng R, Tai YW, Tang CK (2016) Network trimming: a data-driven neuron pruning approach towards efficient deep architectures. arXiv preprint. arXiv:​1607.​03250
Zurück zum Zitat Hu Q, Wang P, Cheng J (2018) From hashing to CNNs: training binary weight networks via hashing. In: 32nd AAAI conference on artificial intelligence, vol 32 Hu Q, Wang P, Cheng J (2018) From hashing to CNNs: training binary weight networks via hashing. In: 32nd AAAI conference on artificial intelligence, vol 32
Zurück zum Zitat Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning (PMLR), pp 448–456 Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning (PMLR), pp 448–456
Zurück zum Zitat Jaderberg M, Vedaldi A, Zisserman A (2014) Speeding up convolutional neural networks with low rank expansions. Comput Sci 4(4):XIII Jaderberg M, Vedaldi A, Zisserman A (2014) Speeding up convolutional neural networks with low rank expansions. Comput Sci 4(4):XIII
Zurück zum Zitat Jegou H, Douze M, Schmid C (2010) Product quantization for nearest neighbor search. IEEE Trans Pattern Anal Mach Intell 33(1):117–128CrossRef Jegou H, Douze M, Schmid C (2010) Product quantization for nearest neighbor search. IEEE Trans Pattern Anal Mach Intell 33(1):117–128CrossRef
Zurück zum Zitat Ji G-P, Fu K, Wu Z, Fan D-P, Shen J, Shao L (2021) Full-duplex strategy for video object segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, pp 4922–4933 Ji G-P, Fu K, Wu Z, Fan D-P, Shen J, Shao L (2021) Full-duplex strategy for video object segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, pp 4922–4933
Zurück zum Zitat Ji M, Shin S, Hwang S, Park G, Moon I-C (2021) Refine myself by teaching myself: Feature refinement via self-knowledge distillation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10664–10673 Ji M, Shin S, Hwang S, Park G, Moon I-C (2021) Refine myself by teaching myself: Feature refinement via self-knowledge distillation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10664–10673
Zurück zum Zitat Jia Y et al (2013) An open source convolutional architecture for fast feature embedding. In: Proceedings of the 22nd ACM international conference on multimedia, Orlando, FL, pp 675–678 Jia Y et al (2013) An open source convolutional architecture for fast feature embedding. In: Proceedings of the 22nd ACM international conference on multimedia, Orlando, FL, pp 675–678
Zurück zum Zitat Jo G, Lee G, Shin D (2020) Exploring group sparsity using dynamic sparse training. In: 2020 IEEE international conference on consumer electronics-Asia (ICCE-Asia). IEEE, pp 1–2 Jo G, Lee G, Shin D (2020) Exploring group sparsity using dynamic sparse training. In: 2020 IEEE international conference on consumer electronics-Asia (ICCE-Asia). IEEE, pp 1–2
Zurück zum Zitat Khan A, Sohail A, Zahoora U, Qureshi AS (2020) A survey of the recent architectures of deep convolutional neural networks. Artif Intell Rev 53(8):5455–5516CrossRef Khan A, Sohail A, Zahoora U, Qureshi AS (2020) A survey of the recent architectures of deep convolutional neural networks. Artif Intell Rev 53(8):5455–5516CrossRef
Zurück zum Zitat Kim YD, Park E, Yoo S, Choi T, Yang L, Shin D (2015) Compression of deep convolutional neural networks for fast and low power mobile applications. Comput Sci 71(2):576–584 Kim YD, Park E, Yoo S, Choi T, Yang L, Shin D (2015) Compression of deep convolutional neural networks for fast and low power mobile applications. Comput Sci 71(2):576–584
Zurück zum Zitat Kim S-W, Kook H-K, Sun J-Y, Kang M-C, Ko S-J (2018) Parallel feature pyramid network for object detection. In: Proceedings of the European conference on computer vision (ECCV), pp 234–250 Kim S-W, Kook H-K, Sun J-Y, Kang M-C, Ko S-J (2018) Parallel feature pyramid network for object detection. In: Proceedings of the European conference on computer vision (ECCV), pp 234–250
Zurück zum Zitat Kong T, Sun F, Tan C, Liu H, Huang W (2018) Deep feature pyramid reconfiguration for object detection. In: Proceedings of the European conference on computer vision (ECCV), pp 169–185 Kong T, Sun F, Tan C, Liu H, Huang W (2018) Deep feature pyramid reconfiguration for object detection. In: Proceedings of the European conference on computer vision (ECCV), pp 169–185
Zurück zum Zitat Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105 Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105
Zurück zum Zitat Lai Z, Lu E, Xie W (2020) Mast: a memory-augmented self-supervised tracker. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6479–6488 Lai Z, Lu E, Xie W (2020) Mast: a memory-augmented self-supervised tracker. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6479–6488
Zurück zum Zitat Lebedev V, Lempitsky V (2016) Fast convnets using group-wise brain damage. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2554–2564 Lebedev V, Lempitsky V (2016) Fast convnets using group-wise brain damage. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2554–2564
Zurück zum Zitat Lebedev V, Ganin Y, Rakhuba M, Oseledets I, Lempitsky V (2014) Speeding-up convolutional neural networks using fine-tuned cp-decomposition. In: International conference on learning representations (ICLR Poster) Lebedev V, Ganin Y, Rakhuba M, Oseledets I, Lempitsky V (2014) Speeding-up convolutional neural networks using fine-tuned cp-decomposition. In: International conference on learning representations (ICLR Poster)
Zurück zum Zitat LeCun Y, Denker JS, Solla SA (1990) Optimal brain damage. In: Touretzky DS (ed) Advances in neural information processing systems. Morgan Kaufmann, San Francisco, pp 598–605 LeCun Y, Denker JS, Solla SA (1990) Optimal brain damage. In: Touretzky DS (ed) Advances in neural information processing systems. Morgan Kaufmann, San Francisco, pp 598–605
Zurück zum Zitat LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324CrossRef LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324CrossRef
Zurück zum Zitat LeCun Y, Kavukcuoglu K, Farabet C (2010) Convolutional networks and applications in vision. In: Proceedings of 2010 IEEE international symposium on circuits and systems. IEEE, pp 253–256 LeCun Y, Kavukcuoglu K, Farabet C (2010) Convolutional networks and applications in vision. In: Proceedings of 2010 IEEE international symposium on circuits and systems. IEEE, pp 253–256
Zurück zum Zitat Lee KH, Verma N (2013) A low-power processor with configurable embedded machine-learning accelerators for high-order and adaptive analysis of medical-sensor signals. IEEE J Solid-State Circuits 48(7):1625–1637CrossRef Lee KH, Verma N (2013) A low-power processor with configurable embedded machine-learning accelerators for high-order and adaptive analysis of medical-sensor signals. IEEE J Solid-State Circuits 48(7):1625–1637CrossRef
Zurück zum Zitat Lee H, Wu Y-H, Lin Y-S, Chien S-Y (2019) Convolutional neural network accelerator with vector quantization. In: 2019 IEEE international symposium on circuits and systems (ISCAS). IEEE, pp 1–5 Lee H, Wu Y-H, Lin Y-S, Chien S-Y (2019) Convolutional neural network accelerator with vector quantization. In: 2019 IEEE international symposium on circuits and systems (ISCAS). IEEE, pp 1–5
Zurück zum Zitat Lee K, Kim H, Lee H, Shin D (2020) Flexible group-level pruning of deep neural networks for on-device machine learning. In: 2020 Design, automation & test in Europe cnference & exhibition (DATE). IEEE, pp 79–84 Lee K, Kim H, Lee H, Shin D (2020) Flexible group-level pruning of deep neural networks for on-device machine learning. In: 2020 Design, automation & test in Europe cnference & exhibition (DATE). IEEE, pp 79–84
Zurück zum Zitat Lee D, Wang D, Yang Y, Deng L, Zhao G, Li G (2021) QTTNet: quantized tensor train neural networks for 3D object and video recognition. Neural Netw 141:420–432 Lee D, Wang D, Yang Y, Deng L, Zhao G, Li G (2021) QTTNet: quantized tensor train neural networks for 3D object and video recognition. Neural Netw 141:420–432
Zurück zum Zitat Leng C, Dou Z, Li H, Zhu S, Jin R (2018) Extremely low bit neural network: Squeeze the last bit out with ADMM. In: Thirty-Second AAAI Conference on Artificial Intelligence Leng C, Dou Z, Li H, Zhu S, Jin R (2018) Extremely low bit neural network: Squeeze the last bit out with ADMM. In: Thirty-Second AAAI Conference on Artificial Intelligence
Zurück zum Zitat Li L, Zhu J, Sun M-T (2019) A spectral clustering based filter-level pruning method for convolutional neural networks. IEICE Trans Inf Syst 102(12):2624–2627CrossRef Li L, Zhu J, Sun M-T (2019) A spectral clustering based filter-level pruning method for convolutional neural networks. IEICE Trans Inf Syst 102(12):2624–2627CrossRef
Zurück zum Zitat Li C, Wang G, Wang B, Liang X, Li Z, Chang X (2021a) Dynamic slimmable network. In: Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, pp 8607–8617 Li C, Wang G, Wang B, Liang X, Li Z, Chang X (2021a) Dynamic slimmable network. In: Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, pp 8607–8617
Zurück zum Zitat Li Y, Ding W, Liu C, Zhang B, Guo G (2021b) TRQ: Ternary neural networks with residual quantization. In: Proceedings of the AAAI conference on artificial intelligence Li Y, Ding W, Liu C, Zhang B, Guo G (2021b) TRQ: Ternary neural networks with residual quantization. In: Proceedings of the AAAI conference on artificial intelligence
Zurück zum Zitat Li Y, Gu S, Mayer C, Gool LV, Timofte R (2020) Group sparsity: The hinge between filter pruning and decomposition for network compression. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8018–8027 Li Y, Gu S, Mayer C, Gool LV, Timofte R (2020) Group sparsity: The hinge between filter pruning and decomposition for network compression. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8018–8027
Zurück zum Zitat Lin D, Talathi S, Annapureddy S (2016) Fixed point quantization of deep convolutional networks. In: International conference on machine learning (PMLR), pp 2849–2858 Lin D, Talathi S, Annapureddy S (2016) Fixed point quantization of deep convolutional networks. In: International conference on machine learning (PMLR), pp 2849–2858
Zurück zum Zitat Lin X, Zhao C, Pan W (2017a) Towards accurate binary convolutional neural network. In: Advances in neural information processing systems Lin X, Zhao C, Pan W (2017a) Towards accurate binary convolutional neural network. In: Advances in neural information processing systems
Zurück zum Zitat Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017b) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117–2125 Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017b) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117–2125
Zurück zum Zitat Lin R, Ko C-Y, He Z, Chen C, Cheng Y, Yu H, Chesi G, Wong N (2020) Hotcake: higher order tucker articulated kernels for deeper CNN compression. In: 2020 IEEE 15th international conference on solid-state & integrated circuit technology (ICSICT). IEEE, pp 1–4 Lin R, Ko C-Y, He Z, Chen C, Cheng Y, Yu H, Chesi G, Wong N (2020) Hotcake: higher order tucker articulated kernels for deeper CNN compression. In: 2020 IEEE 15th international conference on solid-state & integrated circuit technology (ICSICT). IEEE, pp 1–4
Zurück zum Zitat Liu B, Wang M, Foroosh H, Tappen M, Pensky M (2015) Sparse convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 806–814 Liu B, Wang M, Foroosh H, Tappen M, Pensky M (2015) Sparse convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 806–814
Zurück zum Zitat Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) SSD: single shot multibox detector. In: European conference on computer vision. Springer, Cham, pp 21–37 Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) SSD: single shot multibox detector. In: European conference on computer vision. Springer, Cham, pp 21–37
Zurück zum Zitat Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C (2017) Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE international conference on computer vision, pp 2736–2744 Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C (2017) Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE international conference on computer vision, pp 2736–2744
Zurück zum Zitat Liu C, Zoph B, Neumann M, Shlens J, Hua W, Li L-J, Fei-Fei L, Yuille A, Huang J, Murphy K (2018a) Progressive neural architecture search. In: Proceedings of the European conference on computer vision (ECCV), pp 19–34 Liu C, Zoph B, Neumann M, Shlens J, Hua W, Li L-J, Fei-Fei L, Yuille A, Huang J, Murphy K (2018a) Progressive neural architecture search. In: Proceedings of the European conference on computer vision (ECCV), pp 19–34
Zurück zum Zitat Liu C, Chen L-C, Schroff F, Adam H, Hua W, Yuille AL, Fei-Fei L (2019) Auto-deeplab: hierarchical neural architecture search for semantic image segmentation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 82–92 Liu C, Chen L-C, Schroff F, Adam H, Hua W, Yuille AL, Fei-Fei L (2019) Auto-deeplab: hierarchical neural architecture search for semantic image segmentation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 82–92
Zurück zum Zitat Liu Z, Shen Z, Savvides M, Cheng K-T (2020a) Reactnet: towards precise binary neural network with generalized activation functions. In: European conference on computer vision. Springer, Berlin, pp 143–159 Liu Z, Shen Z, Savvides M, Cheng K-T (2020a) Reactnet: towards precise binary neural network with generalized activation functions. In: European conference on computer vision. Springer, Berlin, pp 143–159
Zurück zum Zitat Liu J, Xu Z, Shi R, Cheung RC, So HK (2020b) Dynamic sparse training: Find efficient sparse network from scratch with trainable masked layers. arXiv preprint. arXiv:2005.06870 Liu J, Xu Z, Shi R, Cheung RC, So HK (2020b) Dynamic sparse training: Find efficient sparse network from scratch with trainable masked layers. arXiv preprint. arXiv:​2005.​06870
Zurück zum Zitat Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021b) Swin transformer: hierarchical vision transformer using shifted windows. arXiv preprint. arXiv:2103.14030 Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021b) Swin transformer: hierarchical vision transformer using shifted windows. arXiv preprint. arXiv:​2103.​14030
Zurück zum Zitat Liu L, Zhang S, Kuang Z, Zhou A, Xue J-H, Wang X, Chen Y, Yang W, Liao Q, Zhang W (2021c) Group fisher pruning for practical network compression. In: International conference on machine learning (PMLR), pp 7021–7032 Liu L, Zhang S, Kuang Z, Zhou A, Xue J-H, Wang X, Chen Y, Yang W, Liao Q, Zhang W (2021c) Group fisher pruning for practical network compression. In: International conference on machine learning (PMLR), pp 7021–7032
Zurück zum Zitat Lowe DG (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60(2):91–110CrossRef Lowe DG (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60(2):91–110CrossRef
Zurück zum Zitat Lu Z, Whalen I, Boddeti V, Dhebar Y, Deb K, Goodman E, Banzhaf W (2019) NSGA-Net: neural architecture search using multi-objective genetic algorithm. In: Proceedings of the genetic and evolutionary computation conference, pp 419–427 Lu Z, Whalen I, Boddeti V, Dhebar Y, Deb K, Goodman E, Banzhaf W (2019) NSGA-Net: neural architecture search using multi-objective genetic algorithm. In: Proceedings of the genetic and evolutionary computation conference, pp 419–427
Zurück zum Zitat Luo P, Zhu Z, Liu Z, Wang X, Tang X (2016) Face model compression by distilling knowledge from neurons. In: Proceedings of the 30th AAAI conference on artificial intelligence. AAAI, Phoenix Luo P, Zhu Z, Liu Z, Wang X, Tang X (2016) Face model compression by distilling knowledge from neurons. In: Proceedings of the 30th AAAI conference on artificial intelligence. AAAI, Phoenix
Zurück zum Zitat Luo J-H, Wu J, Lin W (2017) Thinet: a filter level pruning method for deep neural network compression. In: Proceedings of the IEEE international conference on computer vision, pp 5058–5066 Luo J-H, Wu J, Lin W (2017) Thinet: a filter level pruning method for deep neural network compression. In: Proceedings of the IEEE international conference on computer vision, pp 5058–5066
Zurück zum Zitat Ma N, Zhang X, Zheng H-T, Sun J (2018) Shufflenet v2: Practical guidelines for efficient cnn architecture design. In: Proceedings of the European conference on computer vision (ECCV), pp 116–131 Ma N, Zhang X, Zheng H-T, Sun J (2018) Shufflenet v2: Practical guidelines for efficient cnn architecture design. In: Proceedings of the European conference on computer vision (ECCV), pp 116–131
Zurück zum Zitat Ma X, Guo F-M, Niu W, Lin X, Tang J, Ma K, Ren B, Wang Y (2020) PCONV: the missing but desirable sparsity in DNN weight pruning for real-time execution on mobile devices. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 5117–5124 Ma X, Guo F-M, Niu W, Lin X, Tang J, Ma K, Ren B, Wang Y (2020) PCONV: the missing but desirable sparsity in DNN weight pruning for real-time execution on mobile devices. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 5117–5124
Zurück zum Zitat Maas AL, Hannun AY, Ng AY, et al (2013) Rectifier nonlinearities improve neural network acoustic models. Proc ICML 30:3. Citeseer Maas AL, Hannun AY, Ng AY, et al (2013) Rectifier nonlinearities improve neural network acoustic models. Proc ICML 30:3. Citeseer
Zurück zum Zitat Mao H, Han S, Pool J, Li W, Liu X, Wang Y, Dally WJ (2017) Exploring the regularity of sparse structure in convolutional neural networks. arXiv preprint. arXiv:1705.08922 Mao H, Han S, Pool J, Li W, Liu X, Wang Y, Dally WJ (2017) Exploring the regularity of sparse structure in convolutional neural networks. arXiv preprint. arXiv:​1705.​08922
Zurück zum Zitat Mao Y, Wang N, Zhou W, Li H (2021) Joint inductive and transductive learning for video object segmentation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 9670–9679 Mao Y, Wang N, Zhou W, Li H (2021) Joint inductive and transductive learning for video object segmentation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 9670–9679
Zurück zum Zitat Maziarz K, Tan M, Khorlin A, Chang K-YS, Gesmundo A (2019) Evo-nas: Evolutionary-neural hybrid agent for architecture search Maziarz K, Tan M, Khorlin A, Chang K-YS, Gesmundo A (2019) Evo-nas: Evolutionary-neural hybrid agent for architecture search
Zurück zum Zitat Mehta S, Rastegari M (2021) MobileViT: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv preprint. arXiv:2110.02178 Mehta S, Rastegari M (2021) MobileViT: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv preprint. arXiv:​2110.​02178
Zurück zum Zitat Michel G, Alaoui MA, Lebois A, Feriani A, Felhi M (2019) Dvolver: Efficient pareto-optimal neural network architecture search. arXiv preprint. arXiv:1902.01654 Michel G, Alaoui MA, Lebois A, Feriani A, Felhi M (2019) Dvolver: Efficient pareto-optimal neural network architecture search. arXiv preprint. arXiv:​1902.​01654
Zurück zum Zitat Mikolov T, Karafiát M, Burget L, Černocky J, Khudanpur S (2010) Eleventh annual conference of the international speech communication association Mikolov T, Karafiát M, Burget L, Černocky J, Khudanpur S (2010) Eleventh annual conference of the international speech communication association
Zurück zum Zitat Montremerlo M, Beeker J, Bhat S, Dahlkamp H (2008) The stanford entry in the urban challenge. J Field Robot 7(9):468–492 Montremerlo M, Beeker J, Bhat S, Dahlkamp H (2008) The stanford entry in the urban challenge. J Field Robot 7(9):468–492
Zurück zum Zitat Nair V, Hinton GE (2010) Rectified linear units improve restricted Boltzmann machines. In: International conference on international conference on machine learning (ICML) Nair V, Hinton GE (2010) Rectified linear units improve restricted Boltzmann machines. In: International conference on international conference on machine learning (ICML)
Zurück zum Zitat Nguyen DT, Nguyen TN, Kim H, Lee H-J (2019) A high-throughput and power-efficient FPGA implementation of YOLO CNN for object detection. IEEE Trans Very Large Scale Integr (VLSI) Syst 27(8):1861–1873CrossRef Nguyen DT, Nguyen TN, Kim H, Lee H-J (2019) A high-throughput and power-efficient FPGA implementation of YOLO CNN for object detection. IEEE Trans Very Large Scale Integr (VLSI) Syst 27(8):1861–1873CrossRef
Zurück zum Zitat Niu W, Ma X, Lin S, Wang S, Qian X, Lin X, Wang Y, Ren B (2020) PatDNN: achieving real-time DNN execution on mobile devices with pattern-based weight pruning. In: Proceedings of the 25th international conference on architectural support for programming languages and operating systems, pp 907–922 Niu W, Ma X, Lin S, Wang S, Qian X, Lin X, Wang Y, Ren B (2020) PatDNN: achieving real-time DNN execution on mobile devices with pattern-based weight pruning. In: Proceedings of the 25th international conference on architectural support for programming languages and operating systems, pp 907–922
Zurück zum Zitat Peng Z, Huang W, Gu S, Xie L, Wang Y, Jiao J, Ye Q (2021) Conformer: local features coupling global representations for visual recognition. arXiv preprint. arXiv:2105.03889 Peng Z, Huang W, Gu S, Xie L, Wang Y, Jiao J, Ye Q (2021) Conformer: local features coupling global representations for visual recognition. arXiv preprint. arXiv:​2105.​03889
Zurück zum Zitat Punyani P, Gupta R, Kumar A (2020) Neural networks for facial age estimation: a survey on recent advances. Artif Intell Rev 53(5):3299–3347CrossRef Punyani P, Gupta R, Kumar A (2020) Neural networks for facial age estimation: a survey on recent advances. Artif Intell Rev 53(5):3299–3347CrossRef
Zurück zum Zitat Qin Z, Li Z, Zhang Z, Bao Y, Yu G, Peng Y, Sun J (2019) Thundernet: towards real-time generic object detection on mobile devices. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 6718–6727 Qin Z, Li Z, Zhang Z, Bao Y, Yu G, Peng Y, Sun J (2019) Thundernet: towards real-time generic object detection on mobile devices. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 6718–6727
Zurück zum Zitat Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding by generative pre-training Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding by generative pre-training
Zurück zum Zitat Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I et al (2019) Language models are unsupervised multitask learners. OpenAI Blog 1(8):9 Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I et al (2019) Language models are unsupervised multitask learners. OpenAI Blog 1(8):9
Zurück zum Zitat Rastegari M, Ordonez V, Redmon J, Farhadi A (2016) XNOR-Net: Imagenet classification using binary convolutional neural networks. In: European conference on computer vision. Springer, Cham, pp 525–542 Rastegari M, Ordonez V, Redmon J, Farhadi A (2016) XNOR-Net: Imagenet classification using binary convolutional neural networks. In: European conference on computer vision. Springer, Cham, pp 525–542
Zurück zum Zitat Razani R, Morin G, Sari E, Nia VP (2021) Adaptive binary-ternary quantization. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4613–4618 Razani R, Morin G, Sari E, Nia VP (2021) Adaptive binary-ternary quantization. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4613–4618
Zurück zum Zitat Real E, Aggarwal A, Huang Y, Le QV (2019) Regularized evolution for image classifier architecture search. In: Proceedings of the aaai conference on artificial intelligence, vol 33, pp 4780–4789 Real E, Aggarwal A, Huang Y, Le QV (2019) Regularized evolution for image classifier architecture search. In: Proceedings of the aaai conference on artificial intelligence, vol 33, pp 4780–4789
Zurück zum Zitat Redfern AJ, Zhu L, Newquist MK (2021) BCNN: a binary CNN with all matrix OPS quantized to 1 bit precision. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4604–4612 Redfern AJ, Zhu L, Newquist MK (2021) BCNN: a binary CNN with all matrix OPS quantized to 1 bit precision. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4604–4612
Zurück zum Zitat Redmon J, Farhadi A (2017) Yolo9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7263–7271 Redmon J, Farhadi A (2017) Yolo9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7263–7271
Zurück zum Zitat Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779–788 Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779–788
Zurück zum Zitat Ren S, He K, Girshick R, Sun J (2015) Faster R-CNN: towards real-time object detection with region proposal networks. Adv Neural Inf Process Syst 28:91–99 Ren S, He K, Girshick R, Sun J (2015) Faster R-CNN: towards real-time object detection with region proposal networks. Adv Neural Inf Process Syst 28:91–99
Zurück zum Zitat Romero A, Ballas N, Kahou SE, Chassang A, Gatta C, Bengio Y (2014) Fitnets: Hints for thin deep nets. rXiv preprint. arXiv:1412.6550 Romero A, Ballas N, Kahou SE, Chassang A, Gatta C, Bengio Y (2014) Fitnets: Hints for thin deep nets. rXiv preprint. arXiv:​1412.​6550
Zurück zum Zitat Roy AG, Siddiqui S, Pölsterl S, Navab N, Wachinger C (2020) squeeze & exciteguided few-shot segmentation of volumetric images. Med Image Anal 59:101587CrossRef Roy AG, Siddiqui S, Pölsterl S, Navab N, Wachinger C (2020) squeeze & exciteguided few-shot segmentation of volumetric images. Med Image Anal 59:101587CrossRef
Zurück zum Zitat Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510–4520 Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510–4520
Zurück zum Zitat Seong H, Hyun J, Kim E (2020) Kernelized memory network for video object segmentation. In: European conference on computer vision. Springer, Cham, pp 629–645 Seong H, Hyun J, Kim E (2020) Kernelized memory network for video object segmentation. In: European conference on computer vision. Springer, Cham, pp 629–645
Zurück zum Zitat Sharma M, Markopoulos PP, Saber E, Asif MS, Prater-Bennette A (2021) Convolutional auto-encoder with tensor-train factorization. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 198–206 Sharma M, Markopoulos PP, Saber E, Asif MS, Prater-Bennette A (2021) Convolutional auto-encoder with tensor-train factorization. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 198–206
Zurück zum Zitat Simard PY, Steinkraus D, Platt JC et al (2003) Best practices for convolutional neural networks applied to visual document analysis. In: ICDAR, vol 3 Simard PY, Steinkraus D, Platt JC et al (2003) Best practices for convolutional neural networks applied to visual document analysis. In: ICDAR, vol 3
Zurück zum Zitat Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arxiv preprint. arXiv:1409.1556 Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arxiv preprint. arXiv:​1409.​1556
Zurück zum Zitat Srinivas S, Sarvadevabhatla RK, Mopuri KR, Prabhu N, Kruthiventi SS, Babu RV (2016) A taxonomy of deep convolutional neural nets for computer vision. Front Robotics AI 2:36CrossRef Srinivas S, Sarvadevabhatla RK, Mopuri KR, Prabhu N, Kruthiventi SS, Babu RV (2016) A taxonomy of deep convolutional neural nets for computer vision. Front Robotics AI 2:36CrossRef
Zurück zum Zitat Srinivas A, Lin T-Y, Parmar N, Shlens J, Abbeel P, Vaswani A (2021) Bottleneck transformers for visual recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 16519–16529 Srinivas A, Lin T-Y, Parmar N, Shlens J, Abbeel P, Vaswani A (2021) Bottleneck transformers for visual recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 16519–16529
Zurück zum Zitat Srivastava RK, Greff K, Schmidhuber J (2015) Training very deep networks. In: Advances in neural information processing systems Srivastava RK, Greff K, Schmidhuber J (2015) Training very deep networks. In: Advances in neural information processing systems
Zurück zum Zitat Sun W, Zhou A, Stuijk S, Wijnhoven R, Nelson AO, Corporaal H et al (2021) DominoSearch: find layer-wise fine-grained N:M sparse schemes from dense neural networks. In: Advances in neural information processing systems 34 (NeurIPS 2021) Sun W, Zhou A, Stuijk S, Wijnhoven R, Nelson AO, Corporaal H et al (2021) DominoSearch: find layer-wise fine-grained N:M sparse schemes from dense neural networks. In: Advances in neural information processing systems 34 (NeurIPS 2021)
Zurück zum Zitat Sundermeyer M, Schlüter R, Ney H (2012) Lstm neural networks for language modeling. In: 13th annual conference of the international speech communication association Sundermeyer M, Schlüter R, Ney H (2012) Lstm neural networks for language modeling. In: 13th annual conference of the international speech communication association
Zurück zum Zitat Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1–9 Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1–9
Zurück zum Zitat Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: 31st AAAI conference on artificial intelligence Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: 31st AAAI conference on artificial intelligence
Zurück zum Zitat Takahashi N, Mitsufuji Y (2021) Densely connected multi-dilated convolutional networks for dense prediction tasks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 993–1002 Takahashi N, Mitsufuji Y (2021) Densely connected multi-dilated convolutional networks for dense prediction tasks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 993–1002
Zurück zum Zitat Tan M, Chen B, Pang R, Vasudevan V, Sandler M, Howard A, Le QV (2019) Mnasnet: platform-aware neural architecture search for mobile. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2820–2828 Tan M, Chen B, Pang R, Vasudevan V, Sandler M, Howard A, Le QV (2019) Mnasnet: platform-aware neural architecture search for mobile. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2820–2828
Zurück zum Zitat Tang W, Hua G, Wang L (2017) How to train a compact binary neural network with high accuracy? In: 31st AAAI conference on artificial intelligence Tang W, Hua G, Wang L (2017) How to train a compact binary neural network with high accuracy? In: 31st AAAI conference on artificial intelligence
Zurück zum Zitat Tang H, Liu X, Sun S, Yan X, Xie X (2021a) Recurrent mask refinement for few-shot medical image segmentation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 3918–3928 Tang H, Liu X, Sun S, Yan X, Xie X (2021a) Recurrent mask refinement for few-shot medical image segmentation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 3918–3928
Zurück zum Zitat Tang Y, Wang Y, Xu Y, Deng Y, Xu C, Tao D, Xu C (2021a) Manifold regularized dynamic network pruning. In: Proceedings of the IEEE/CVF international conference on computer vision and pattern recognition, pp 5018–5028 Tang Y, Wang Y, Xu Y, Deng Y, Xu C, Tao D, Xu C (2021a) Manifold regularized dynamic network pruning. In: Proceedings of the IEEE/CVF international conference on computer vision and pattern recognition, pp 5018–5028
Zurück zum Zitat Theis L, Korshunova I, Tejani A, Huszár F (2018) Faster gaze prediction with dense networks and fisher pruning. arXiv preprint. arXiv:1801.05787 Theis L, Korshunova I, Tejani A, Huszár F (2018) Faster gaze prediction with dense networks and fisher pruning. arXiv preprint. arXiv:​1801.​05787
Zurück zum Zitat Touvron H, Cord M, Douze M, Massa F, Sablayrolles A, Jégou H (2021) Training data-efficient image transformers and distillation through attention. In: International conference on machine learning (PMLR), pp 10347–10357 Touvron H, Cord M, Douze M, Massa F, Sablayrolles A, Jégou H (2021) Training data-efficient image transformers and distillation through attention. In: International conference on machine learning (PMLR), pp 10347–10357
Zurück zum Zitat Uijlings JR, Van De Sande KE, Gevers T, Smeulders AW (2013) Selective search for object recognition. Int J Comput Vis 104(2):154–171CrossRef Uijlings JR, Van De Sande KE, Gevers T, Smeulders AW (2013) Selective search for object recognition. Int J Comput Vis 104(2):154–171CrossRef
Zurück zum Zitat Vanhoucke V, Senior A, Mao MZ (2011) Improving the speed of neural networks on CPUS. In: Deep learning and unsupervised feature learning workshop Vanhoucke V, Senior A, Mao MZ (2011) Improving the speed of neural networks on CPUS. In: Deep learning and unsupervised feature learning workshop
Zurück zum Zitat Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008 Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008
Zurück zum Zitat Wang P, Cheng J (2016) Accelerating convolutional neural networks for mobile applications. In: Proceedings of the 24th ACM international conference on Multimedia, pp 541–545 Wang P, Cheng J (2016) Accelerating convolutional neural networks for mobile applications. In: Proceedings of the 24th ACM international conference on Multimedia, pp 541–545
Zurück zum Zitat Wang P, Cheng J (2017) Fixed-point factorized networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4012–4020 Wang P, Cheng J (2017) Fixed-point factorized networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4012–4020
Zurück zum Zitat Wang P, Hu Q, Zhang Y, Zhang C, Liu Y, Cheng J (2018) Two-step quantization for low-bit neural networks. In: Proceedings of the IEEE Conference on computer vision and pattern recognition, pp 4376–4384 Wang P, Hu Q, Zhang Y, Zhang C, Liu Y, Cheng J (2018) Two-step quantization for low-bit neural networks. In: Proceedings of the IEEE Conference on computer vision and pattern recognition, pp 4376–4384
Zurück zum Zitat Wang Z, Lin S, Xie J, Lin Y (2019a) Pruning blocks for cnn compression and acceleration via online ensemble distillation. IEEE Access 7:175703–175716CrossRef Wang Z, Lin S, Xie J, Lin Y (2019a) Pruning blocks for cnn compression and acceleration via online ensemble distillation. IEEE Access 7:175703–175716CrossRef
Zurück zum Zitat Wang W, Fu C, Guo J, Cai D, He X (2019b) COP: customized deep model compression via regularized correlation-based filter-level pruning. Neurocomputing 464:533–545 Wang W, Fu C, Guo J, Cai D, He X (2019b) COP: customized deep model compression via regularized correlation-based filter-level pruning. Neurocomputing 464:533–545
Zurück zum Zitat Wang Z, Lu J, Tao C, Zhou J, Tian Q (2019c) Learning channel-wise interactions for binary convolutional neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 568–577 Wang Z, Lu J, Tao C, Zhou J, Tian Q (2019c) Learning channel-wise interactions for binary convolutional neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 568–577
Zurück zum Zitat Wang C-Y, Liao H-YM, Wu Y-H, Chen P-Y, Hsieh J-W, Yeh I-H (2020a) Cspnet: a new backbone that can enhance learning capability of CNN. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp 390–391 Wang C-Y, Liao H-YM, Wu Y-H, Chen P-Y, Hsieh J-W, Yeh I-H (2020a) Cspnet: a new backbone that can enhance learning capability of CNN. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp 390–391
Zurück zum Zitat Wang N, Zhou W, Li H (2020b) Contrastive transformation for self-supervised correspondence learning. arXiv preprint. arXiv:2012.05057 Wang N, Zhou W, Li H (2020b) Contrastive transformation for self-supervised correspondence learning. arXiv preprint. arXiv:​2012.​05057
Zurück zum Zitat Wang D, Li M, Gong C, Chandra V (2021a) Attentivenas: improving neural architecture search via attentive sampling. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6418–6427 Wang D, Li M, Gong C, Chandra V (2021a) Attentivenas: improving neural architecture search via attentive sampling. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6418–6427
Zurück zum Zitat Wang Y, Xu Z, Wang X, Shen C, Cheng B, Shen H, Xia H (2021b) End-to-end video instance segmentation with transformers. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8741–8750 Wang Y, Xu Z, Wang X, Shen C, Cheng B, Shen H, Xia H (2021b) End-to-end video instance segmentation with transformers. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8741–8750
Zurück zum Zitat Wang D, Zhao G, Chen H, Liu Z, Deng L, Li G (2021c) Nonlinear tensor train format for deep neural network compression. Neural Netw 144:320–333CrossRef Wang D, Zhao G, Chen H, Liu Z, Deng L, Li G (2021c) Nonlinear tensor train format for deep neural network compression. Neural Netw 144:320–333CrossRef
Zurück zum Zitat Wang C, Liu B, Liu L, Zhu Y, Hou J, Liu P, Li X (2021d) A review of deep learning used in the hyperspectral image analysis for agriculture. Artif Intell Rev 54:5205–5253 Wang C, Liu B, Liu L, Zhu Y, Hou J, Liu P, Li X (2021d) A review of deep learning used in the hyperspectral image analysis for agriculture. Artif Intell Rev 54:5205–5253
Zurück zum Zitat Wang Z, Xiao H, Lu J, Zhou J (2021e) Generalizable mixed-precision quantization via attribution rank preservation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 5291–5300 Wang Z, Xiao H, Lu J, Zhou J (2021e) Generalizable mixed-precision quantization via attribution rank preservation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 5291–5300
Zurück zum Zitat Wang W, Xie E, Li X, Fan D-P, Song K, Liang D, Lu T, Luo P, Shao L (2021f) Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. arXiv preprint. arXiv:2102.12122 Wang W, Xie E, Li X, Fan D-P, Song K, Liang D, Lu T, Luo P, Shao L (2021f) Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. arXiv preprint. arXiv:​2102.​12122
Zurück zum Zitat Wen W, Wu C, Wang Y, Chen Y, Li H (2016) Learning structured sparsity in deep neural networks. Adv Neural Inf Process Syst 29:2074–2082 Wen W, Wu C, Wang Y, Chen Y, Li H (2016) Learning structured sparsity in deep neural networks. Adv Neural Inf Process Syst 29:2074–2082
Zurück zum Zitat Wen W, Xu C, Yan F, Wu C, Wang Y, Chen Y, Li H (2017) Terngrad: Ternary gradients to reduce communication in distributed deep learning. arXiv preprint. arXiv:1705.07878 Wen W, Xu C, Yan F, Wu C, Wang Y, Chen Y, Li H (2017) Terngrad: Ternary gradients to reduce communication in distributed deep learning. arXiv preprint. arXiv:​1705.​07878
Zurück zum Zitat Wu J, Leng C, Wang Y, Hu Q, Cheng J (2016) Quantized convolutional neural networks for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4820–4828 Wu J, Leng C, Wang Y, Hu Q, Cheng J (2016) Quantized convolutional neural networks for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4820–4828
Zurück zum Zitat Wu Y, Wu Y, Gong R, Lv Y, Chen K, Liang D, Hu X, Liu X, Yan J (2020a) Rotation consistent margin loss for efficient low-bit face recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6866–6876 Wu Y, Wu Y, Gong R, Lv Y, Chen K, Liang D, Hu X, Liu X, Yan J (2020a) Rotation consistent margin loss for efficient low-bit face recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6866–6876
Zurück zum Zitat Wu B, Xu C, Dai X, Wan A, Zhang P, Yan Z, Tomizuka M, Gonzalez J, Keutzer K, Vajda P (2020b) Visual transformers: token-based image representation and processing for computer vision. arXiv preprint. arXiv:2006.03677 Wu B, Xu C, Dai X, Wan A, Zhang P, Yan Z, Tomizuka M, Gonzalez J, Keutzer K, Vajda P (2020b) Visual transformers: token-based image representation and processing for computer vision. arXiv preprint. arXiv:​2006.​03677
Zurück zum Zitat Xie L, Yuille A (2017) Genetic CNN. In: Proceedings of the IEEE international conference on computer vision, pp 1379–1388 Xie L, Yuille A (2017) Genetic CNN. In: Proceedings of the IEEE international conference on computer vision, pp 1379–1388
Zurück zum Zitat Xie S, Girshick R, Dollár P, Tu Z, He K (2017) Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1492–1500 Xie S, Girshick R, Dollár P, Tu Z, He K (2017) Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1492–1500
Zurück zum Zitat Xu H (2020) Pnfm: a filter level pruning method for cnn compression. In: Proceedings of the 3rd international conference on information technologies and electrical engineering, pp 49–54 Xu H (2020) Pnfm: a filter level pruning method for cnn compression. In: Proceedings of the 3rd international conference on information technologies and electrical engineering, pp 49–54
Zurück zum Zitat Xu B, Wang N, Chen T, Li M (2015) Empirical evaluation of rectified activations in convolutional network. arXiv preprint. arXiv:1505.00853 Xu B, Wang N, Chen T, Li M (2015) Empirical evaluation of rectified activations in convolutional network. arXiv preprint. arXiv:​1505.​00853
Zurück zum Zitat Xu Y, Xie L, Zhang X, Chen X, Qi G-J, Tian Q, Xiong H (2019) PC-DARTS: partial channel connections for memory-efficient architecture search. arXiv preprint. arXiv:1907.05737 Xu Y, Xie L, Zhang X, Chen X, Qi G-J, Tian Q, Xiong H (2019) PC-DARTS: partial channel connections for memory-efficient architecture search. arXiv preprint. arXiv:​1907.​05737
Zurück zum Zitat Xu Y, Wang Y, Han K, Tang Y, Jui S, Xu C, Xu C (2021) Renas: relativistic evaluation of neural architecture search. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4411–4420 Xu Y, Wang Y, Han K, Tang Y, Jui S, Xu C, Xu C (2021) Renas: relativistic evaluation of neural architecture search. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4411–4420
Zurück zum Zitat Yamamoto K (2021) Learnable companding quantization for accurate low-bit neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5029–5038 Yamamoto K (2021) Learnable companding quantization for accurate low-bit neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5029–5038
Zurück zum Zitat Yang Z, Wang Y, Chen X, Shi B, Xu C, Xu C, Tian Q, Xu C (2020a) CARS: continuous evolution for efficient neural architecture search. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1829–1838 Yang Z, Wang Y, Chen X, Shi B, Xu C, Xu C, Tian Q, Xu C (2020a) CARS: continuous evolution for efficient neural architecture search. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1829–1838
Zurück zum Zitat Yang Z, Wang Y, Han K, Xu C, Xu C, Tao D, Xu C (2020b) Searching for low-bit weights in quantized neural networks. arXiv preprint. arXiv:2009.08695 Yang Z, Wang Y, Han K, Xu C, Xu C, Tao D, Xu C (2020b) Searching for low-bit weights in quantized neural networks. arXiv preprint. arXiv:​2009.​08695
Zurück zum Zitat Yang Z, Wang Y, Chen X, Guo J, Zhang W, Xu C, Xu C, Tao D, Xu C (2021) Hournas: extremely fast neural architecture search through an hourglass lens. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10896–10906 Yang Z, Wang Y, Chen X, Guo J, Zhang W, Xu C, Xu C, Tao D, Xu C (2021) Hournas: extremely fast neural architecture search through an hourglass lens. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10896–10906
Zurück zum Zitat Yao L, Pi R, Xu H, Zhang W, Li Z, Zhang T (2021) G-DetKD: towards general distillation framework for object detectors via contrastive and semantic-guided feature imitation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 3591–3600 Yao L, Pi R, Xu H, Zhang W, Li Z, Zhang T (2021) G-DetKD: towards general distillation framework for object detectors via contrastive and semantic-guided feature imitation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 3591–3600
Zurück zum Zitat Ye J, Li G, Chen D, Yang H, Zhe S, Xu Z (2020) Block-term tensor neural networks. Neural Netw 130:11–21CrossRef Ye J, Li G, Chen D, Yang H, Zhe S, Xu Z (2020) Block-term tensor neural networks. Neural Netw 130:11–21CrossRef
Zurück zum Zitat Ye M, Kanski M, Yang D, Chang Q, Yan Z, Huang Q, Axel L, Metaxas D (2021) DeepTag: an unsupervised deep learning method for motion tracking on cardiac tagging magnetic resonance images. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7261–7271 Ye M, Kanski M, Yang D, Chang Q, Yan Z, Huang Q, Axel L, Metaxas D (2021) DeepTag: an unsupervised deep learning method for motion tracking on cardiac tagging magnetic resonance images. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7261–7271
Zurück zum Zitat Yuan L, Wang T, Zhang X, Tay FE, Jie Z, Liu W, Feng J (2020) Central similarity quantization for efficient image and video retrieval. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3083–3092 Yuan L, Wang T, Zhang X, Tay FE, Jie Z, Liu W, Feng J (2020) Central similarity quantization for efficient image and video retrieval. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3083–3092
Zurück zum Zitat Yuan L, Chen Y, Wang T, Yu W, Shi Y, Jiang Z, Tay FE, Feng J, Yan S (2021) Tokens-to-token vit: training vision transformers from scratch on imagenet. arXiv preprint. arXiv:2101.11986 Yuan L, Chen Y, Wang T, Yu W, Shi Y, Jiang Z, Tay FE, Feng J, Yan S (2021) Tokens-to-token vit: training vision transformers from scratch on imagenet. arXiv preprint. arXiv:​2101.​11986
Zurück zum Zitat Zagoruyko S, Komodakis N (2016a) Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. arXiv preprint. arXiv:1612.03928 Zagoruyko S, Komodakis N (2016a) Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. arXiv preprint. arXiv:​1612.​03928
Zurück zum Zitat Zagoruyko S, Komodakis N (2016b) Wide residual networks, British Machine Vision Conference Zagoruyko S, Komodakis N (2016b) Wide residual networks, British Machine Vision Conference
Zurück zum Zitat Zeiler MD, Fergus R (2014) Visualizing and understanding convolutional networks. In: European conference on computer vision. Springer, Cham, pp 818–833 Zeiler MD, Fergus R (2014) Visualizing and understanding convolutional networks. In: European conference on computer vision. Springer, Cham, pp 818–833
Zurück zum Zitat Zhang X, Zou J, He K, Sun J (2015) Accelerating very deep convolutional networks for classification and detection. IEEE Trans Pattern Anal Mach Intell 38(10):1943–1955CrossRef Zhang X, Zou J, He K, Sun J (2015) Accelerating very deep convolutional networks for classification and detection. IEEE Trans Pattern Anal Mach Intell 38(10):1943–1955CrossRef
Zurück zum Zitat Zhang Q, Zhang M, Chen T, Sun Z, Ma Y, Yu B (2019) Recent advances in convolutional neural network acceleration. Neurocomputing 323:37–51CrossRef Zhang Q, Zhang M, Chen T, Sun Z, Ma Y, Yu B (2019) Recent advances in convolutional neural network acceleration. Neurocomputing 323:37–51CrossRef
Zurück zum Zitat Zhang T, Cheng H-P, Li Z, Yan F, Huang C, Li H, Chen Y (2020) Autoshrink: a topology-aware NAS for discovering efficient neural architecture. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 6829–6836 Zhang T, Cheng H-P, Li Z, Yan F, Huang C, Li H, Chen Y (2020) Autoshrink: a topology-aware NAS for discovering efficient neural architecture. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 6829–6836
Zurück zum Zitat Zhang Z, Lu X, Cao G, Yang Y, Jiao L, Liu F (2021a) Vit-yolo: Transformer-based yolo for object detection. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 2799–2808 Zhang Z, Lu X, Cao G, Yang Y, Jiao L, Liu F (2021a) Vit-yolo: Transformer-based yolo for object detection. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 2799–2808
Zurück zum Zitat Zhang C, Yuan G, Niu W, Tian J, Jin S, Zhuang D, Jiang Z, Wang Y, Ren B, Song SL et al (2021b) Clicktrain: efficient and accurate end-to-end deep learning training via fine-grained architecture-preserving pruning. In: Proceedings of the ACM international conference on supercomputing, pp 266–278 Zhang C, Yuan G, Niu W, Tian J, Jin S, Zhuang D, Jiang Z, Wang Y, Ren B, Song SL et al (2021b) Clicktrain: efficient and accurate end-to-end deep learning training via fine-grained architecture-preserving pruning. In: Proceedings of the ACM international conference on supercomputing, pp 266–278
Zurück zum Zitat Zhao Q, Sheng T, Wang Y, Tang Z, Chen Y, Cai L, Ling H (2019) M2det: a single-shot object detector based on multi-level feature pyramid network. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 9259–9266 Zhao Q, Sheng T, Wang Y, Tang Z, Chen Y, Cai L, Ling H (2019) M2det: a single-shot object detector based on multi-level feature pyramid network. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 9259–9266
Zurück zum Zitat Zhao T, Cao K, Yao J, Nogues I, Lu L, Huang L, Xiao J, Yin Z, Zhang L (2021a) 3D graph anatomy geometry-integrated network for pancreatic mass segmentation, diagnosis, and quantitative patient management. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13743–13752 Zhao T, Cao K, Yao J, Nogues I, Lu L, Huang L, Xiao J, Yin Z, Zhang L (2021a) 3D graph anatomy geometry-integrated network for pancreatic mass segmentation, diagnosis, and quantitative patient management. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13743–13752
Zurück zum Zitat Zhao H, Zhou W, Chen D, Wei T, Zhang W, Yu N (2021b) Multi-attentional deepfake detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2185–2194 Zhao H, Zhou W, Chen D, Wei T, Zhang W, Yu N (2021b) Multi-attentional deepfake detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2185–2194
Zurück zum Zitat Zhong Z, Yang Z, Deng B, Yan J, Wu W, Shao J, Liu C-L (2020) Blockqnn: efficient block-wise neural network architecture generation. IEEE Trans Pattern Anal Mach Intell 43(7):2314–2328 Zhong Z, Yang Z, Deng B, Yan J, Wu W, Shao J, Liu C-L (2020) Blockqnn: efficient block-wise neural network architecture generation. IEEE Trans Pattern Anal Mach Intell 43(7):2314–2328
Zurück zum Zitat Zhou H, Alvarez JM, Porikli F (2016) Less is more: towards compact CNNs. In: European conference on computer vision. Springer, Berlin, pp 662–677 Zhou H, Alvarez JM, Porikli F (2016) Less is more: towards compact CNNs. In: European conference on computer vision. Springer, Berlin, pp 662–677
Zurück zum Zitat Zhou M, Liu Y, Long Z, Chen L, Zhu C (2019) Tensor rank learning in CP decomposition via convolutional neural network. Signal Process Image Commun 73:12–21 Zhou M, Liu Y, Long Z, Chen L, Zhu C (2019) Tensor rank learning in CP decomposition via convolutional neural network. Signal Process Image Commun 73:12–21
Zurück zum Zitat Zhou S, Wang Y, Chen D, Chen J, Wang X, Wang C, Bu J (2021) Distilling holistic knowledge with graph neural networks. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 10387–10396 Zhou S, Wang Y, Chen D, Chen J, Wang X, Wang C, Bu J (2021) Distilling holistic knowledge with graph neural networks. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 10387–10396
Zurück zum Zitat Zhou S, Wu Y, Ni Z, Zhou X, Wen H, Zou Y (2016) DoReFa-Net: training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint. arXiv:2004.10934 Zhou S, Wu Y, Ni Z, Zhou X, Wen H, Zou Y (2016) DoReFa-Net: training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint. arXiv:​2004.​10934
Zurück zum Zitat Zhu X, Lyu S, Wang X, Zhao Q (2021a) TPH-YOLOv5: improved yolov5 based on transformer prediction head for object detection on drone-captured scenarios. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 2778–2788 Zhu X, Lyu S, Wang X, Zhao Q (2021a) TPH-YOLOv5: improved yolov5 based on transformer prediction head for object detection on drone-captured scenarios. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 2778–2788
Zurück zum Zitat Zhu J, Tang S, Chen D, Yu S, Liu Y, Rong M, Yang A, Wang X (2021b) Complementary relation contrastive distillation. In: Proceedings of the IEEE/CVF international conference on computer visionand pattern recognition, pp 9260–9269 Zhu J, Tang S, Chen D, Yu S, Liu Y, Rong M, Yang A, Wang X (2021b) Complementary relation contrastive distillation. In: Proceedings of the IEEE/CVF international conference on computer visionand pattern recognition, pp 9260–9269
Zurück zum Zitat Zoph B, Vasudevan V, Shlens J, Le QV (2018) Learning transferable architectures for scalable image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8697–8710 Zoph B, Vasudevan V, Shlens J, Le QV (2018) Learning transferable architectures for scalable image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8697–8710
Metadaten
Titel
A review of convolutional neural network architectures and their optimizations
verfasst von
Shuang Cong
Yang Zhou
Publikationsdatum
22.06.2022
Verlag
Springer Netherlands
Erschienen in
Artificial Intelligence Review / Ausgabe 3/2023
Print ISSN: 0269-2821
Elektronische ISSN: 1573-7462
DOI
https://doi.org/10.1007/s10462-022-10213-5

Weitere Artikel der Ausgabe 3/2023

Artificial Intelligence Review 3/2023 Zur Ausgabe

Premium Partner