Skip to main content
Top
Published in: Neural Processing Letters 1/2018

20-11-2017

Deep Learning with Darwin: Evolutionary Synthesis of Deep Neural Networks

Authors: Mohammad Javad Shafiee, Akshaya Mishra, Alexander Wong

Published in: Neural Processing Letters | Issue 1/2018

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Taking inspiration from biological evolution, we explore the idea of “Can deep neural networks evolve naturally over successive generations into highly efficient deep neural networks?” by introducing the notion of synthesizing new highly efficient, yet powerful deep neural networks over successive generations via an evolutionary process from ancestor deep neural networks. The architectural traits of ancestor deep neural networks are encoded using synaptic probability models, which can be viewed as the ‘DNA’ of these networks. New descendant networks with differing network architectures are synthesized based on these synaptic probability models from the ancestor networks and computational environmental factor models, in a random manner to mimic heredity, natural selection, and random mutation. These offspring networks are then trained into fully functional networks, like one would train a newborn, and have more efficient, more diverse network architectures than their ancestor networks, while achieving powerful modeling capabilities. Experimental results for the task of visual saliency demonstrated that the synthesized ‘evolved’ offspring networks can achieve state-of-the-art performance while having network architectures that are significantly more efficient (with a staggering \(\sim \) 48-fold decrease in synapses by the fourth generation) compared to the original ancestor network.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
2.
go back to reference Graves A, Mohamed A-R, Hinton G (2013) Speech recognition with deep recurrent neural networks. In: IEEE international conference on acoustics, speech and signal processing, pp 6645–6649 Graves A, Mohamed A-R, Hinton G (2013) Speech recognition with deep recurrent neural networks. In: IEEE international conference on acoustics, speech and signal processing, pp 6645–6649
3.
4.
go back to reference Tompson J, Jain A, LeCun Y, Bregler C (2014) Joint training of a convolutional network and a graphicalmodel for human pose estimation. In: Proceedings of advances in neural information processing systems (NIPS), pp 1799–1807 Tompson J, Jain A, LeCun Y, Bregler C (2014) Joint training of a convolutional network and a graphicalmodel for human pose estimation. In: Proceedings of advances in neural information processing systems (NIPS), pp 1799–1807
5.
go back to reference Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: Proceedings of advances in neural information processing systems (NIPS), pp 1097–1105 Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: Proceedings of advances in neural information processing systems (NIPS), pp 1097–1105
6.
go back to reference Farabet C, Couprie C, Najman L, LeCun Y (2013) Learning hierarchical features for scene labeling. In: IEEE transactions on pattern analysis and machine intelligence (TPAMI) Farabet C, Couprie C, Najman L, LeCun Y (2013) Learning hierarchical features for scene labeling. In: IEEE transactions on pattern analysis and machine intelligence (TPAMI)
7.
go back to reference Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. CoRR, abs/1409.1556 Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. CoRR, abs/1409.1556
8.
go back to reference Hinton G, Deng L, Yu D, Dahl GE, Mohamed A-R, Jaitly N, Senior A, Vanhoucke V, Nguyen P, Sainath TN et al (2012) Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. In: IEEE signal processing magazine Hinton G, Deng L, Yu D, Dahl GE, Mohamed A-R, Jaitly N, Senior A, Vanhoucke V, Nguyen P, Sainath TN et al (2012) Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. In: IEEE signal processing magazine
9.
go back to reference Hannun A, Case C, Casper J, Catanzaro B, Diamos G, Elsen E, Prenger R, Satheesh S, Sengupta S, Coates A et al (2014) Deep speech: scaling up end-to-end speech recognition. CoRR, abs/1412.5567 Hannun A, Case C, Casper J, Catanzaro B, Diamos G, Elsen E, Prenger R, Satheesh S, Sengupta S, Coates A et al (2014) Deep speech: scaling up end-to-end speech recognition. CoRR, abs/1412.5567
10.
go back to reference Amodei D, Anubhai R, Battenberg E, Case C, Casper J, Catanzaro B, Chen J, Chrzanowski M, Coates A, Diamos G et al (2015) Deep speech 2: end-to-end speech recognition in English and Mandarin. CoRR, abs/1512.02595 Amodei D, Anubhai R, Battenberg E, Case C, Casper J, Catanzaro B, Chen J, Chrzanowski M, Coates A, Diamos G et al (2015) Deep speech 2: end-to-end speech recognition in English and Mandarin. CoRR, abs/1512.02595
11.
go back to reference Srivastava RK, Greff K, Schmidhuber J (2015) Training very deep networks. In: Proceedings of advances in neural information processing systems (NIPS), pp 2377–2385 Srivastava RK, Greff K, Schmidhuber J (2015) Training very deep networks. In: Proceedings of advances in neural information processing systems (NIPS), pp 2377–2385
12.
go back to reference Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 1–9 Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 1–9
13.
go back to reference LeCun Y, Denker JS, Solla SA, Howard RE, Jackel LD (1989) Optimal brain damage. In: Advances in neural information processing systems (NIPS) LeCun Y, Denker JS, Solla SA, Howard RE, Jackel LD (1989) Optimal brain damage. In: Advances in neural information processing systems (NIPS)
14.
go back to reference Gong Y, Liu L, Yang M, Bourdev L (2014) Compressing deep convolutional networks using vector quantization. CoRR, abs/1412.6115 Gong Y, Liu L, Yang M, Bourdev L (2014) Compressing deep convolutional networks using vector quantization. CoRR, abs/1412.6115
15.
go back to reference Han S, Mao H, Dally WJ (2015) Deep compression: compressing deep neural networks with pruning, trained quantization and huffman coding. CoRR, abs/1510.00149 Han S, Mao H, Dally WJ (2015) Deep compression: compressing deep neural networks with pruning, trained quantization and huffman coding. CoRR, abs/1510.00149
16.
go back to reference Han S, Pool J, Tran J, Dally W (2015) Learning both weights and connections for efficient neural network. In: Advances in neural information processing systems (NIPS) Han S, Pool J, Tran J, Dally W (2015) Learning both weights and connections for efficient neural network. In: Advances in neural information processing systems (NIPS)
17.
go back to reference Chen W, Wilson JT, Tyree S, Weinberger KQ, Chen Y (2015) Compressing neural networks with the hashing trick. CoRR, abs/1504.04788 Chen W, Wilson JT, Tyree S, Weinberger KQ, Chen Y (2015) Compressing neural networks with the hashing trick. CoRR, abs/1504.04788
18.
go back to reference Moran D, Softley R, Warrant EJ (2015) The energetic cost of vision and the evolution of eyeless Mexican cavefish. Sci Adv 1:e1500363CrossRef Moran D, Softley R, Warrant EJ (2015) The energetic cost of vision and the evolution of eyeless Mexican cavefish. Sci Adv 1:e1500363CrossRef
19.
go back to reference Peter GMS, Angeline J, Pollack JB (1994) An evolutionary algorithm that constructs recurrent neural networks. In: IEEE transactions on neural networks Peter GMS, Angeline J, Pollack JB (1994) An evolutionary algorithm that constructs recurrent neural networks. In: IEEE transactions on neural networks
20.
go back to reference Stanley KO, Bryant BD, Miikkulainen R (2005) Real-time neuroevolution in the NERO video game. In: IEEE transactions on evolutionary computation Stanley KO, Bryant BD, Miikkulainen R (2005) Real-time neuroevolution in the NERO video game. In: IEEE transactions on evolutionary computation
21.
go back to reference Stanley KO, Miikkulainen R (2002) Evolving neural networks through augmenting topologies. Evol Comput 10:99–127CrossRef Stanley KO, Miikkulainen R (2002) Evolving neural networks through augmenting topologies. Evol Comput 10:99–127CrossRef
22.
go back to reference Gauci J, Stanley KO (2007) Generating large-scale neural networks through discovering geometric regularities. In: Proceedings of the 9th annual conference on Genetic and evolutionary computation. pp 997–1004 Gauci J, Stanley KO (2007) Generating large-scale neural networks through discovering geometric regularities. In: Proceedings of the 9th annual conference on Genetic and evolutionary computation. pp 997–1004
23.
go back to reference Tirumala SS, Ali S, Ramesh CP (2016) Evolving deep neural networks: a new prospect. In: 12th International conference on natural computation, Fuzzy systems and knowledge discovery (ICNC-FSKD). pp 69-74 Tirumala SS, Ali S, Ramesh CP (2016) Evolving deep neural networks: a new prospect. In: 12th International conference on natural computation, Fuzzy systems and knowledge discovery (ICNC-FSKD). pp 69-74
24.
go back to reference Liu T, Yuan Z, Sun J, Wang J, Zheng N, Tang X, Shum H-Y (2011) Learning to detect a salient object. In: IEEE transactions on pattern analysis and machine intelligence (TPAMI). pp 353–367 Liu T, Yuan Z, Sun J, Wang J, Zheng N, Tang X, Shum H-Y (2011) Learning to detect a salient object. In: IEEE transactions on pattern analysis and machine intelligence (TPAMI). pp 353–367
25.
go back to reference Li G, Yu Y (2015) Visual saliency based on multiscale deep features. In: IEEE conference on computer vision and pattern recognition (CVPR) Li G, Yu Y (2015) Visual saliency based on multiscale deep features. In: IEEE conference on computer vision and pattern recognition (CVPR)
26.
go back to reference Nitish S, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res JMLR 15:1929–1958MathSciNetMATH Nitish S, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res JMLR 15:1929–1958MathSciNetMATH
27.
go back to reference Wan L, Zeiler M, Zhang S, LeCun Y, Fergus R (2013) Regularization of neural networks using dropconnect. In: International conference on machine learning (ICML) Wan L, Zeiler M, Zhang S, LeCun Y, Fergus R (2013) Regularization of neural networks using dropconnect. In: International conference on machine learning (ICML)
28.
go back to reference Ioannou Y, Robertson D, Shotton J, Cipolla R, Criminisi A (2015) Training CNNS with low-rank filters for efficient image classification. arXiv preprint arXiv:1511.06744 Ioannou Y, Robertson D, Shotton J, Cipolla R, Criminisi A (2015) Training CNNS with low-rank filters for efficient image classification. arXiv preprint arXiv:​1511.​06744
29.
go back to reference Jaderberg M, Vedaldi A, Zisserman A (2014) Speeding up convolutional neural networks with low rank expansions. arXiv preprint arXiv:1405.3866, Jaderberg M, Vedaldi A, Zisserman A (2014) Speeding up convolutional neural networks with low rank expansions. arXiv preprint arXiv:​1405.​3866,
30.
go back to reference Denton E, Zaremba W, Bruna J, LeCun Y, Fergus R (2014) Exploiting linear structure within convolutional networks for efficient evaluation. In: Proceedings of advances in neural information processing systems (NIPS). pp 1269–1277 Denton E, Zaremba W, Bruna J, LeCun Y, Fergus R (2014) Exploiting linear structure within convolutional networks for efficient evaluation. In: Proceedings of advances in neural information processing systems (NIPS). pp 1269–1277
31.
go back to reference Feng J, Darrell T (2015) Learning the structure of deep convolutional networks. In: Proceedings of the IEEE international conference on computer vision, pp 2749–2757 Feng J, Darrell T (2015) Learning the structure of deep convolutional networks. In: Proceedings of the IEEE international conference on computer vision, pp 2749–2757
32.
go back to reference Liu B, Wang M, Foroosh H, Tappen M, Pensky M (2015) Sparse convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 806–814 Liu B, Wang M, Foroosh H, Tappen M, Pensky M (2015) Sparse convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 806–814
33.
Metadata
Title
Deep Learning with Darwin: Evolutionary Synthesis of Deep Neural Networks
Authors
Mohammad Javad Shafiee
Akshaya Mishra
Alexander Wong
Publication date
20-11-2017
Publisher
Springer US
Published in
Neural Processing Letters / Issue 1/2018
Print ISSN: 1370-4621
Electronic ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-017-9733-0

Other articles of this Issue 1/2018

Neural Processing Letters 1/2018 Go to the issue