skip to main content
10.1145/2897937.2898092acmotherconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article

Simplifying deep neural networks for neuromorphic architectures

Published:05 June 2016Publication History

ABSTRACT

Deep learning using deep neural networks is taking machine intelligence to the next level in computer vision, speech recognition, natural language processing, etc. Brain-like hardware platforms for the brain-inspired computational models are being studied, but none of such platforms deals with the huge size of practical deep neural networks. This paper presents two techniques, factorization and pruning, that not only compress the models but also maintain the form of the models for the execution on neuromorphic architectures. We also propose a novel method to combine the two techniques. The proposed method shows significant improvements in reducing the number of model parameters over standalone use of each method while maintaining the performance. Our experimental results show that the proposed method can achieve 31× reduction rate without loss of accuracy for the largest layer of AlexNet.

References

  1. Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. Deep learning. Nature, 521(7553):436--444, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  2. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pages 1097--1105, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Ouais Alsharif and Joelle Pineau. End-to-end text recognition with hybrid hmm maxout models. arXiv preprint arXiv:1310.1811, 2013.Google ScholarGoogle Scholar
  4. Ali Sharif Razavian, Hossein Azizpour, Josephine Sullivan, and Stefan Carlsson. Cnn features off-the-shelf: an astounding baseline for recognition. In Proc. IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 512--519, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Yaniv Taigman, Ming Yang, Marc'Aurelio Ranzato, and Lars Wolf. Deepface: Closing the gap to human-level performance in face verification. In Proc. Conf. on Computer Vision and Pattern Recognition, pages 1701--1708, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Pierre Sermanet, David Eigen, Xiang Zhang, Michaël Mathieu, Rob Fergus, and Yann LeCun. Overfeat: Integrated recognition, localization and detection using convolutional networks. arXiv preprint arXiv:1312.6229, 2013.Google ScholarGoogle Scholar
  7. Clément Farabet, Berin Martini, Polina Akselrod, Selçuk Talay, Yann LeCun, and Eugenio Culurciello. Hardware accelerated convolutional neural networks for synthetic vision systems. In Proc. IEEE Int. Symp. on Circuits and Systems, pages 257--260, 2010.Google ScholarGoogle ScholarCross RefCross Ref
  8. Srimat Chakradhar, Murugan Sankaradas, Venkata Jakkula, and Srihari Cadambi. A dynamically configurable coprocessor for convolutional neural networks. In ACM SIGARCH Computer Architecture News, volume 38, pages 247--257, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Andrew S Cassidy et al. Real-time scalable cortical computing at 46 giga-synaptic ops/watt with ~100x speed up in time-to-solution and ~100,000x reduction in energy-to-solution. In Proc. the International Conference for High Performance Computing, Networking, Storage and Analysis, pages 27--38, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Ben Varkey Benjamin et al. Neurogrid: A mixed-analog-digital multichip system for large-scale neural simulations. Proceedings of the IEEE, 102(5):699--716, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  11. Johannes Schemmel, D Bruderle, A Grubl, Matthias Hock, Karlheinz Meier, and Sebastian Millner. A wafer-scale neuromorphic hardware system for large-scale neural modeling. In Proc. IEEE Int. Symp. on Circuits and Systems, pages 1947--1950, 2010.Google ScholarGoogle ScholarCross RefCross Ref
  12. John V Arthur et al. Building block of a programmable neuromorphic substrate: A digital neurosynaptic core. In Proc. Int. Joint Conf. on Neural Networks, pages 1--8, 2012.Google ScholarGoogle Scholar
  13. Steven K Esser et al. Cognitive computing systems: Algorithms and applications for networks of neurosynaptic cores. In Proc. Int. Joint Conf. on Neural Networks, pages 1--10, 2013.Google ScholarGoogle Scholar
  14. Arnon Amir et al. Cognitive computing programming paradigm: a corelet language for composing networks of neurosynaptic cores. In Proc. Int. Joint Conf. on Neural Networks, pages 1--10, 2013.Google ScholarGoogle Scholar
  15. Andrew S. Cassidy et al. Cognitive computing building block: A versatile and efficient digital neuron model for neurosynaptic cores. In Proc. Int. Joint Conf. on Neural Networks, pages 1--10, 2013.Google ScholarGoogle Scholar
  16. Greg Snider. Molecular-junction-nanowire crossbar-based neural network. U.S. Patent 20040150010.Google ScholarGoogle Scholar
  17. Wenlin Chen, James T Wilson, Stephen Tyree, Kilian Q Weinberger, and Yixin Chen. Compressing neural networks with the hashing trick. arXiv preprint arXiv:1504.04788, 2015.Google ScholarGoogle Scholar
  18. John A Hertz, Anders S Krogh, and Richard G Palmer. Introduction to the theory of neural computation. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Yann LeCun, John S Denker, Sara A Solla, Richard E Howard, and Lawrence D Jackel. Optimal brain damage. In Advances in Neural Information Processing Systems, pages 598--605, 1990. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Babak Hassibi, David G Stork, and Gregory J Wolff. Optimal brain surgeon and general network pruning. In Proc. Int. Conf. on Neural Networks, pages 293--299, 1993.Google ScholarGoogle ScholarCross RefCross Ref
  21. Tianxing He, Yuchen Fan, Yanmin Qian, Tian Tan, and Kai Yu. Reshaping deep neural network for fast decoding by node-pruning. In Proc. IEEE Int. Conf. on Acoustics, Speech and Signal Processing, pages 245--249, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  22. Tara N Sainath, Brian Kingsbury, Vikas Sindhwani, Ebru Arisoy, and Bhuvana Ramabhadran. Low-rank matrix factorization for deep neural network training with high-dimensional output targets. In Proc. IEEE Int. Conf. on Acoustics, Speech and Signal Processing, pages 6655--6659, 2013.Google ScholarGoogle ScholarCross RefCross Ref
  23. Jian Xue, Jinyu Li, and Yifan Gong. Restructuring of deep neural network acoustic models with singular value decomposition. In INTERSPEECH, pages 2365--2369, 2013.Google ScholarGoogle Scholar
  24. Emily L Denton, Wojciech Zaremba, Joan Bruna, Yann LeCun, and Rob Fergus. Exploiting linear structure within convolutional networks for efficient evaluation. In Advances in Neural Information Processing Systems, pages 1269--1277, 2014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Ian J. Goodfellow, David Warde-Farley, Pascal Lamblin, Vincent Dumoulin, Mehdi Mirza, Razvan Pascanu, James Bergstra, Frédéric Bastien, and Yoshua Bengio. Pylearn2: a machine learning research library. arXiv preprint arXiv:1308.4214, 2013.Google ScholarGoogle Scholar
  26. Weiguang Ding, Ruoyan Wang, Fei Mao, and Graham Taylor. Theano-based large-scale visual recognition with multiple gpus. arXiv preprint arXiv:1412.2302, 2014.Google ScholarGoogle Scholar
  27. Song Han, Jeff Pool, John Tran, and William J Dally. Learning both weights and connections for efficient neural networks. arXiv preprint arXiv:1506.02626, 2015.Google ScholarGoogle Scholar
  28. Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.Google ScholarGoogle Scholar
  29. Wenlin Chen, James T Wilson, Stephen Tyree, Kilian Q Weinberger, and Yixin Chen. Compressing neural networks with the hashing trick. arXiv preprint arXiv:1504.04788, 2015.Google ScholarGoogle Scholar
  30. Jaeyong Chung, Taehwan Shin, and Yongshin Kang. Insight: A neuromorphic computing system for evaluation of large neural networks. arXiv preprint arXiv:1508.01008, 2015.Google ScholarGoogle Scholar
  31. Taehwan Shin, Yongshin Kang, Seungho Yang, Seban Kim, and Jaeyong Chung. Live demonstration: Real-time image classification on a neuromorphic computing system with zero off-chip memory access. In Proc. IEEE Int. Symp. on Circuits and Systems, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Simplifying deep neural networks for neuromorphic architectures

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          DAC '16: Proceedings of the 53rd Annual Design Automation Conference
          June 2016
          1048 pages
          ISBN:9781450342360
          DOI:10.1145/2897937

          Copyright © 2016 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 5 June 2016

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          Overall Acceptance Rate1,770of5,499submissions,32%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader