skip to main content
10.1145/2897937.2898003acmotherconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article

DeepBurning: automatic generation of FPGA-based learning accelerators for the neural network family

Authors Info & Claims
Published:05 June 2016Publication History

ABSTRACT

Recent advances in Neural Networks (NN) are enabling more and more innovative applications. As an energy-efficient hardware solution, machine learning accelerators for CNNs or traditional ANNs are also gaining popularity in the area of embedded vision, robotics and cyberphysics. However, the design parameters of NN models vary significantly from application to application. Hence, it's hard to provide one general and highly-efficient hardware solution to accommodate all of them, and it is also impractical for the domain-specific developers to customize their flown hardware targeting on a specific NN model. To deal with this dilemma, this study proposes a design automation tool, DeepBurning, allowing the application developers to build from scratch learning accelerators that targets their specific NN models with custom configurations and optimized performance. DeepBurning includes a RTL-level accelerator generator and a coordinated compiler that generates the control flow and data layout under the user-specified constraints. The results can be used to implement FPGA-based NN accelerator or help generate chip design for early design stage. In general, DeepBurning supports a large family of NN models, and greatly simplifies the design flow of NN accelerators for the machine learning or AI application developers. The evaluation shows that the generated learning accelerators burnt to our FPGA board exhibit great power efficiency compared to state-of-the-art FPGA-based solutions.

References

  1. Esmaeilzadeh et al., "Neural acceleration for general-purposeapproximate programs,". In Proc. MICRO, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. G. Hinton et al., "Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups," in IEEE Signal Processing Magazine, vol. 29, no. 6, pp. 82--97, 2012.Google ScholarGoogle ScholarCross RefCross Ref
  3. Y Jia et al., "Caffe: Convolutional architecture for fast feature embedding," in ACM Proc. Multimedia, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. T. Chen et al., "DianNao: a small-footprint high-throughput accelerator for ubiquitous machine-learning," in Proc. ASPLOS, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. A. Krizhevsky et al., "Imagenet classification with deep convolutional neural networks," in Proc. NIPS, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. M. Peemen et al., "Memory-centric accelerator design for convolutional neural networks," in Proc. ICCD, 2013.Google ScholarGoogle Scholar
  7. C. Zhang et al., "Optimizing fpga-based accelerator design for deep convolutional neural networks," in Proc. FPGA, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. W. Ouyang et al,. "DeepID-Net: Deformable Deep Convolutional Neural Networks for Object Detection," in Proc. CVPR, 2015.Google ScholarGoogle Scholar
  9. Y. Pan et. al., "Jointly Modeling Embedding and Translation to Bridge Video and Language," in arXiv, 2015.Google ScholarGoogle Scholar
  10. V. Vapnik, "Does Deep Learning Come from the Devil?," Yandex conference on machine learning prospects and applications, Berlin, 2015Google ScholarGoogle Scholar
  11. R. Beigel et al., "Sorting n Objects With a k-Sorter," IEEE Trans, on Computers, 1990. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. DeepBurning: automatic generation of FPGA-based learning accelerators for the neural network family

          Recommendations

          Reviews

          Stewart Mark Godwin

          Technically complex, this paper has numerous acronyms that are commonly used in specialist areas like electronics and engineering. However, the topic of DeepBurning can be summarized as a design automation tool that allows application developers the option to build a learning accelerator for specific neural networks. This process uses field-programmable gate arrays (FPGAs) that are designed to be modified and configured to suit problems in areas like machine learning and artificial learning. In the paper, the DeepBurning framework is evaluated using eight neural networks, with comparisons of performance, power consumption, and accuracy. In conclusion, the authors indicate they have proved that DeepBurning enables an instant generation of hardware and software solutions for specific neural networks. This paper and the general topic are not for the layperson and would only be of interest to industry-specific experts and academics within this field.

          Access critical reviews of Computing literature here

          Become a reviewer for Computing Reviews.

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Other conferences
            DAC '16: Proceedings of the 53rd Annual Design Automation Conference
            June 2016
            1048 pages
            ISBN:9781450342360
            DOI:10.1145/2897937

            Copyright © 2016 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 5 June 2016

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            Overall Acceptance Rate1,770of5,499submissions,32%

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader