skip to main content
10.1145/3061639.3062326acmconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article

TIME: A Training-in-memory Architecture for Memristor-based Deep Neural Networks

Authors Info & Claims
Published:18 June 2017Publication History

ABSTRACT

The training of neural network (NN) is usually time-consuming and resource intensive. Memristor has shown its potential in computation of NN. Especially for the metal-oxide resistive random access memory (RRAM), its crossbar structure and multi-bit characteristic can perform the matrix-vector product in high precision, which is the most common operation of NN. However, there exist two challenges on realizing the training of NN. Firstly, the current architecture can only support the inference phase of training and cannot perform the backpropagation (BP), the weights update of NN. Secondly, the training of NN requires enormous iterations and constantly updates the weights to reach the convergence, which leads to large energy consumption because of lots of write and read operations. In this work, we propose a novel architecture, TIME, and peripheral circuit designs to enable the training of NN in RRAM. TIME supports the BP and the weights update while maximizing the reuse of peripheral circuits for the inference operation on RRAM. Meanwhile, a variability-free tuning scheme and gradually-write circuits are designed to reduce the cost of tuning RRAM. We explore the performance of both SL (supervised learning) and DRL (deep reinforcement learning) in TIME, and a specific mapping method of DRL is also introduced to further improve the energy efficiency. Experimental results show that, in SL, TIME can achieve 5.3x higher energy efficiency on average compared with the most powerful application-specific integrated circuits (ASIC) in the literature. In DRL, TIME can perform averagely 126x higher than GPU in energy efficiency. If the cost of tuning RRAM can be further reduced, TIME have the potential of boosting the energy efficiency by 2 orders of magnitude compared with ASIC.

References

  1. Volodymyr Mnih et al. Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602, 2013.Google ScholarGoogle Scholar
  2. Kaiming He et al. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, pages 1026--1034, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Stephen W. Keckler et al. GPUs and the Future of Parallel Computing. IEEE Micro, 31(5):7--17, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Yunji Chen et al. Dadiannao: A machine-learning supercomputer. In IEEE/ACM ISM, pages 609--622, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Boxun Li et al. Training itself: Mixed-signal training acceleration for memristor-based neural network. In ASPDAC, pages 361--366, 2014.Google ScholarGoogle Scholar
  6. Ping Chi et al. PRIME: A Novel Processing-in-Memory Architecture for Neural Network Computation in ReRAM-Based Main Memory. In ISCA, pages 27--39, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Ragib Hasan et al. Enabling back propagation training of memristor crossbar neuromorphic processors. In IJCNN, pages 21--28. IEEE, 2014.Google ScholarGoogle Scholar
  8. Lixue Xia et al. Switched by input: power efficient structure for rram-based convolutional neural network. In DAC, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. LeCun et al. The mnist database of handwritten digits, 1998.Google ScholarGoogle Scholar
  10. D Soudry et al. Memristor-based multilayer neural networks with online gradient descent training. IEEE TNNLS, 26(10):1, 2015.Google ScholarGoogle Scholar
  11. Boxun Li et al. RRAM-Based Analog Approximate Computing. TCAD, 34(12):1--1, 2015.Google ScholarGoogle Scholar
  12. Shimeng Yu et al. A neuromorphic visual system using RRAM synaptic devices with Sub-pJ energy and tolerance to variability: Experimental characterization and large-scale modeling. IEDM, 2012:10.4.1--10.4.4, 2012.Google ScholarGoogle Scholar
  13. Y. Lecun. A theoretical framework for back-propagation. In Artificial Neural Networks: concepts and theory, 1992.Google ScholarGoogle Scholar
  14. G. W. Burr et al. Large-scale neural networks implemented with non-volatile memory as the synaptic weight element: Comparative performance analysis (accuracy, speed, and power). In IEDM, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  15. Shuchang Zhou et al. Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160, 2016.Google ScholarGoogle Scholar
  16. Cong Xu et al. Understanding the trade-offs in multi-level cell reram memory design. In DAC, pages 1--6, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Lixue Xia et al. MNSIM: Simulation platform for memristor-based neuromorphic computing system. In DATE, pages 469--474, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. H-S Philip Wong et al. Recent progress of phase change memory (PCM) and resistive switching random access memory (RRAM). In IEEE IMW, pages 1--5. IEEE, 2011.Google ScholarGoogle Scholar
  19. Xiangyu Dong et al. Nvsim: A circuit-level performance, energy, and area model for emerging non-volatile memory. In Emerging Memory Technologies, pages 15--50. Springer, 2014.Google ScholarGoogle ScholarCross RefCross Ref

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    DAC '17: Proceedings of the 54th Annual Design Automation Conference 2017
    June 2017
    533 pages
    ISBN:9781450349277
    DOI:10.1145/3061639

    Copyright © 2017 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 18 June 2017

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate1,770of5,499submissions,32%

    Upcoming Conference

    DAC '24
    61st ACM/IEEE Design Automation Conference
    June 23 - 27, 2024
    San Francisco , CA , USA

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader