Skip to main content
Top

2020 | OriginalPaper | Chapter

Neuron Fault Tolerance Capability Based Computation Reuse in DNNs

Authors : Pengnian Qi, Jing Wang, Xiaoyan Zhu, Weigong Zhang

Published in: Algorithms and Architectures for Parallel Processing

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

For applications of speech and video, the consecutive inputs exhibit a high degree of similarity, hence, some results of previous execution can be reused. The technique of quantization can efficiently increase the similarity of consecutive inputs. However, when using quantization, the smaller the number of quantization bits the higher the similarity, as the inputs are constrained to a smaller set of values, but the larger the accuracy loss since input errors are increased. Therefore, we observe that existing reuse schema just applied unique the number of quantization bits in the entire network. If the number of quantization bits is too long, it will directly reduce the similarity between the inputs and thus reduce the reuse ratio. Hence, it is important that exploits the tradeoff among the number of quantization bits, reuse rate, and accuracy. There is an opportunity to significantly improve the performance and efficiency of DNN execution by use multiple quantization bits simultaneously according to the technique of neuron criticality analysis. To do so, we propose a novel reuse schema called Mquans based on neuron criticality analysis without accuracy loss. And evaluation results show that our proposed design achieves 2.7 speedups and 38% energy saving on average over the baseline.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Riera, M., Arnau, J.-M., Gonzalez, A.: Computation reuse in DNNs by exploiting input similarity. In: ISCA (2018) Riera, M., Arnau, J.-M., Gonzalez, A.: Computation reuse in DNNs by exploiting input similarity. In: ISCA (2018)
2.
go back to reference Zhou, S., Wu, Y., Ni, Z., et al.: DoReFa-net: training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160 (2016) Zhou, S., Wu, Y., Ni, Z., et al.: DoReFa-net: training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:​1606.​06160 (2016)
3.
go back to reference Zhang, Q., Wang, T., Tian, Y., Yuan, F., Xu, Q.: ApproxANN: an approximate computing framework for artificial neural network. In: DATE (2015) Zhang, Q., Wang, T., Tian, Y., Yuan, F., Xu, Q.: ApproxANN: an approximate computing framework for artificial neural network. In: DATE (2015)
4.
go back to reference Weaver, V.M., et al.: Measuring energy and power with PAPI. In: ICCP (2012) Weaver, V.M., et al.: Measuring energy and power with PAPI. In: ICCP (2012)
5.
go back to reference Farabet, C., Martini, B., Corda, B., Akselrod, P., Culurciello, E., LeCun, Y.: NeuFlow: a runtime reconfigurable dataflow processor for vision. In: CVPRW (2011) Farabet, C., Martini, B., Corda, B., Akselrod, P., Culurciello, E., LeCun, Y.: NeuFlow: a runtime reconfigurable dataflow processor for vision. In: CVPRW (2011)
6.
go back to reference Chen, T., et al.: DianNao: a small-footprint High-throughput accelerator for ubiquitous machine-learning. In: ASPLOS (2014) Chen, T., et al.: DianNao: a small-footprint High-throughput accelerator for ubiquitous machine-learning. In: ASPLOS (2014)
7.
go back to reference Chen, Y., et al.: Dadiannao: a machine-learning supercomputer. In: MICRO (2014) Chen, Y., et al.: Dadiannao: a machine-learning supercomputer. In: MICRO (2014)
8.
go back to reference Hegde, K., Yu, J., Agrawal, R., et al.: UCNN: exploiting computational reuse in deep neural networks via weight repetition. In: ISCA (2018) Hegde, K., Yu, J., Agrawal, R., et al.: UCNN: exploiting computational reuse in deep neural networks via weight repetition. In: ISCA (2018)
9.
go back to reference Fuchs, A., Wentzlaff, D.: Scaling datacenter accelerators with compute-reuse architectures. In: ISCA (2018) Fuchs, A., Wentzlaff, D.: Scaling datacenter accelerators with compute-reuse architectures. In: ISCA (2018)
10.
go back to reference Chen, Y.-H., Emer, J., Sze, V.: Eyeriss: a spatial architecture for energy-efficient dataflow for convolutional neural networks. In ISCA (2016) Chen, Y.-H., Emer, J., Sze, V.: Eyeriss: a spatial architecture for energy-efficient dataflow for convolutional neural networks. In ISCA (2016)
Metadata
Title
Neuron Fault Tolerance Capability Based Computation Reuse in DNNs
Authors
Pengnian Qi
Jing Wang
Xiaoyan Zhu
Weigong Zhang
Copyright Year
2020
DOI
https://doi.org/10.1007/978-3-030-38961-1_23

Premium Partner