Skip to main content

2019 | OriginalPaper | Buchkapitel

On Construction of a Caffe Deep Learning Framework based on Intel Xeon Phi

verfasst von : Chao-Tung Yang, Jung-Chun Liu, Yu-Wei Chan, Endah Kristiani, Chan-Fu Kuo

Erschienen in: Advances on P2P, Parallel, Grid, Cloud and Internet Computing

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

With the increase of processor computing power, also a substantial rise in the development of many scientific applications, such as weather forecast, financial market analysis, medical technology and so on. The need for more intelligent data increases significantly. Deep Learning as a framework that able to understand the abstract information such as images, text, and sound has a challenging area in recent research works. This phenomenon makes the accuracy and speed are essential for implementing a large neural network. Therefore in this paper, we intend to implement Caffe deep learning framework on Intel Xeon Phi and measure the performance of this environment. In this case, we conduct three experiments. First, we evaluated the accuracy of Caffe deep learning framework in several numbers of iterations on Intel Xeon Phi. For the speed evaluation, in the second experiment we compared the training time before and after optimization on Intel Xeon E5-2650 and Intel Xeon Phi 7210 . In this case, we use vectorization, OpenMP parallel processing, message transfer Interface (MPI) for optimization. In the third experiment, we compared multinode execution results on two nodes of Intel Xeon E5-2650 and two nodes of Intel Xeon Phi 7210.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Heinecke, A.: Accelerators in scientific computing is it worth the effort? In: 2013 International Conference on High Performance Computing Simulation (HPCS), pp. 504–504, July 2013 Heinecke, A.: Accelerators in scientific computing is it worth the effort? In: 2013 International Conference on High Performance Computing Simulation (HPCS), pp. 504–504, July 2013
2.
Zurück zum Zitat Dagum, Leonardo, Menon, Ramesh: Openmp: an industry standard api for shared-memory programming. IEEE Comput. Sci. Eng. 5(1), 46–55 (1998)CrossRef Dagum, Leonardo, Menon, Ramesh: Openmp: an industry standard api for shared-memory programming. IEEE Comput. Sci. Eng. 5(1), 46–55 (1998)CrossRef
3.
Zurück zum Zitat Chapman, B., Jost, G., Van Der Pas, R.: Using OpenMP: portable shared memory parallel programming, vol. 10. MIT press, Cambridge (2008) Chapman, B., Jost, G., Van Der Pas, R.: Using OpenMP: portable shared memory parallel programming, vol. 10. MIT press, Cambridge (2008)
4.
Zurück zum Zitat Chandra, R.: Parallel programming in OpenMP. Morgan kaufmann (2001) Chandra, R.: Parallel programming in OpenMP. Morgan kaufmann (2001)
6.
Zurück zum Zitat Zhang, C., Fang, Z., Zhou, P., Pan, P., Cong, J.: Caffeine: towards uniformed representation and acceleration for deep convolutional neural networks. In: IEEE/ACM International Conference on Computer-Aided Design, Digest of Technical Papers, ICCAD, 07–10 Nov 2016 Zhang, C., Fang, Z., Zhou, P., Pan, P., Cong, J.: Caffeine: towards uniformed representation and acceleration for deep convolutional neural networks. In: IEEE/ACM International Conference on Computer-Aided Design, Digest of Technical Papers, ICCAD, 07–10 Nov 2016
7.
Zurück zum Zitat Hegde, G., Siddhartha, Ramasamy, N., Kapre, N.: Caffepresso: an optimized library for deep learning on embedded accelerator-based platforms. In: Proceedings of the International Conference on Compilers, Architectures and Synthesis for Embedded Systems, CASES 2016 (2016) Hegde, G., Siddhartha, Ramasamy, N., Kapre, N.: Caffepresso: an optimized library for deep learning on embedded accelerator-based platforms. In: Proceedings of the International Conference on Compilers, Architectures and Synthesis for Embedded Systems, CASES 2016 (2016)
8.
Zurück zum Zitat Jia, Y., et al.: Caffe: convolutional architecture for fast feature embedding. In: MM 2014 - Proceedings of the 2014 ACM Conference on Multimedia, pp. 675–678 (2014) Jia, Y., et al.: Caffe: convolutional architecture for fast feature embedding. In: MM 2014 - Proceedings of the 2014 ACM Conference on Multimedia, pp. 675–678 (2014)
9.
Zurück zum Zitat Krizhevsky, A., Hinton, G.: Convolutional deep belief networks on cifar-10. Unpublished manuscript 40 (2010) Krizhevsky, A., Hinton, G.: Convolutional deep belief networks on cifar-10. Unpublished manuscript 40 (2010)
10.
Zurück zum Zitat Coates, A., Ng, A., Lee, H.: An analysis of single-layer networks in unsupervised feature learning. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pp. 215–223 (2011) Coates, A., Ng, A., Lee, H.: An analysis of single-layer networks in unsupervised feature learning. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pp. 215–223 (2011)
13.
Zurück zum Zitat Roska, T., et al.: The use of cnn models in the subcortical visual pathway. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 40(3), 182–195 (1993)MathSciNetCrossRef Roska, T., et al.: The use of cnn models in the subcortical visual pathway. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 40(3), 182–195 (1993)MathSciNetCrossRef
14.
Zurück zum Zitat Zarándy, Ákos, Orzó, László, Grawes, Edward, Werblin, Frank: CNN-based models for color vision and visual illusions. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 46(2), 229–238 (1999)CrossRef Zarándy, Ákos, Orzó, László, Grawes, Edward, Werblin, Frank: CNN-based models for color vision and visual illusions. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 46(2), 229–238 (1999)CrossRef
15.
Zurück zum Zitat Bottou, L., et al.: Comparison of classifier methods: a case study in handwritten digit recognition. In: Proceedings of the 12th IAPR International. Conference on Pattern Recognition, 1994. Vol. 2-Conference B: Computer Vision & Image Processing, vol. 2, pp. 77–82. IEEE (1994) Bottou, L., et al.: Comparison of classifier methods: a case study in handwritten digit recognition. In: Proceedings of the 12th IAPR International. Conference on Pattern Recognition, 1994. Vol. 2-Conference B: Computer Vision & Image Processing, vol. 2, pp. 77–82. IEEE (1994)
16.
Zurück zum Zitat Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. In: Advances in Neural Information Processing Systems, pp. 1135–1143 (2015) Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. In: Advances in Neural Information Processing Systems, pp. 1135–1143 (2015)
17.
Zurück zum Zitat Susan Blackford, L., Petitet, Antoine, Pozo, Roldan, Karin Remington, R., Whaley, Clint, Demmel, James, Dongarra, Jack, Duff, Iain, Hammarling, Sven, Henry, Greg: An updated set of basic linear algebra subprograms (blas). ACM Trans. Math. Softw. 28(2), 135–151 (2002)MathSciNetCrossRef Susan Blackford, L., Petitet, Antoine, Pozo, Roldan, Karin Remington, R., Whaley, Clint, Demmel, James, Dongarra, Jack, Duff, Iain, Hammarling, Sven, Henry, Greg: An updated set of basic linear algebra subprograms (blas). ACM Trans. Math. Softw. 28(2), 135–151 (2002)MathSciNetCrossRef
18.
Zurück zum Zitat Nath, R., Tomov, S., Dongarra, J.: Accelerating GPU kernels for dense linear algebra. In: VECPAR, pp. 83–92. Springer, Berlin (2010) Nath, R., Tomov, S., Dongarra, J.: Accelerating GPU kernels for dense linear algebra. In: VECPAR, pp. 83–92. Springer, Berlin (2010)
20.
Zurück zum Zitat Nair, V., Hinton, G.E.: Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning (ICML-10), pp. 807–814 (2010) Nair, V., Hinton, G.E.: Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning (ICML-10), pp. 807–814 (2010)
21.
Zurück zum Zitat Dahl, G.E., Sainath, T.N., Hinton, G.E.: Improving deep neural networks for LVCSR using rectified linear units and dropout. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 8609–8613. IEEE (2013) Dahl, G.E., Sainath, T.N., Hinton, G.E.: Improving deep neural networks for LVCSR using rectified linear units and dropout. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 8609–8613. IEEE (2013)
Metadaten
Titel
On Construction of a Caffe Deep Learning Framework based on Intel Xeon Phi
verfasst von
Chao-Tung Yang
Jung-Chun Liu
Yu-Wei Chan
Endah Kristiani
Chan-Fu Kuo
Copyright-Jahr
2019
DOI
https://doi.org/10.1007/978-3-030-02607-3_9

Neuer Inhalt