Skip to main content
Erschienen in: World Wide Web 2/2018

22.04.2017

Scalable and fast SVM regression using modern hardware

verfasst von: Zeyi Wen, Rui Zhang, Kotagiri Ramamohanarao, Li Yang

Erschienen in: World Wide Web | Ausgabe 2/2018

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Support Vector Machine (SVM) regression is an important technique in data mining. The SVM training is expensive and its cost is dominated by: (i) the kernel value computation, and (ii) a search operation which finds extreme training data points for adjusting the regression function in every training iteration. Existing training algorithms for SVM regression are not scalable to large datasets because: (i) each training iteration repeatedly performs expensive kernel value computations, which is inefficient and requires holding the whole training dataset in memory; (ii) the search operation used in each training iteration considers the whole search space which is very expensive. In this article, we significantly improve the scalability and efficiency of SVM regression by exploiting the high performance of Graphics Processing Units (GPUs) and solid state drives (SSDs). Our key ideas are as follows. (i) To reduce the cost of repeated kernel value computations and avoid holding the whole training dataset in the GPU memory, we precompute all the kernel values and store them in the CPU memory extended by the SSD; together with an efficient strategy to read the precomputed kernel values, reusing precomputed kernel values with an efficient retrieval is much faster than computing them on-the-fly. This also alleviates the restriction that the training dataset has to fit into the GPU memory, and hence makes our algorithm scalable to large datasets, especially for large datasets with very high dimensionality. (ii) To enhance the performance of the frequently used search operation, we design an algorithm that minimizes the search space and the number of accesses to the GPU global memory; this optimized search algorithm also avoids branch divergence (one of the causes for poor performance) among GPU threads to achieve high utilization of the GPU resources. Our proposed techniques together form a scalable solution to the SVM regression which we call SIGMA. Our extensive experimental results show that SIGMA is highly efficient and can handle very large datasets which the state-of-the-art GPU-based algorithm cannot handle. On the datasets of size that the state-of-the-art GPU-based algorithm can handle, SIGMA consistently outperforms the state-of-the-art GPU-based algorithm by an order of magnitude and achieves up to 86 times speedup.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Fußnoten
1
When the context is clear, we omit “SVM” in the rest of this article, similarly for the SVM training
 
2
The datasets are found in LibSVM site and UCI repository.
 
3
To distinguish from the GPU memory, we use “the CPU memory” instead of “main memory” in this article.
 
4
Without confusion, we use “an element” and “an optimality indicator” in the optimality indicator vector interchangeably
 
Literatur
1.
Zurück zum Zitat Athanasopoulos, A., Dimou, A., Mezaris, V., Kompatsiaris, I.: GPU acceleration for support vector machines International Workshop on Image Analysis for Multimedia Interactive Services (2011) Athanasopoulos, A., Dimou, A., Mezaris, V., Kompatsiaris, I.: GPU acceleration for support vector machines International Workshop on Image Analysis for Multimedia Interactive Services (2011)
2.
Zurück zum Zitat Carpenter, A.: Cusvm: A cuda implementation of support vector classification and regression. patternsonscreen. net/cuSVMDesc.pdf (2009) Carpenter, A.: Cusvm: A cuda implementation of support vector classification and regression. patternsonscreen. net/​cuSVMDesc.​pdf (2009)
3.
Zurück zum Zitat Caruana, G., Li, M., Qi, M.: A MapReduce based parallel SVM for large scale spam filtering The International Conference on Fuzzy Systems and Knowledge Discovery, vol. 4, pp 2659–2662 (2011) Caruana, G., Li, M., Qi, M.: A MapReduce based parallel SVM for large scale spam filtering The International Conference on Fuzzy Systems and Knowledge Discovery, vol. 4, pp 2659–2662 (2011)
4.
Zurück zum Zitat Catak, F.O., Balaban, M.E.: CloudSVM: training an SVM classifier in cloud computing systems Pervasive Computing and the Networked World, pp 57–68. Springer-Verlag (2013) Catak, F.O., Balaban, M.E.: CloudSVM: training an SVM classifier in cloud computing systems Pervasive Computing and the Networked World, pp 57–68. Springer-Verlag (2013)
5.
Zurück zum Zitat Catanzaro, B., Sundaram, N., Keutzer, K.: Fast support vector machine training and classification on graphics processors ICML, pp 104–111. ACM (2008) Catanzaro, B., Sundaram, N., Keutzer, K.: Fast support vector machine training and classification on graphics processors ICML, pp 104–111. ACM (2008)
6.
Zurück zum Zitat Chang, C.-C., Lin, C.-J.: LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology (TIST) 2(3), 27 (2011) Chang, C.-C., Lin, C.-J.: LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology (TIST) 2(3), 27 (2011)
7.
Zurück zum Zitat Codreanu, V., Dröge, B., Williams, D., Yasar, B., Yang, P., Liu, B, Dong, F., Surinta, O., Schomaker, L., Roerdink, J., et al: Evaluating automatically parallelized versions of the SVM. Concurrency and Computation: Practice and Experience (2014) Codreanu, V., Dröge, B., Williams, D., Yasar, B., Yang, P., Liu, B, Dong, F., Surinta, O., Schomaker, L., Roerdink, J., et al: Evaluating automatically parallelized versions of the SVM. Concurrency and Computation: Practice and Experience (2014)
8.
Zurück zum Zitat Cotter, A., Srebro, N., Keshet, J.: A GPU-tailored approach for training kernelized SVMs KDD, pp 805–813 (2011) Cotter, A., Srebro, N., Keshet, J.: A GPU-tailored approach for training kernelized SVMs KDD, pp 805–813 (2011)
9.
Zurück zum Zitat CUDA Nvidia: NVIDIA CUDA programming guide (2011) CUDA Nvidia: NVIDIA CUDA programming guide (2011)
10.
Zurück zum Zitat Fan, R.-E., Chen, P.-H., Lin, C.-J.: Working set selection using second order information for training support vector machines. JMLR 6, 1889–1918 (2005)MathSciNetMATH Fan, R.-E., Chen, P.-H., Lin, C.-J.: Working set selection using second order information for training support vector machines. JMLR 6, 1889–1918 (2005)MathSciNetMATH
11.
Zurück zum Zitat Flake, G.W., Lawrence, S.: Efficient SVM regression training with SMO. Mach. Learn. 46(1-3), 271–290 (2002)CrossRefMATH Flake, G.W., Lawrence, S.: Efficient SVM regression training with SMO. Mach. Learn. 46(1-3), 271–290 (2002)CrossRefMATH
12.
Zurück zum Zitat Forero, P.A., Cano, A., Giannakis, G.B.: Consensus-based distributed support vector machines. The Journal of Machine Learning Research 99, 1663–1707 (2010)MathSciNetMATH Forero, P.A., Cano, A., Giannakis, G.B.: Consensus-based distributed support vector machines. The Journal of Machine Learning Research 99, 1663–1707 (2010)MathSciNetMATH
13.
Zurück zum Zitat He, B., Fang, W., Luo, Q., Govindaraju, N.K., Wang, T.: Mars: a mapreduce framework on graphics processors Proceedings of the 17th international conference on Parallel architectures and compilation techniques, pp 260–269. ACM (2008) He, B., Fang, W., Luo, Q., Govindaraju, N.K., Wang, T.: Mars: a mapreduce framework on graphics processors Proceedings of the 17th international conference on Parallel architectures and compilation techniques, pp 260–269. ACM (2008)
14.
Zurück zum Zitat Hsu, C.-W., Chang, C.-C., Lin, C.-J., et al: A practical guide to support vector classification (2003) Hsu, C.-W., Chang, C.-C., Lin, C.-J., et al: A practical guide to support vector classification (2003)
15.
Zurück zum Zitat Hu, M., Hao, W.: A parallel approach for SVM with multi-core CPU International Conference on Computer Application and System Modeling, vol. 15, pp V15–373. IEEE (2010) Hu, M., Hao, W.: A parallel approach for SVM with multi-core CPU International Conference on Computer Application and System Modeling, vol. 15, pp V15–373. IEEE (2010)
16.
Zurück zum Zitat Joachims, T.: Making large-scale SVM learning practical Advances in kernel methods, pp 169–184. MIT Press (1999) Joachims, T.: Making large-scale SVM learning practical Advances in kernel methods, pp 169–184. MIT Press (1999)
17.
Zurück zum Zitat Joachims, T.: Training linear SVMs in linear time KDD, pp 217–226 (2006) Joachims, T.: Training linear SVMs in linear time KDD, pp 217–226 (2006)
18.
Zurück zum Zitat Jordaan, E.M., Smits, G.F.: Robust outlier detection using SVM regression IEEE International Joint Conference on Neural Networks, vol. 3, pp 2017–2022. IEEE (2004) Jordaan, E.M., Smits, G.F.: Robust outlier detection using SVM regression IEEE International Joint Conference on Neural Networks, vol. 3, pp 2017–2022. IEEE (2004)
19.
Zurück zum Zitat Kang, S., Park, S., Jung, H., Shim, H., Cha, J.: Performance trade-offs in using nvram write buffer for flash memory-based storage devices. IEEE Trans. Comput. 58(6), 744–758 (2009)MathSciNetCrossRef Kang, S., Park, S., Jung, H., Shim, H., Cha, J.: Performance trade-offs in using nvram write buffer for flash memory-based storage devices. IEEE Trans. Comput. 58(6), 744–758 (2009)MathSciNetCrossRef
20.
Zurück zum Zitat Kim, K.-J.: Financial time series forecasting using support vector machines. Neurocomputing 55(1), 307–319 (2003)CrossRef Kim, K.-J.: Financial time series forecasting using support vector machines. Neurocomputing 55(1), 307–319 (2003)CrossRef
21.
Zurück zum Zitat Li, Y., Gong, S., Liddell, H.: Support vector regression and classification based multi-view face detection and recognition International Conference on Automatic Face and Gesture Recognition, pp 300–305. IEEE (2000) Li, Y., Gong, S., Liddell, H.: Support vector regression and classification based multi-view face detection and recognition International Conference on Automatic Face and Gesture Recognition, pp 300–305. IEEE (2000)
22.
Zurück zum Zitat Nocedal, J., Wright, S.: Numerical optimization, series in operations research and financial engineering. Springer (2006) Nocedal, J., Wright, S.: Numerical optimization, series in operations research and financial engineering. Springer (2006)
23.
Zurück zum Zitat Nvidia CUDA: Cublas library. NVIDIA Corporation, Santa Clara, California, 15, 2008 Nvidia CUDA: Cublas library. NVIDIA Corporation, Santa Clara, California, 15, 2008
24.
Zurück zum Zitat Osuna, E., Freund, R., Girosi, F.: An improved training algorithm for support vector machines IEEE Workshop on Neural Networks for Signal Processing, pp 276–285. IEEE (1997) Osuna, E., Freund, R., Girosi, F.: An improved training algorithm for support vector machines IEEE Workshop on Neural Networks for Signal Processing, pp 276–285. IEEE (1997)
25.
Zurück zum Zitat Platt, J.C.: Fast training of SVMs using sequential minimal optimization Advances in kernel methods, pp 185–208. MIT Press (1999) Platt, J.C.: Fast training of SVMs using sequential minimal optimization Advances in kernel methods, pp 185–208. MIT Press (1999)
26.
Zurück zum Zitat Scholkopf, B., Smola, A.: Learning with kernels (2002) Scholkopf, B., Smola, A.: Learning with kernels (2002)
27.
Zurück zum Zitat Shalev-Shwartz, S., Singer, Y., Srebro, N., Cotter, A.: Pegasos: Primal estimated sub-gradient solver for svm. Math. Program. 127(1), 3–30 (2011)MathSciNetCrossRefMATH Shalev-Shwartz, S., Singer, Y., Srebro, N., Cotter, A.: Pegasos: Primal estimated sub-gradient solver for svm. Math. Program. 127(1), 3–30 (2011)MathSciNetCrossRefMATH
29.
Zurück zum Zitat Sun, Y., Yuan, N. J., Wang, Y., Xie, X., McDonald, K., Zhang, R.: Contextual intent tracking for personal assistants (2016) Sun, Y., Yuan, N. J., Wang, Y., Xie, X., McDonald, K., Zhang, R.: Contextual intent tracking for personal assistants (2016)
30.
Zurück zum Zitat Volkov, V.: Better performance at lower occupancy The GPU Technology Conference, vol. 10 (2010) Volkov, V.: Better performance at lower occupancy The GPU Technology Conference, vol. 10 (2010)
31.
Zurück zum Zitat Ward, P.G.D., He, Z., Zhang, R., Qi, J.: Real-time continuous intersection joins over large sets of moving objects using graphic processing units. The VLDB Journal 23(6), 965–985 (2014)CrossRef Ward, P.G.D., He, Z., Zhang, R., Qi, J.: Real-time continuous intersection joins over large sets of moving objects using graphic processing units. The VLDB Journal 23(6), 965–985 (2014)CrossRef
32.
Zurück zum Zitat Wen, Z., Zhang, R., Ramamohanarao, K.: Enabling precision/recall preferences for semi-supervised svm training Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, pp 421–430. ACM (2014) Wen, Z., Zhang, R., Ramamohanarao, K.: Enabling precision/recall preferences for semi-supervised svm training Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, pp 421–430. ACM (2014)
33.
Zurück zum Zitat Wen, Z., Zhang, R., Ramamohanarao, K., Qi, J., Taylor, K.: Mascot: Fast and highly scalable svm cross-validation using GPUs and SSDs ICDM. IEEE (2014) Wen, Z., Zhang, R., Ramamohanarao, K., Qi, J., Taylor, K.: Mascot: Fast and highly scalable svm cross-validation using GPUs and SSDs ICDM. IEEE (2014)
34.
Zurück zum Zitat Yoo, B., Won, Y., Cho, S., Kang, S., Choi, J., Yoon, S.: SSD characterization: From energy consumption’s perspective Proceedings of HotStorage (2011) Yoo, B., Won, Y., Cho, S., Kang, S., Choi, J., Yoon, S.: SSD characterization: From energy consumption’s perspective Proceedings of HotStorage (2011)
35.
Zurück zum Zitat Yang, L., Zhou, F., Xia, Y.: An improved caching strategy for training SVMs International Conference on Intelligent Systems and Knowledge Engineering, pp 1397–1401 (2007) Yang, L., Zhou, F., Xia, Y.: An improved caching strategy for training SVMs International Conference on Intelligent Systems and Knowledge Engineering, pp 1397–1401 (2007)
36.
Zurück zum Zitat Yang, Q., Ren, J.: I-cash: Intelligently coupled array of ssd and hdd 2011 IEEE 17th International Symposium on High Performance Computer Architecture, pp 278–289. IEEE (2011) Yang, Q., Ren, J.: I-cash: Intelligently coupled array of ssd and hdd 2011 IEEE 17th International Symposium on High Performance Computer Architecture, pp 278–289. IEEE (2011)
37.
Zurück zum Zitat Zhao, H.X., Magoules, F.: Parallel support vector machines on multi-core and multiprocessor systems International Conference on Artificial Intelligence and Applications. IASTED (2011) Zhao, H.X., Magoules, F.: Parallel support vector machines on multi-core and multiprocessor systems International Conference on Artificial Intelligence and Applications. IASTED (2011)
Metadaten
Titel
Scalable and fast SVM regression using modern hardware
verfasst von
Zeyi Wen
Rui Zhang
Kotagiri Ramamohanarao
Li Yang
Publikationsdatum
22.04.2017
Verlag
Springer US
Erschienen in
World Wide Web / Ausgabe 2/2018
Print ISSN: 1386-145X
Elektronische ISSN: 1573-1413
DOI
https://doi.org/10.1007/s11280-017-0445-1

Weitere Artikel der Ausgabe 2/2018

World Wide Web 2/2018 Zur Ausgabe