Skip to main content
Erschienen in: Cluster Computing 2/2017

10.04.2017

K- local maximum margin feature extraction algorithm for churn prediction in telecom

verfasst von: Long Zhao, Qian Gao, XiangJun Dong, Aimei Dong, Xue Dong

Erschienen in: Cluster Computing | Ausgabe 2/2017

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Telecom customer churn data is not publicly available because involving users’ personal privacy. In 2009, the French telecommunications company Orange for knowledge discovery and data mining (KDD) competition provides a telecom customer churn data set KDD Cup 09. In order to solve the high dimensional problem of KDD Cup 09, a new feature reduction method is used to explore the influence of different features on the prediction of classification model. In this paper, a new K- local maximum margin feature extraction algorithm (KLMM) is proposed. Through researching on the diversification subspace partition rules, the corresponding potential field structure is constructed. According to the data source in the dimension of scalability, the intrinsic link between data attributes and classification results is revealed. The extracted features can reduce the dimension of the churn prediction in telecom data. The KLMM method adapts auto selection sigma factor to reflect the anisotropy of features. The potential function is used to assess the weights of attributes and find the potential important weight. Experiments and analysis show that the extracted features by KLMM are more likely to find a classification hyperplane which can separate data points of the different classes.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Xu, H., Zhang, Z., Zhang, Y.: Churn prediction in telecom using a hybrid two-phase feature selection method[C] international symposium on intelligent information technology application. 576–579 (2009) Xu, H., Zhang, Z., Zhang, Y.: Churn prediction in telecom using a hybrid two-phase feature selection method[C] international symposium on intelligent information technology application. 576–579 (2009)
2.
Zurück zum Zitat Idris, A., Khan, A., Lee, Y.S.: Intelligent churn prediction in telecom: employing mRMR feature selection and RotBoost based ensemble classification. Appl. Intell. 39(3), 659–672 (2013)CrossRef Idris, A., Khan, A., Lee, Y.S.: Intelligent churn prediction in telecom: employing mRMR feature selection and RotBoost based ensemble classification. Appl. Intell. 39(3), 659–672 (2013)CrossRef
3.
Zurück zum Zitat Fathian, M., Hoseinpoor, Y., Minaei-Bidgoli, B.: Offering a hybrid approach of data mining to predict the customer churn based on bagging and boosting methods. Kybernetes 45(5), 732–743 (2016)MathSciNetCrossRef Fathian, M., Hoseinpoor, Y., Minaei-Bidgoli, B.: Offering a hybrid approach of data mining to predict the customer churn based on bagging and boosting methods. Kybernetes 45(5), 732–743 (2016)MathSciNetCrossRef
4.
Zurück zum Zitat Idris, A., Khan, A., Lee, Y.S.: Intelligent churn prediction in telecom: employing mrmr feature selection and rotboost based ensemble classification. Appl. Intell. 39(3), 659–672 (2013)CrossRef Idris, A., Khan, A., Lee, Y.S.: Intelligent churn prediction in telecom: employing mrmr feature selection and rotboost based ensemble classification. Appl. Intell. 39(3), 659–672 (2013)CrossRef
5.
Zurück zum Zitat Xiao, J., Jiang, X., He, C., Teng, G.: Churn prediction in customer relationship management via gmdh-based multiple classifiers ensemble. IEEE Intell. Syst. 31(2), 37–44 (2016)CrossRef Xiao, J., Jiang, X., He, C., Teng, G.: Churn prediction in customer relationship management via gmdh-based multiple classifiers ensemble. IEEE Intell. Syst. 31(2), 37–44 (2016)CrossRef
6.
Zurück zum Zitat Yang, B., Xu, J., Yang, J., Li, M.: Localization algorithm in wireless sensor networks based on semi-supervised manifold learning and its application. Clus. Comput. 13(4), 435–446 (2010)CrossRef Yang, B., Xu, J., Yang, J., Li, M.: Localization algorithm in wireless sensor networks based on semi-supervised manifold learning and its application. Clus. Comput. 13(4), 435–446 (2010)CrossRef
7.
Zurück zum Zitat Mirebeau, J.M.: Anisotropic fast-marching on cartesian grids using lattice basis reduction. Siam J. Numer. Anal. 52, 1573–1599 (2014)MathSciNetCrossRefMATH Mirebeau, J.M.: Anisotropic fast-marching on cartesian grids using lattice basis reduction. Siam J. Numer. Anal. 52, 1573–1599 (2014)MathSciNetCrossRefMATH
8.
Zurück zum Zitat Daniel, S.F., Connolly, A., Schneider, J., Vanderplas, J., Xiong, L.: Classification of stellar spectra with local linear embedding. Astron. J. 142(6), 557–561 (2011)CrossRef Daniel, S.F., Connolly, A., Schneider, J., Vanderplas, J., Xiong, L.: Classification of stellar spectra with local linear embedding. Astron. J. 142(6), 557–561 (2011)CrossRef
10.
Zurück zum Zitat Li, B., Zheng, C.H., Huang, D.S.: Locally linear discriminant embedding: an efficient method for face recognition. Pattern Recogn. 41(12), 3813–3821 (2008)CrossRefMATH Li, B., Zheng, C.H., Huang, D.S.: Locally linear discriminant embedding: an efficient method for face recognition. Pattern Recogn. 41(12), 3813–3821 (2008)CrossRefMATH
11.
Zurück zum Zitat Li, J.B., Pan, J.S., Chu, S.C.: Kernel class-wise locality preserving projection. Inf. Sci. 178(7), 1825–1835 (2008)CrossRefMATH Li, J.B., Pan, J.S., Chu, S.C.: Kernel class-wise locality preserving projection. Inf. Sci. 178(7), 1825–1835 (2008)CrossRefMATH
12.
Zurück zum Zitat Monge, D.A., Holec, M., Železný, F., Garino, C.G.: Ensemble learning of runtime prediction models for gene-expression analysis workflows. Clus. Comput. 18(4), 1317–1329 (2015)CrossRef Monge, D.A., Holec, M., Železný, F., Garino, C.G.: Ensemble learning of runtime prediction models for gene-expression analysis workflows. Clus. Comput. 18(4), 1317–1329 (2015)CrossRef
13.
Zurück zum Zitat Kwak, N.: Nonlinear projection trick in kernel methods: an alternative to the kernel trick. IEEE Trans. Neural Netw. Learn. Syst. 24(12), 2113 (2013)CrossRef Kwak, N.: Nonlinear projection trick in kernel methods: an alternative to the kernel trick. IEEE Trans. Neural Netw. Learn. Syst. 24(12), 2113 (2013)CrossRef
14.
Zurück zum Zitat Jang, J., Lee, Y., Lee, S., Shin, S., Kim, D., Rim, H.: A novel density-based clustering method using word embedding features for dialogue intention recognition. Clust. Comput. 19, 2315–2326 (2016)CrossRef Jang, J., Lee, Y., Lee, S., Shin, S., Kim, D., Rim, H.: A novel density-based clustering method using word embedding features for dialogue intention recognition. Clust. Comput. 19, 2315–2326 (2016)CrossRef
15.
Zurück zum Zitat Yang, J., Zhang, L., Yang, J.Y., Zhang, D.: From classifiers to discriminators: a nearest neighbor rule induced discriminant analysis. Pattern Recogn. 44(7), 1387–1402 (2011)CrossRefMATH Yang, J., Zhang, L., Yang, J.Y., Zhang, D.: From classifiers to discriminators: a nearest neighbor rule induced discriminant analysis. Pattern Recogn. 44(7), 1387–1402 (2011)CrossRefMATH
16.
Zurück zum Zitat Villegas, M., Paredes, R.: Dimensionality reduction by minimizing nearest-neighbor classification error. Pattern Recogn. Lett. 32(4), 633–639 (2011)CrossRef Villegas, M., Paredes, R.: Dimensionality reduction by minimizing nearest-neighbor classification error. Pattern Recogn. Lett. 32(4), 633–639 (2011)CrossRef
17.
Zurück zum Zitat Guyon, I., Lemaire, V., Dror, G., Vogel, D.: Design and analysis of the kdd cup 2009: fast scoring on a large orange customer database. ACM Sigkdd Explor. Newslett. 11(2), 68–76 (2010)CrossRef Guyon, I., Lemaire, V., Dror, G., Vogel, D.: Design and analysis of the kdd cup 2009: fast scoring on a large orange customer database. ACM Sigkdd Explor. Newslett. 11(2), 68–76 (2010)CrossRef
18.
Zurück zum Zitat Rodan, A., Faris, H., Al-Sakran, J., Al-Kadi, O.: A support vector machine approach for churn prediction in telecom industry. Int. J. Inf. 17(8), 3961 (2014) Rodan, A., Faris, H., Al-Sakran, J., Al-Kadi, O.: A support vector machine approach for churn prediction in telecom industry. Int. J. Inf. 17(8), 3961 (2014)
19.
Zurück zum Zitat Li, D., Wang, S., Gan, W., Li, D.: Data field for hierarchical clustering. Int. J. Data Warehous. Min. 7(4), 43–63 (2011)CrossRef Li, D., Wang, S., Gan, W., Li, D.: Data field for hierarchical clustering. Int. J. Data Warehous. Min. 7(4), 43–63 (2011)CrossRef
20.
Zurück zum Zitat Li, C., Liu, Q., Dong, W., Wei, F., Zhang, X., Yang, L.: Max-margin-based discriminative feature learning. IEEE Trans. Neural Netw. Learning Syst. 27(12), 2768–2775 (2016)CrossRef Li, C., Liu, Q., Dong, W., Wei, F., Zhang, X., Yang, L.: Max-margin-based discriminative feature learning. IEEE Trans. Neural Netw. Learning Syst. 27(12), 2768–2775 (2016)CrossRef
21.
Zurück zum Zitat Yong-Zhi, L.I., Yang, J.Y., Zheng, Y.J., Xia, Y.Q.: New and efficient feature extraction methods based on maximum margin criterion. J. Syst. Simul. 19(5), 1061–1066 (2007) Yong-Zhi, L.I., Yang, J.Y., Zheng, Y.J., Xia, Y.Q.: New and efficient feature extraction methods based on maximum margin criterion. J. Syst. Simul. 19(5), 1061–1066 (2007)
22.
Zurück zum Zitat Sang, Y.O., Chung, K.: Vocabulary optimization process using similar phoneme recognition and feature extraction. Clust. Comput. 19, 1683–1690 (2016)CrossRef Sang, Y.O., Chung, K.: Vocabulary optimization process using similar phoneme recognition and feature extraction. Clust. Comput. 19, 1683–1690 (2016)CrossRef
23.
Zurück zum Zitat Zhu, Q., Feng, J., Huang, J.: Weighted natural neighborhood graph: an adaptive structure for clustering and outlier detection with no neighborhood parameter. Clust. Comput. 19(3), 1–13 (2016) Zhu, Q., Feng, J., Huang, J.: Weighted natural neighborhood graph: an adaptive structure for clustering and outlier detection with no neighborhood parameter. Clust. Comput. 19(3), 1–13 (2016)
24.
Zurück zum Zitat Yang, H.H., Moody, J.: Data visualization and feature selection: new algorithms for nongaussian data. Adv. Neural Inf. Process. Syst. 12, 687–693 (2000) Yang, H.H., Moody, J.: Data visualization and feature selection: new algorithms for nongaussian data. Adv. Neural Inf. Process. Syst. 12, 687–693 (2000)
25.
Zurück zum Zitat Meyer, P.E., Bontempi, G.: On the Use of Variable Complementarity for Feature Selection in Cancer Classification. Applications of Evolutionary Computing, Springer (2006)CrossRef Meyer, P.E., Bontempi, G.: On the Use of Variable Complementarity for Feature Selection in Cancer Classification. Applications of Evolutionary Computing, Springer (2006)CrossRef
26.
Zurück zum Zitat Lin, D., Tang, X.: (2006). Conditional Infomax Learning: an integrated framework for feature extraction and fusion. Computer vision - ECCV 2006, European Conference on Computer Vision, Graz, Austria, Proceedings vol. 3951, pp. 68–82. May 7–13 2006 Lin, D., Tang, X.: (2006). Conditional Infomax Learning: an integrated framework for feature extraction and fusion. Computer vision - ECCV 2006, European Conference on Computer Vision, Graz, Austria, Proceedings vol. 3951, pp. 68–82. May 7–13 2006
27.
Zurück zum Zitat Bratko, I.: (2005). Machine learning based on attribute interactions: PhD dissertation Bratko, I.: (2005). Machine learning based on attribute interactions: PhD dissertation
28.
Zurück zum Zitat Cheng, H., Qin, Z., Feng, C., Wang, Y., Li, F.: Conditional mutual information-based feature selection analyzing for synergy and redundancy. Etri J. 33(2), 210–218 (2011)CrossRef Cheng, H., Qin, Z., Feng, C., Wang, Y., Li, F.: Conditional mutual information-based feature selection analyzing for synergy and redundancy. Etri J. 33(2), 210–218 (2011)CrossRef
Metadaten
Titel
K- local maximum margin feature extraction algorithm for churn prediction in telecom
verfasst von
Long Zhao
Qian Gao
XiangJun Dong
Aimei Dong
Xue Dong
Publikationsdatum
10.04.2017
Verlag
Springer US
Erschienen in
Cluster Computing / Ausgabe 2/2017
Print ISSN: 1386-7857
Elektronische ISSN: 1573-7543
DOI
https://doi.org/10.1007/s10586-017-0843-2

Weitere Artikel der Ausgabe 2/2017

Cluster Computing 2/2017 Zur Ausgabe