Skip to main content

2022 | OriginalPaper | Buchkapitel

4. Feature Selection

verfasst von : Yong Shi

Erschienen in: Advances in Big Data Analytics

Verlag: Springer Nature Singapore

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In big data analytics, irrelevant and redundant features may not only deteriorate the performances of classifiers, but also slow down the prediction process. Although there is the availability of many classification models for prediction, it is a challenge to choose a set of important features that can lead to a satisfactory classifier. This chapter outlines some achievements of feature selection research in the last decade. Section 4.1 has three subsections. The first is an integrated scheme for feature selection and classifier evaluation in the context of prediction [1]. The second is about two-stage hybrid feature selection algorithms [2]. The third one is the feature selection with attributes clustering by maximal information coefficient [3]. Section 4.2 presents two regularizations for feature selections. They are feature selection with MCP2 regularization [4] and feature selection with 2, 1 − 2 regularization [5]. Finally, Sect. 4.3 describes two distance-based feature selections. They are the spatial distance join based feature selection [1] and a domain driven two-phase feature selection method based on bhattacharyya distance and kernel distance measurements [6].

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Peng, Y., Kou, G., Ergu, D., Wu, W., Shi, Y.: An integrated feature selection and classification scheme. Stud. Inf. Contr. 21(3), 241–248 (2012) Peng, Y., Kou, G., Ergu, D., Wu, W., Shi, Y.: An integrated feature selection and classification scheme. Stud. Inf. Contr. 21(3), 241–248 (2012)
2.
Zurück zum Zitat Xie, J., Lei, J., Xie, W., Shi, Y., Liu, X.: Two-stage hybrid feature selection algorithms for diagnosing erythemato-squamous diseases. Health Inf. Sci. Syst. 1(1), 1–14 (2013) Xie, J., Lei, J., Xie, W., Shi, Y., Liu, X.: Two-stage hybrid feature selection algorithms for diagnosing erythemato-squamous diseases. Health Inf. Sci. Syst. 1(1), 1–14 (2013)
3.
Zurück zum Zitat Zhao, X., Deng, W., Shi, Y.: Feature selection with attributes clustering by maximal information coefficient. Proc. Comput. Sci. 17, 70–79 (2013) Zhao, X., Deng, W., Shi, Y.: Feature selection with attributes clustering by maximal information coefficient. Proc. Comput. Sci. 17, 70–79 (2013)
4.
Zurück zum Zitat Shi, Y., Miao, J., Niu, L.: Feature selection with MCP2 regularization. Neural Comput. Applic. 31(10), 6699–6709 (2019) Shi, Y., Miao, J., Niu, L.: Feature selection with MCP2 regularization. Neural Comput. Applic. 31(10), 6699–6709 (2019)
6.
Zurück zum Zitat Chen, Y., Zhang, L., Li, J., Shi, Y.: Domain driven two-phase feature selection method based on Bhattacharyya distance and Kernel distance measurements. In: 2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology, vol. 3, pp. 217–220. IEEE, New York (2011) Chen, Y., Zhang, L., Li, J., Shi, Y.: Domain driven two-phase feature selection method based on Bhattacharyya distance and Kernel distance measurements. In: 2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology, vol. 3, pp. 217–220. IEEE, New York (2011)
7.
Zurück zum Zitat Myrtveit, I., Stensrud, E., Shepperd, M.: Reliability and validity in comparative studies of software prediction models. IEEE Trans. Softw. Eng. 31(5), 380–391 (2005) Myrtveit, I., Stensrud, E., Shepperd, M.: Reliability and validity in comparative studies of software prediction models. IEEE Trans. Softw. Eng. 31(5), 380–391 (2005)
8.
Zurück zum Zitat Witten, I.H., Frank, E.: Data mining: practical machine learning tools and techniques with java implementations. ACM SIGMOD Rec. 31(1), 76–77 (2002) Witten, I.H., Frank, E.: Data mining: practical machine learning tools and techniques with java implementations. ACM SIGMOD Rec. 31(1), 76–77 (2002)
9.
Zurück zum Zitat Brans, J.P., De Smet, Y.: Promethee methods. In: Multiple Criteria Decision Analysis, pp. 187–219. Springer, New York (2016) Brans, J.P., De Smet, Y.: Promethee methods. In: Multiple Criteria Decision Analysis, pp. 187–219. Springer, New York (2016)
10.
Zurück zum Zitat Opricovic, S.: Multicriteria Optimization of Civil Engineering Systems. Faculty of Civil Engineering, Belgrade 2(1), 5–21 (1998) Opricovic, S.: Multicriteria Optimization of Civil Engineering Systems. Faculty of Civil Engineering, Belgrade 2(1), 5–21 (1998)
11.
Zurück zum Zitat Opricovic, S., Tzeng, G.H.: Compromise solution by mcdm methods: a comparative analysis of vikor and topsis. Eur. J. Oper. Res. 156(2), 445–455 (2004)MATH Opricovic, S., Tzeng, G.H.: Compromise solution by mcdm methods: a comparative analysis of vikor and topsis. Eur. J. Oper. Res. 156(2), 445–455 (2004)MATH
12.
Zurück zum Zitat Charnes, A., Cooper, W.W., Rhodes, E.: Measuring the efficiency of decision making units. Eur. J. Oper. Res. 2(6), 429–444 (1978)MathSciNetMATH Charnes, A., Cooper, W.W., Rhodes, E.: Measuring the efficiency of decision making units. Eur. J. Oper. Res. 2(6), 429–444 (1978)MathSciNetMATH
13.
Zurück zum Zitat Cooper, W.W., Seiford, L.M., Zhu, J.: Data envelopment analysis: history, models, and interpretations. In: Handbook on Data Envelopment Analysis, pp. 1–39. Springer, New York (2004)MATH Cooper, W.W., Seiford, L.M., Zhu, J.: Data envelopment analysis: history, models, and interpretations. In: Handbook on Data Envelopment Analysis, pp. 1–39. Springer, New York (2004)MATH
14.
Zurück zum Zitat Nakhaeizadeh, G., Schnabl, A.: Development of multi-criteria metrics for evaluation of data mining algorithms. In: KDD, pp. 37–42 (1997) Nakhaeizadeh, G., Schnabl, A.: Development of multi-criteria metrics for evaluation of data mining algorithms. In: KDD, pp. 37–42 (1997)
15.
Zurück zum Zitat Roy, B.: Classement et choix en présence de points de vue multiples. RAIRO Oper. Res. Recherche Opérationnelle. 2(V1), 57–75 (1968) Roy, B.: Classement et choix en présence de points de vue multiples. RAIRO Oper. Res. Recherche Opérationnelle. 2(V1), 57–75 (1968)
16.
Zurück zum Zitat Kou, G., Lou, C.: Multiple factor hierarchical clustering algorithm for large scale web page and search engine clickstream data. Ann. Oper. Res. 197(1), 123–134 (2012) Kou, G., Lou, C.: Multiple factor hierarchical clustering algorithm for large scale web page and search engine clickstream data. Ann. Oper. Res. 197(1), 123–134 (2012)
17.
Zurück zum Zitat Kou, G., Lu, Y., Peng, Y., Shi, Y.: Evaluation of classification algorithms using mcdm and rank correlation. Int. J. Inf. Technol. Decis. Mak. 11(01), 197–225 (2012) Kou, G., Lu, Y., Peng, Y., Shi, Y.: Evaluation of classification algorithms using mcdm and rank correlation. Int. J. Inf. Technol. Decis. Mak. 11(01), 197–225 (2012)
18.
Zurück zum Zitat Peng, Y., Kou, G., Wang, G., Wu, W., Shi, Y.: Ensemble of software defect predictors: an ahp-based evaluation method. Int. J. Inf. Technol. Decis. Mak. 10(01), 187–206 (2011) Peng, Y., Kou, G., Wang, G., Wu, W., Shi, Y.: Ensemble of software defect predictors: an ahp-based evaluation method. Int. J. Inf. Technol. Decis. Mak. 10(01), 187–206 (2011)
19.
Zurück zum Zitat Fu, K.S., Min, P.J., Li, T.J.: Feature selection in pattern recognition. IEEE Trans. Syst. Sci. Cybernet. 6(1), 33–39 (1970)MATH Fu, K.S., Min, P.J., Li, T.J.: Feature selection in pattern recognition. IEEE Trans. Syst. Sci. Cybernet. 6(1), 33–39 (1970)MATH
20.
Zurück zum Zitat Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3(Mar), 1157–1182 (2003)MATH Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3(Mar), 1157–1182 (2003)MATH
21.
Zurück zum Zitat Hua, J., Tembe, W.D., Dougherty, E.R.: Performance of feature-selection methods in the classification of high-dimension data. Pattern Recogn. 42(3), 409–424 (2009)MATH Hua, J., Tembe, W.D., Dougherty, E.R.: Performance of feature-selection methods in the classification of high-dimension data. Pattern Recogn. 42(3), 409–424 (2009)MATH
22.
Zurück zum Zitat Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artif. Intell. 97(1–2), 245–271 (1997)MathSciNetMATH Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artif. Intell. 97(1–2), 245–271 (1997)MathSciNetMATH
23.
Zurück zum Zitat Kohavi, R., John, G.H.: Wrappers for feature subset selection. Artif. Intell. 97(1–2), 273–324 (1997)MATH Kohavi, R., John, G.H.: Wrappers for feature subset selection. Artif. Intell. 97(1–2), 273–324 (1997)MATH
24.
Zurück zum Zitat Xie, J., Xie, W., Wang, C., Gao, X.: A novel hybrid feature selection method based on ifsffs and svm for the diagnosis of erythemato-squamous diseases. In: Proceedings of the First Workshop on Applications of Pattern Analysis, pp. 142–151. PMLR (2010) Xie, J., Xie, W., Wang, C., Gao, X.: A novel hybrid feature selection method based on ifsffs and svm for the diagnosis of erythemato-squamous diseases. In: Proceedings of the First Workshop on Applications of Pattern Analysis, pp. 142–151. PMLR (2010)
25.
Zurück zum Zitat Whitney, A.W.: A direct method of nonparametric measurement selection. IEEE Trans. Comput. 100(9), 1100–1103 (1971)MATH Whitney, A.W.: A direct method of nonparametric measurement selection. IEEE Trans. Comput. 100(9), 1100–1103 (1971)MATH
26.
Zurück zum Zitat Marill, T., Green, D.: On the effectiveness of receptors in recognition systems. IEEE Trans. Inf. Theory. 9(1), 11–17 (1963) Marill, T., Green, D.: On the effectiveness of receptors in recognition systems. IEEE Trans. Inf. Theory. 9(1), 11–17 (1963)
27.
Zurück zum Zitat Pudil, P., Novovičová, J., Kittler, J.: Floating search methods in feature selection. Pattern Recogn. Lett. 15(11), 1119–1125 (1994) Pudil, P., Novovičová, J., Kittler, J.: Floating search methods in feature selection. Pattern Recogn. Lett. 15(11), 1119–1125 (1994)
28.
Zurück zum Zitat Pereira, F., Tishby, N., Lee, L.: Distributional clustering of English words. In: Proceedings of the 31st Annual Meeting on Association for Computational Linguistics, pp. 183–190 (1993) Pereira, F., Tishby, N., Lee, L.: Distributional clustering of English words. In: Proceedings of the 31st Annual Meeting on Association for Computational Linguistics, pp. 183–190 (1993)
29.
Zurück zum Zitat Tishby, N., Pereira, F.C., Bialek, W.: The information bottleneck method. arXiv preprint physics/0004057 (2000) Tishby, N., Pereira, F.C., Bialek, W.: The information bottleneck method. arXiv preprint physics/0004057 (2000)
30.
Zurück zum Zitat Baker, L.D., McCallum, A.K.: Distributional clustering of words for text classification. In: Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 96–103 (1998) Baker, L.D., McCallum, A.K.: Distributional clustering of words for text classification. In: Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 96–103 (1998)
31.
Zurück zum Zitat Dhillon, I.S., Mallela, S., Kumar, R.: A divisive information theoretic feature clustering algorithm for text classification. J. Mach. Learn. Res. 3, 1265–1287 (2003)MathSciNetMATH Dhillon, I.S., Mallela, S., Kumar, R.: A divisive information theoretic feature clustering algorithm for text classification. J. Mach. Learn. Res. 3, 1265–1287 (2003)MathSciNetMATH
32.
Zurück zum Zitat Au, W.H., Chan, K.C., Wong, A.K., Wang, Y.: Attribute clustering for grouping, selection, and classification of gene expression data. IEEE/ACM Trans. Comput. Biol. Bioinform. 2(2), 83–101 (2005) Au, W.H., Chan, K.C., Wong, A.K., Wang, Y.: Attribute clustering for grouping, selection, and classification of gene expression data. IEEE/ACM Trans. Comput. Biol. Bioinform. 2(2), 83–101 (2005)
33.
Zurück zum Zitat Jiang, J.Y., Liou, R.J., Lee, S.J.: A fuzzy self-constructing feature clustering algorithm for text classification. IEEE Trans. Knowl. Data Eng. 23(3), 335–349 (2010) Jiang, J.Y., Liou, R.J., Lee, S.J.: A fuzzy self-constructing feature clustering algorithm for text classification. IEEE Trans. Knowl. Data Eng. 23(3), 335–349 (2010)
34.
Zurück zum Zitat Amaldi, E., Kann, V.: On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems. Theor. Comput. Sci. 209(1–2), 237–260 (1998)MathSciNetMATH Amaldi, E., Kann, V.: On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems. Theor. Comput. Sci. 209(1–2), 237–260 (1998)MathSciNetMATH
35.
Zurück zum Zitat Combarro, E.F., Montanes, E., Diaz, I., Ranilla, J., Mones, R.: Introducing a family of linear measures for feature selection in text categorization. IEEE Trans. Knowl. Data Eng. 17(9), 1223–1232 (2005) Combarro, E.F., Montanes, E., Diaz, I., Ranilla, J., Mones, R.: Introducing a family of linear measures for feature selection in text categorization. IEEE Trans. Knowl. Data Eng. 17(9), 1223–1232 (2005)
36.
Zurück zum Zitat Reshef, D.N., Reshef, Y.A., Finucane, H.K., Grossman, S.R., McVean, G., Turnbaugh, P.J., Lander, E.S., Mitzenmacher, M., Sabeti, P.C.: Detecting novel associations in large data sets. Science. 334(6062), 1518–1524 (2011)MATH Reshef, D.N., Reshef, Y.A., Finucane, H.K., Grossman, S.R., McVean, G., Turnbaugh, P.J., Lander, E.S., Mitzenmacher, M., Sabeti, P.C.: Detecting novel associations in large data sets. Science. 334(6062), 1518–1524 (2011)MATH
37.
Zurück zum Zitat Frey, B.J., Dueck, D.: Clustering by passing messages between data points. Science. 315(5814), 972–976 (2007)MathSciNetMATH Frey, B.J., Dueck, D.: Clustering by passing messages between data points. Science. 315(5814), 972–976 (2007)MathSciNetMATH
38.
Zurück zum Zitat Cai, X., Nie, F., Huang, H., Ding, C.: Multi-class l2, 1-norm support vector machine. In: 2011 IEEE 11th International Conference on Data Mining, pp. 91–100. IEEE, New York (2011) Cai, X., Nie, F., Huang, H., Ding, C.: Multi-class l2, 1-norm support vector machine. In: 2011 IEEE 11th International Conference on Data Mining, pp. 91–100. IEEE, New York (2011)
39.
Zurück zum Zitat Ma, Z., Nie, F., Yang, Y., Uijlings, J.R., Sebe, N.: Web image annotation via subspace-sparsity collaborated feature selection. IEEE Trans. Multimedia. 14(4), 1021–1030 (2012) Ma, Z., Nie, F., Yang, Y., Uijlings, J.R., Sebe, N.: Web image annotation via subspace-sparsity collaborated feature selection. IEEE Trans. Multimedia. 14(4), 1021–1030 (2012)
40.
Zurück zum Zitat Nie, F., Huang, H., Cai, X., Ding, C.: Efficient and robust feature selection via joint 2, 1-norms minimization. Adv. Neural Inf. Proces. Syst. 23, 1813–1821 (2010) Nie, F., Huang, H., Cai, X., Ding, C.: Efficient and robust feature selection via joint 2, 1-norms minimization. Adv. Neural Inf. Proces. Syst. 23, 1813–1821 (2010)
41.
Zurück zum Zitat Xiang, S., Nie, F., Meng, G., Pan, C., Zhang, C.: Discriminative least squares regression for multiclass classification and feature selection. IEEE Trans. Neural Netw. Learn. Syst. 23(11), 1738–1754 (2012) Xiang, S., Nie, F., Meng, G., Pan, C., Zhang, C.: Discriminative least squares regression for multiclass classification and feature selection. IEEE Trans. Neural Netw. Learn. Syst. 23(11), 1738–1754 (2012)
42.
Zurück zum Zitat Zhang, M., Ding, C., Zhang, Y., Nie, F.: Feature selection at the discrete limit. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 28, (2014) Zhang, M., Ding, C., Zhang, Y., Nie, F.: Feature selection at the discrete limit. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 28, (2014)
43.
Zurück zum Zitat Sriperumbudur, B.K., Lanckriet, G.R.: On the convergence of the concave-convex procedure. In: Nips, vol. 9, pp. 1759–1767. Citeseer (2009) Sriperumbudur, B.K., Lanckriet, G.R.: On the convergence of the concave-convex procedure. In: Nips, vol. 9, pp. 1759–1767. Citeseer (2009)
44.
Zurück zum Zitat Yen, I.E., Peng, N., Wang, P.W., Lin, S.D.: On convergence rate of concave-convex procedure. In: Proceedings of the NIPS 2012 Optimization Work-shop, pp. 31–35 (2012) Yen, I.E., Peng, N., Wang, P.W., Lin, S.D.: On convergence rate of concave-convex procedure. In: Proceedings of the NIPS 2012 Optimization Work-shop, pp. 31–35 (2012)
45.
Zurück zum Zitat Yuille, A.L., Rangarajan, A., Yuille, A.: The concave-convex procedure (cccp). Adv. Neural Inf. Proces. Syst. 2, 1033–1040 (2002)MATH Yuille, A.L., Rangarajan, A., Yuille, A.: The concave-convex procedure (cccp). Adv. Neural Inf. Proces. Syst. 2, 1033–1040 (2002)MATH
46.
Zurück zum Zitat Yin, P., Lou, Y., He, Q., Xin, J.: Minimization of 1-2 for compressed sensing. SIAM J. Sci. Comput. 37(1), A536–A563 (2015)MathSciNetMATH Yin, P., Lou, Y., He, Q., Xin, J.: Minimization of 1-2 for compressed sensing. SIAM J. Sci. Comput. 37(1), A536–A563 (2015)MathSciNetMATH
47.
Zurück zum Zitat Esser, E., Lou, Y., Xin, J.: A method for finding structured sparse solutions to nonnegative least squares problems with applications. SIAM J. Imag. Sci. 6(4), 2010–2046 (2013)MathSciNetMATH Esser, E., Lou, Y., Xin, J.: A method for finding structured sparse solutions to nonnegative least squares problems with applications. SIAM J. Imag. Sci. 6(4), 2010–2046 (2013)MathSciNetMATH
48.
Zurück zum Zitat Lou, Y., Osher, S., Xin, J.: Computational aspects of constrained l 1-l 2 minimization for compressive sensing. In: Modelling, Computation and Optimization in Information Systems and Management Sciences, pp. 169–180. Springer, New York (2015) Lou, Y., Osher, S., Xin, J.: Computational aspects of constrained l 1-l 2 minimization for compressive sensing. In: Modelling, Computation and Optimization in Information Systems and Management Sciences, pp. 169–180. Springer, New York (2015)
49.
Zurück zum Zitat Lou, Y., Yin, P., He, Q., Xin, J.: Computing sparse representation in a highly coherent dictionary based on difference of l 1 and l 2. J. Sci. Comput. 64(1), 178–196 (2015)MathSciNetMATH Lou, Y., Yin, P., He, Q., Xin, J.: Computing sparse representation in a highly coherent dictionary based on difference of l 1 and l 2. J. Sci. Comput. 64(1), 178–196 (2015)MathSciNetMATH
50.
Zurück zum Zitat Zangwill, W.I.: Nonlinear Programming: A Unified Approach, vol. 52. Prentice-Hall, Englewood Cliffs, NJ (1969)MATH Zangwill, W.I.: Nonlinear Programming: A Unified Approach, vol. 52. Prentice-Hall, Englewood Cliffs, NJ (1969)MATH
51.
Zurück zum Zitat Bertsekas, D.P.: Convex Optimization Theory. Athena Scientific, Belmont (2009)MATH Bertsekas, D.P.: Convex Optimization Theory. Athena Scientific, Belmont (2009)MATH
52.
Zurück zum Zitat Rockafellar, R.T., Wets, R.J.B.: Variational Analysis, vol. 317. Springer Science & Business Media, New York (2009)MATH Rockafellar, R.T., Wets, R.J.B.: Variational Analysis, vol. 317. Springer Science & Business Media, New York (2009)MATH
53.
Zurück zum Zitat Meyer, R.R.: Sufficient conditions for the convergence of monotonic mathematical programming algorithms. J. Comput. Syst. Sci. 12(1), 108–121 (1976)MathSciNetMATH Meyer, R.R.: Sufficient conditions for the convergence of monotonic mathematical programming algorithms. J. Comput. Syst. Sci. 12(1), 108–121 (1976)MathSciNetMATH
54.
Zurück zum Zitat Boyd, S., Parikh, N., Chu, E.: Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Now Publishers Inc, Delft, Netherlands (2011)MATH Boyd, S., Parikh, N., Chu, E.: Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Now Publishers Inc, Delft, Netherlands (2011)MATH
55.
Zurück zum Zitat Li, Z., Liu, J., Yang, Y., Zhou, X., Lu, H.: Clustering-guided sparse structural learning for unsupervised feature selection. IEEE Trans. Knowl. Data Eng. 26(9), 2138–2150 (2013) Li, Z., Liu, J., Yang, Y., Zhou, X., Lu, H.: Clustering-guided sparse structural learning for unsupervised feature selection. IEEE Trans. Knowl. Data Eng. 26(9), 2138–2150 (2013)
56.
Zurück zum Zitat Yang, Y., Shen, H.T., Ma, Z., Huang, Z., Zhou, X.: 2, 1-norm regularized discriminative feature selection for unsupervised learning. In: IJCAI International Joint Conference on Artificial Intelligence (2011) Yang, Y., Shen, H.T., Ma, Z., Huang, Z., Zhou, X.: 2, 1-norm regularized discriminative feature selection for unsupervised learning. In: IJCAI International Joint Conference on Artificial Intelligence (2011)
57.
Zurück zum Zitat Huang, J., Nie, F., Huang, H., Ding, C.: Robust manifold nonnegative matrix factorization. ACM Trans. Knowl. Discov. Data. 8(3), 1–21 (2014) Huang, J., Nie, F., Huang, H., Ding, C.: Robust manifold nonnegative matrix factorization. ACM Trans. Knowl. Discov. Data. 8(3), 1–21 (2014)
58.
Zurück zum Zitat Jiang, W., Nie, F., Huang, H.: Robust dictionary learning with capped l1-norm. In: Twenty-Fourth International Joint Conference on Artificial Intelligence (2015) Jiang, W., Nie, F., Huang, H.: Robust dictionary learning with capped l1-norm. In: Twenty-Fourth International Joint Conference on Artificial Intelligence (2015)
59.
Zurück zum Zitat Fan, J., Li, R.: Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 96(456), 1348–1360 (2001)MathSciNetMATH Fan, J., Li, R.: Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 96(456), 1348–1360 (2001)MathSciNetMATH
60.
Zurück zum Zitat Zhang, C.H., et al.: Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 38(2), 894–942 (2010)MathSciNetMATH Zhang, C.H., et al.: Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 38(2), 894–942 (2010)MathSciNetMATH
61.
Zurück zum Zitat Collobert, R., Sinz, F., Weston, J., Bottou, L., Joachims, T.: Large scale transductive svms. J. Mach. Learn. Res. 7(8) (2006) Collobert, R., Sinz, F., Weston, J., Bottou, L., Joachims, T.: Large scale transductive svms. J. Mach. Learn. Res. 7(8) (2006)
62.
Zurück zum Zitat Zhen, Y., Yeung, D.Y.: Co-regularized hashing for multimodal data. Adv. Neural Inf. Proces. Syst. 2, 1376 (2012) Zhen, Y., Yeung, D.Y.: Co-regularized hashing for multimodal data. Adv. Neural Inf. Proces. Syst. 2, 1376 (2012)
63.
Zurück zum Zitat de Sousa, E.P., Traina, C., Traina, A.J., Wu, L., Faloutsos, C.: A fast and effective method to find correlations among attributes in databases. Data Min. Knowl. Disc. 14(3), 367–407 (2007)MathSciNet de Sousa, E.P., Traina, C., Traina, A.J., Wu, L., Faloutsos, C.: A fast and effective method to find correlations among attributes in databases. Data Min. Knowl. Disc. 14(3), 367–407 (2007)MathSciNet
64.
Zurück zum Zitat Traina Jr., C., Traina, A., Wu, L., Faloutsos, C.: Fast feature selection using fractal dimension. J. Inf. Data Manag. 1(1), 3–3 (2010) Traina Jr., C., Traina, A., Wu, L., Faloutsos, C.: Fast feature selection using fractal dimension. J. Inf. Data Manag. 1(1), 3–3 (2010)
65.
Zurück zum Zitat Faloutsos, C., Seeger, B., Traina, A., Traina Jr., C.: Spatial join selectivity using power laws. In: Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, pp. 177–188 (2000) Faloutsos, C., Seeger, B., Traina, A., Traina Jr., C.: Spatial join selectivity using power laws. In: Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, pp. 177–188 (2000)
66.
Zurück zum Zitat Lee, H.D., Monard, M.C., Wu, F.C.: A fractal dimension based filter algorithm to select features for supervised learning. In: Advances in Artificial Intelligence-IBERAMIA-SBIA 2006, pp. 278–288. Springer, New York (2006) Lee, H.D., Monard, M.C., Wu, F.C.: A fractal dimension based filter algorithm to select features for supervised learning. In: Advances in Artificial Intelligence-IBERAMIA-SBIA 2006, pp. 278–288. Springer, New York (2006)
67.
Zurück zum Zitat Ni, L.P., Ni, Z.W., Gao, Y.Z.: Stock trend prediction based on fractal feature selection and support vector machine. Expert Syst. Appl. 38(5), 5569–5576 (2011) Ni, L.P., Ni, Z.W., Gao, Y.Z.: Stock trend prediction based on fractal feature selection and support vector machine. Expert Syst. Appl. 38(5), 5569–5576 (2011)
68.
Zurück zum Zitat Pham, D., Packianather, M., Garcia, M., Castellani, M.: Novel feature selection method using mutual information and fractal dimension. In: 2009 35th Annual Conference of IEEE Industrial Electronics, pp. 3393–3398. IEEE, New York (2009) Pham, D., Packianather, M., Garcia, M., Castellani, M.: Novel feature selection method using mutual information and fractal dimension. In: 2009 35th Annual Conference of IEEE Industrial Electronics, pp. 3393–3398. IEEE, New York (2009)
69.
Zurück zum Zitat Schroeder, M.: Fractals, Chaos, Power Laws: Minutes from an Infinite Paradise. Courier Corporation, Chelmsford, MA (2009)MATH Schroeder, M.: Fractals, Chaos, Power Laws: Minutes from an Infinite Paradise. Courier Corporation, Chelmsford, MA (2009)MATH
70.
Zurück zum Zitat Belussi, A., Faloutsos, C.: Estimating the selectivity of spatial queries using the ‘correlation’ fractal dimension. Tech. rep. (1998) Belussi, A., Faloutsos, C.: Estimating the selectivity of spatial queries using the ‘correlation’ fractal dimension. Tech. rep. (1998)
71.
Zurück zum Zitat Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, New York (2006)MATH Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, New York (2006)MATH
72.
Zurück zum Zitat Madden, S.: From databases to big data. IEEE Internet Comput. 16(3), 4–6 (2012) Madden, S.: From databases to big data. IEEE Internet Comput. 16(3), 4–6 (2012)
73.
Zurück zum Zitat Liu, R., Shi, Y.: Spatial distance join based feature selection. Eng. Appl. Artif. Intell. 26(10), 2597–2607 (2013) Liu, R., Shi, Y.: Spatial distance join based feature selection. Eng. Appl. Artif. Intell. 26(10), 2597–2607 (2013)
74.
Zurück zum Zitat Golub, T.R., Slonim, D.K., Tamayo, P., Huard, C., Gaasenbeek, M., Mesirov, J.P., Coller, H., Loh, M.L., Downing, J.R., Caligiuri, M.A., et al.: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 286(5439), 531–537 (1999) Golub, T.R., Slonim, D.K., Tamayo, P., Huard, C., Gaasenbeek, M., Mesirov, J.P., Coller, H., Loh, M.L., Downing, J.R., Caligiuri, M.A., et al.: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 286(5439), 531–537 (1999)
Metadaten
Titel
Feature Selection
verfasst von
Yong Shi
Copyright-Jahr
2022
Verlag
Springer Nature Singapore
DOI
https://doi.org/10.1007/978-981-16-3607-3_4