Skip to main content
Erschienen in: Soft Computing 20/2020

11.04.2020 | Methodologies and Application

A new ensemble feature selection approach based on genetic algorithm

verfasst von: Hongzhi Wang, Chengquan He, Zhuping Li

Erschienen in: Soft Computing | Ausgabe 20/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In the ensemble feature selection method, if the weight adjustment is performed on each feature subset used, the ensemble effect can be significantly different; therefore, how to find the optimized weight vector is a key and challenging problem. Aiming at this optimization problem, this paper proposes an ensemble feature selection approach based on genetic algorithm (EFS-BGA). After each base feature selector generates a feature subset, the EFS-BGA method obtains the optimized weight of each feature subset through genetic algorithm, which is different from traditional genetic algorithm directly processing single features. We divide the EFS-BGA algorithm into two types. The first is a complete ensemble feature selection method; based on the first, we further propose the selective EFS-BGA model. After that, through mathematical analysis, we theoretically explain why weight adjustment is an optimization problem and how to optimize. Finally, through the comparative experiments on multiple data sets, the advantages of the EFS-BGA algorithm in this paper over the previous ensemble feature selection algorithms are explained in practice.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Baraniuk RG (2007) Compressive sensing [lecture notes]. IEEE Signal Process Mag 24(4):118–121CrossRef Baraniuk RG (2007) Compressive sensing [lecture notes]. IEEE Signal Process Mag 24(4):118–121CrossRef
Zurück zum Zitat Buitinck L, Louppe G, Blondel M, Pedregosa F, Mueller A, Grisel O et al (2013) Api design for machine learning software: experiences from the scikit-learn project. Eprint Arxiv Buitinck L, Louppe G, Blondel M, Pedregosa F, Mueller A, Grisel O et al (2013) Api design for machine learning software: experiences from the scikit-learn project. Eprint Arxiv
Zurück zum Zitat Das AK, Das S, Ghosh A (2017) Ensemble feature selection using bi-objective genetic algorithm. Knowl-Based Syst 123:116–127CrossRef Das AK, Das S, Ghosh A (2017) Ensemble feature selection using bi-objective genetic algorithm. Knowl-Based Syst 123:116–127CrossRef
Zurück zum Zitat Fortin F-A, De Rainville F-M, Gardner M-A, Marc P, Christian G (2012) DEAP: evolutionary algorithms made easy. J Mach Learn Res 13:2171–2175MathSciNet Fortin F-A, De Rainville F-M, Gardner M-A, Marc P, Christian G (2012) DEAP: evolutionary algorithms made easy. J Mach Learn Res 13:2171–2175MathSciNet
Zurück zum Zitat Friedman JH (1997) On bias, variance, 0/1–loss, and the curse-of-dimensionality. Data Min Knowl Discov 1(1):55–77MathSciNetCrossRef Friedman JH (1997) On bias, variance, 0/1–loss, and the curse-of-dimensionality. Data Min Knowl Discov 1(1):55–77MathSciNetCrossRef
Zurück zum Zitat Ghamisi P, Benediktsson JA (2015) Feature selection based on hybridization of genetic algorithm and particle swarm optimization. IEEE Geosci Remote Sens Lett 12(2):309–313CrossRef Ghamisi P, Benediktsson JA (2015) Feature selection based on hybridization of genetic algorithm and particle swarm optimization. IEEE Geosci Remote Sens Lett 12(2):309–313CrossRef
Zurück zum Zitat Ghareb AS, Bakar AA, Hamdan AR (2016) Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Syst Appl 49:31–47CrossRef Ghareb AS, Bakar AA, Hamdan AR (2016) Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Syst Appl 49:31–47CrossRef
Zurück zum Zitat Goldberg D (1989) Genetic algorithms in search, optimization and machine learning. Addison-Wesley Professional, Reading, MA. ISBN 978-0201157673 Goldberg D (1989) Genetic algorithms in search, optimization and machine learning. Addison-Wesley Professional, Reading, MA. ISBN 978-0201157673
Zurück zum Zitat Goldberg DE, Holland JH (1988) Genetic algorithms and machine learning. Mach Learn 3(2):95–99CrossRef Goldberg DE, Holland JH (1988) Genetic algorithms and machine learning. Mach Learn 3(2):95–99CrossRef
Zurück zum Zitat Huan L, Hiroshi M (2007) Computational methods of feature selection. Chapman & Hall/CRC Data Mining and Knowledge Discovery Series, Taylor & Francis, New York. Chapman & Hall/CRC Huan L, Hiroshi M (2007) Computational methods of feature selection. Chapman & Hall/CRC Data Mining and Knowledge Discovery Series, Taylor & Francis, New York. Chapman & Hall/CRC
Zurück zum Zitat Jiang S, Chin KS, Wang L et al (2017) Modified genetic algorithm-based feature selection combined with pre-trained deep neural network for demand forecasting in outpatient department. Expert Syst Appl 82:216–230CrossRef Jiang S, Chin KS, Wang L et al (2017) Modified genetic algorithm-based feature selection combined with pre-trained deep neural network for demand forecasting in outpatient department. Expert Syst Appl 82:216–230CrossRef
Zurück zum Zitat Kira K, Rendell L (1992) The feature selection problem: traditional methods and a new algorithm. In: National conference on artificial intelligence Kira K, Rendell L (1992) The feature selection problem: traditional methods and a new algorithm. In: National conference on artificial intelligence
Zurück zum Zitat Kohavi R, John GH (1997) Wrappers for feature subset selection. Artif Intell 97(1–2):273–324MATHCrossRef Kohavi R, John GH (1997) Wrappers for feature subset selection. Artif Intell 97(1–2):273–324MATHCrossRef
Zurück zum Zitat Kononenko I (1994) Estimating attributes: analysis and extensions of RELIEF. In: European conference on machine learning Kononenko I (1994) Estimating attributes: analysis and extensions of RELIEF. In: European conference on machine learning
Zurück zum Zitat Lehmann EL, Casella G (1998) Theory of point estimation (2nd ed.). New York: Springer. ISBN 0-387-98502-6. MR 1639875 Lehmann EL, Casella G (1998) Theory of point estimation (2nd ed.). New York: Springer. ISBN 0-387-98502-6. MR 1639875
Zurück zum Zitat Liu H, Setiono R (1996) Feature selection and classification—a probabilistic wrapper approach. industrial and engineering applications of artificial intelligence and expert systems Liu H, Setiono R (1996) Feature selection and classification—a probabilistic wrapper approach. industrial and engineering applications of artificial intelligence and expert systems
Zurück zum Zitat Mitchell L, Sloan TM, Mewissen M et al (2014) Parallel classification and feature selection in microarray data using SPRINT. Concurr Comput Pract Exp 26(4):854–865CrossRef Mitchell L, Sloan TM, Mewissen M et al (2014) Parallel classification and feature selection in microarray data using SPRINT. Concurr Comput Pract Exp 26(4):854–865CrossRef
Zurück zum Zitat Saeys Y, Abeel T, Peer Y (2008) Robust feature selection using ensemble feature selection techniques. In: European conference on machine learning and knowledge discovery in databases. Springer, Berlin, pp 313–325 Saeys Y, Abeel T, Peer Y (2008) Robust feature selection using ensemble feature selection techniques. In: European conference on machine learning and knowledge discovery in databases. Springer, Berlin, pp 313–325
Zurück zum Zitat Tibshirani R (1996) Regression shrinkage and selection via the lasso. J R Stat Soc Ser B (Methodol) 58(1):267–88. JSTOR 2346178 Tibshirani R (1996) Regression shrinkage and selection via the lasso. J R Stat Soc Ser B (Methodol) 58(1):267–88. JSTOR 2346178
Zurück zum Zitat Weston J, Elisseeff A, Schölkopf B, Tipping M (2003) Use of the zero-norm with linear models and kernel methods. J Mach Learn Res 3(2003):1439–1461MathSciNetMATH Weston J, Elisseeff A, Schölkopf B, Tipping M (2003) Use of the zero-norm with linear models and kernel methods. J Mach Learn Res 3(2003):1439–1461MathSciNetMATH
Zurück zum Zitat Xue X, Yao M, Wu Z (2018) A novel ensemble-based wrapper method for feature selection using extreme learning machine and genetic algorithm. Knowl Inf Syst 57(2):389–412CrossRef Xue X, Yao M, Wu Z (2018) A novel ensemble-based wrapper method for feature selection using extreme learning machine and genetic algorithm. Knowl Inf Syst 57(2):389–412CrossRef
Metadaten
Titel
A new ensemble feature selection approach based on genetic algorithm
verfasst von
Hongzhi Wang
Chengquan He
Zhuping Li
Publikationsdatum
11.04.2020
Verlag
Springer Berlin Heidelberg
Erschienen in
Soft Computing / Ausgabe 20/2020
Print ISSN: 1432-7643
Elektronische ISSN: 1433-7479
DOI
https://doi.org/10.1007/s00500-020-04911-x

Weitere Artikel der Ausgabe 20/2020

Soft Computing 20/2020 Zur Ausgabe

Premium Partner