Skip to main content
Top
Published in: Neural Computing and Applications 2/2023

06-10-2022 | Original Article

Learning the transfer function in binary metaheuristic algorithm for feature selection in classification problems

Authors: Zahra Nassiri, Hesam Omranpour

Published in: Neural Computing and Applications | Issue 2/2023

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

One of the most challenging issues in pattern recognition is the data attribution selection process. Feature selection plays a key role in solving problems with high-dimensional data and is a fundamental step in pre-processing many classifications and machine learning problems. The feature selection method reduces the amount of data and increases the category precision. Unrelated data, which can lead to inappropriate classification, is thus removed to obtain fewer features. In this paper, the Binary Gray Wolf Optimization algorithm uses the Wrapper method for feature selection. The transfer function is an essential part of BGWO to map a continuous value to a binary value. In this study, eight transfer functions are divided into two families: S-shaped and V-shaped. Previous research has used only one transfer function for the whole algorithm, and all wolves in the whole algorithm deal with this transfer function. In this paper, each wolf has its own transfer function. Because algorithms are evolutionary meta-innovations and can optimize themselves, each wolf can play a role in the whole algorithm at any stage while optimizing itself and adapting to its community, and not just depend on one transfer function. In the proposed method, eight transfer functions are divided into two families, S-shaped and V-shaped. This article proposes two approaches for learning the transfer function, by selecting the transfer function and the slope of these functions. In the first approach, we add three or two binary bits to the initial population. If two bits are added, four modes of the transfer function are available, and if three bits are added, eight transfer functions are achievable. These bits are used as a criterion for selecting a predefined transfer function for each wolf. So, in the proposed method, each wolf has its transfer function. During the implementation of the algorithm, the wolves update their position according to the evaluation function and learning. In the second approach, ten or twenty-one binary bits are added to the initial population. If ten binary bits are used, we will have a transfer function, and \(2^{10}\) coefficient modes are available for the slope of the transfer function. If twenty-one binary bits are used, we have two transfer functions available. So, there are \(2^{10}\) modes for the gradient of the transfer function. These bits are used as a criterion for selecting the transfer function and the coefficient affecting the slope of these functions. In both ideas, after each iteration of the algorithm, the position of the wolves is updated and based on the evaluation function, the alpha wolf is identified and the transfer function is selected. With subsequent iterations, the algorithm learns and optimizes the transfer function to achieve the best feature selection with the smallest error. Experimental results on ten UCI datasets show that selecting the obtained feature subset with high classification accuracy is efficient.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Ziemba P, Piwowarski M (2011) Feature selection methods in data mining techniques, Prace Naukowe Uniwersytetu Ekonomicznego we Wrocławiu, 213–223 Ziemba P, Piwowarski M (2011) Feature selection methods in data mining techniques, Prace Naukowe Uniwersytetu Ekonomicznego we Wrocławiu, 213–223
2.
go back to reference Chen R-C, Dewi C, Huang S-W, Caraka RE (2020) Selecting critical features for data classification based on machine learning methods. J Big Data 7:1–6CrossRef Chen R-C, Dewi C, Huang S-W, Caraka RE (2020) Selecting critical features for data classification based on machine learning methods. J Big Data 7:1–6CrossRef
3.
go back to reference Zawbaa HM, Emary E, Grosan C, Snasel V (2018) Large-dimensionality small-instance set feature selection: a hybrid bio-inspired heuristic approach. Swarm Evolution Comput 42:29–42CrossRef Zawbaa HM, Emary E, Grosan C, Snasel V (2018) Large-dimensionality small-instance set feature selection: a hybrid bio-inspired heuristic approach. Swarm Evolution Comput 42:29–42CrossRef
4.
go back to reference Zamani H, Nadimi-Shahraki M-H (2016) Feature selection based on whale optimization algorithm for diseases diagnosis. Int J Comput Sci Inform Security 14(9):1243 Zamani H, Nadimi-Shahraki M-H (2016) Feature selection based on whale optimization algorithm for diseases diagnosis. Int J Comput Sci Inform Security 14(9):1243
5.
go back to reference Hsu H-H, Hsieh C-W, Lu M-D (2011) Hybrid feature selection by combining filters and wrappers. Expert Syst Appl 38:8144–8150CrossRef Hsu H-H, Hsieh C-W, Lu M-D (2011) Hybrid feature selection by combining filters and wrappers. Expert Syst Appl 38:8144–8150CrossRef
6.
go back to reference Gangavarapu T, Patil N (2019) A novel filter–wrapper hybrid greedy ensemble approach optimized using the genetic algorithm to reduce the dimensionality of high-dimensional biomedical datasets. Applied Soft Computing 81:105538CrossRef Gangavarapu T, Patil N (2019) A novel filter–wrapper hybrid greedy ensemble approach optimized using the genetic algorithm to reduce the dimensionality of high-dimensional biomedical datasets. Applied Soft Computing 81:105538CrossRef
7.
go back to reference Hoque N, Bhattacharyya DK, Kalita JK (2014) MIFS-ND: a mutual information-based feature selection method. Expert Syst Appl 41(14):6371–6385CrossRef Hoque N, Bhattacharyya DK, Kalita JK (2014) MIFS-ND: a mutual information-based feature selection method. Expert Syst Appl 41(14):6371–6385CrossRef
8.
go back to reference Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybern 43(6):1656–1671CrossRef Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybern 43(6):1656–1671CrossRef
9.
go back to reference Zhang R, Nie F, Li X, Wei X (2019) Feature selection with multiview data: a survey. Inform Fusion 50:158–167CrossRef Zhang R, Nie F, Li X, Wei X (2019) Feature selection with multiview data: a survey. Inform Fusion 50:158–167CrossRef
10.
go back to reference Sánchez-Maroño N, Alonso-Betanzos A, Tombilla-Sanromán M (2007) Filter methods for feature selection–a comparative study. In: Intelligent data engineering and automated learning–ideal, pp 178–187 Sánchez-Maroño N, Alonso-Betanzos A, Tombilla-Sanromán M (2007) Filter methods for feature selection–a comparative study. In: Intelligent data engineering and automated learning–ideal, pp 178–187
11.
go back to reference Nnamoko NA, Arshad FN, England D, Vora J, Norman J (2014) Evaluation of filter and wrapper methods for feature selection in supervised machine learning. In: Conference: the 15th annual postgraduate symposium on the convergence of telecommunication, networking and broadcasting at: liverpool Nnamoko NA, Arshad FN, England D, Vora J, Norman J (2014) Evaluation of filter and wrapper methods for feature selection in supervised machine learning. In: Conference: the 15th annual postgraduate symposium on the convergence of telecommunication, networking and broadcasting at: liverpool
12.
go back to reference Cai J, Luo J, Wang S, Yang S (2018) Feature selection in machine learning: a new perspective. Neurocomputing 300:70–79CrossRef Cai J, Luo J, Wang S, Yang S (2018) Feature selection in machine learning: a new perspective. Neurocomputing 300:70–79CrossRef
13.
go back to reference Venkatesh B, Anuradha J (2019) A review of feature selection and its methods. Cybern Inform Technol 19(1):3–26MathSciNet Venkatesh B, Anuradha J (2019) A review of feature selection and its methods. Cybern Inform Technol 19(1):3–26MathSciNet
14.
go back to reference Wu Y, Liu Y, Wang Y, Shi Y, Zhao X (2018) JCDSA: a joint covariate detection tool for survival analysis on tumor expression profiles. BMC Bioinform 19(1):1–8CrossRef Wu Y, Liu Y, Wang Y, Shi Y, Zhao X (2018) JCDSA: a joint covariate detection tool for survival analysis on tumor expression profiles. BMC Bioinform 19(1):1–8CrossRef
16.
go back to reference Masoudi Sobhanzadeh Y, Motieghader H, Masoudi-Nejad A (2019) Feature Select: a software for feature selection based on machine learning approaches. BMC Bioinform 20(170):1–17 Masoudi Sobhanzadeh Y, Motieghader H, Masoudi-Nejad A (2019) Feature Select: a software for feature selection based on machine learning approaches. BMC Bioinform 20(170):1–17
17.
go back to reference Metin SK (2018) Feature selection in multiword expression recognition. Expert Syst Appl 92(C):106–123CrossRef Metin SK (2018) Feature selection in multiword expression recognition. Expert Syst Appl 92(C):106–123CrossRef
18.
go back to reference Saxena AK, Dubey VK, Wang J (2017) Hybrid feature selection methods for high-dimensional multi-class datasets. Int J Data Min Modell Manag 9(4):315 Saxena AK, Dubey VK, Wang J (2017) Hybrid feature selection methods for high-dimensional multi-class datasets. Int J Data Min Modell Manag 9(4):315
19.
go back to reference Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A (2012) An ensemble of filters and classifiers for microarray data classification. Pattern Recogn 45(1):531–539CrossRef Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A (2012) An ensemble of filters and classifiers for microarray data classification. Pattern Recogn 45(1):531–539CrossRef
20.
go back to reference Hoque N, Singh M, Bhattacharyya DK (2018) EFS-MI: an ensemble feature selection method for classification. Complex Intell Syst 4:105–118CrossRef Hoque N, Singh M, Bhattacharyya DK (2018) EFS-MI: an ensemble feature selection method for classification. Complex Intell Syst 4:105–118CrossRef
21.
go back to reference Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A, Benitez JM, Herrera F (2014) A review of microarray datasets and applied feature selection methods. Inf Sci 282:111–135CrossRef Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A, Benitez JM, Herrera F (2014) A review of microarray datasets and applied feature selection methods. Inf Sci 282:111–135CrossRef
22.
go back to reference Agrawal P, Abutarboush HF, Ganesh T, Mohamed AW (2021) Metaheuristic Algorithms on feature selection: A survey of one decade of research (2009–2019). IEEE Access 9:26766–26791CrossRef Agrawal P, Abutarboush HF, Ganesh T, Mohamed AW (2021) Metaheuristic Algorithms on feature selection: A survey of one decade of research (2009–2019). IEEE Access 9:26766–26791CrossRef
23.
go back to reference Jiang Y, Luo Q, Wei Y, Abualigah L, Zhou Y (2021) An efficient binary Gradient-based optimizer for feature selection. Math Biosci Eng 18(4):3813–3854CrossRefMATH Jiang Y, Luo Q, Wei Y, Abualigah L, Zhou Y (2021) An efficient binary Gradient-based optimizer for feature selection. Math Biosci Eng 18(4):3813–3854CrossRefMATH
24.
go back to reference Emary E, Zawbaa HM, Hassanien AE (2016) Binary grey wolf optimization approaches for feature selection. Neuro Comput 172:371–381 Emary E, Zawbaa HM, Hassanien AE (2016) Binary grey wolf optimization approaches for feature selection. Neuro Comput 172:371–381
25.
go back to reference Mirjalili S, Lewis A (2013) S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol Comput 9:114CrossRef Mirjalili S, Lewis A (2013) S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol Comput 9:114CrossRef
26.
go back to reference Mirjalili S, Yang X-S, Mirjalili SM (2014) Binary bat algorithm. Neurl Comput Appl 25(3):663–681CrossRef Mirjalili S, Yang X-S, Mirjalili SM (2014) Binary bat algorithm. Neurl Comput Appl 25(3):663–681CrossRef
27.
go back to reference Mirjalili S, Mirjalili SM, Lewis A (2014) Grey wolf optimizer. Adv Eng Softw 69:46–61CrossRef Mirjalili S, Mirjalili SM, Lewis A (2014) Grey wolf optimizer. Adv Eng Softw 69:46–61CrossRef
28.
go back to reference Zhang J, Hong L, Liu Q (2021) An improved whale optimization algorithm for the traveling salesman problem. Symmetry 13(1):48CrossRef Zhang J, Hong L, Liu Q (2021) An improved whale optimization algorithm for the traveling salesman problem. Symmetry 13(1):48CrossRef
29.
go back to reference Hussien AG, Houssein EH, Hassanien AE (2017) A binary whale optimization algorithm with hyperbolic tangent fitness function for feature selection. In: Eighth international conference on intelligent computing and information systems (ICICIS) Hussien AG, Houssein EH, Hassanien AE (2017) A binary whale optimization algorithm with hyperbolic tangent fitness function for feature selection. In: Eighth international conference on intelligent computing and information systems (ICICIS)
30.
go back to reference Hu P, Pan JS, Chu SC (2020) Improved Binary Grey Wolf Optimizer and Its application for a feature selection. Knowledg Based Syst 195:105746CrossRef Hu P, Pan JS, Chu SC (2020) Improved Binary Grey Wolf Optimizer and Its application for a feature selection. Knowledg Based Syst 195:105746CrossRef
31.
go back to reference Altman NS (1992) An introduction to kernel and nearest-neighbor nonpara-metric regression. Am Statist 46(3):175–185 Altman NS (1992) An introduction to kernel and nearest-neighbor nonpara-metric regression. Am Statist 46(3):175–185
32.
go back to reference Ghosh M et al (2019) Genetic algorithm based cancerous gene identification from microarray data using ensemble of filter methods. Med Biol Eng Comput 57(1):159–176CrossRef Ghosh M et al (2019) Genetic algorithm based cancerous gene identification from microarray data using ensemble of filter methods. Med Biol Eng Comput 57(1):159–176CrossRef
33.
go back to reference Chellammal S, Sharmila R (2019) Recommendation of attributes for heart disease prediction using correlation measure. Int J Recent Technol Eng (IJRTE) 8(23):870–875 Chellammal S, Sharmila R (2019) Recommendation of attributes for heart disease prediction using correlation measure. Int J Recent Technol Eng (IJRTE) 8(23):870–875
Metadata
Title
Learning the transfer function in binary metaheuristic algorithm for feature selection in classification problems
Authors
Zahra Nassiri
Hesam Omranpour
Publication date
06-10-2022
Publisher
Springer London
Published in
Neural Computing and Applications / Issue 2/2023
Print ISSN: 0941-0643
Electronic ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-022-07869-z

Other articles of this Issue 2/2023

Neural Computing and Applications 2/2023 Go to the issue

Premium Partner