Skip to main content
Erschienen in: Pattern Analysis and Applications 2/2023

17.02.2023 | Theoretical Advances

Feature selection using max dynamic relevancy and min redundancy

verfasst von: Kexin Yin, Junren Zhai, Aifeng Xie, Jianqi Zhu

Erschienen in: Pattern Analysis and Applications | Ausgabe 2/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Feature selection algorithms based on three-way interaction information have been widely studied. However, most of these traditional algorithms only consider class-dependent redundancy, which can lead to an underestimation of redundancy. To address this issue, a feature selection algorithm based on maximum dynamic relevancy minimum redundancy is proposed. The algorithm first proposes a quality coefficient to estimate the feature relevancy. Then we introduce class-independent redundancy to solve the issue of not fully considering redundancy, and propose adaptive coefficients to optimize the algorithm. To ensure the effectiveness of the algorithm, experimental comparisons are carried on 19 benchmark data sets with six algorithms. The results show that the proposed algorithm outperforms other algorithms in terms of performance.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Bahassine S, Madani A, Al-Sarem M, Kissi M (2020) Feature selection using an improved chi-square for arabic text classification. J King Saud Univ Comput Inform Sci 32(2):225–231 Bahassine S, Madani A, Al-Sarem M, Kissi M (2020) Feature selection using an improved chi-square for arabic text classification. J King Saud Univ Comput Inform Sci 32(2):225–231
2.
Zurück zum Zitat Ansari G, Ahmad T, Doja MN (2019) Hybrid filter-wrapper feature selection method for sentiment classification. Arab J Sci Eng 44(11):9191–9208CrossRef Ansari G, Ahmad T, Doja MN (2019) Hybrid filter-wrapper feature selection method for sentiment classification. Arab J Sci Eng 44(11):9191–9208CrossRef
3.
Zurück zum Zitat Selvakumar B, Muneeswaran K (2019) Firefly algorithm based feature selection for network intrusion detection. Comput Secur 81:148–155CrossRef Selvakumar B, Muneeswaran K (2019) Firefly algorithm based feature selection for network intrusion detection. Comput Secur 81:148–155CrossRef
4.
Zurück zum Zitat Mohammadi S, Mirvaziri H, Ghazizadeh-Ahsaee M, Karimipour H (2019) Cyber intrusion detection by combined feature selection algorithm. J Inform Secur Appl 44:80–88 Mohammadi S, Mirvaziri H, Ghazizadeh-Ahsaee M, Karimipour H (2019) Cyber intrusion detection by combined feature selection algorithm. J Inform Secur Appl 44:80–88
5.
Zurück zum Zitat Ghosh M, Adhikary S, Ghosh KK, Sardar A, Begum S, Sarkar R (2019) Genetic algorithm based cancerous gene identification from microarray data using ensemble of filter methods. Med Biol Eng Comput 57(1):159–176CrossRef Ghosh M, Adhikary S, Ghosh KK, Sardar A, Begum S, Sarkar R (2019) Genetic algorithm based cancerous gene identification from microarray data using ensemble of filter methods. Med Biol Eng Comput 57(1):159–176CrossRef
6.
Zurück zum Zitat Abdulla M, Khasawneh MT (2020) G-forest: an ensemble method for cost-sensitive feature selection in gene expression microarrays. Artif Intell Med 108:101941CrossRef Abdulla M, Khasawneh MT (2020) G-forest: an ensemble method for cost-sensitive feature selection in gene expression microarrays. Artif Intell Med 108:101941CrossRef
7.
Zurück zum Zitat Sawant SS, Manoharan P (2020) Unsupervised band selection based on weighted information entropy and 3d discrete cosine transform for hyperspectral image classification. Int J Remote Sens 41(10):3948–3969CrossRef Sawant SS, Manoharan P (2020) Unsupervised band selection based on weighted information entropy and 3d discrete cosine transform for hyperspectral image classification. Int J Remote Sens 41(10):3948–3969CrossRef
8.
Zurück zum Zitat Sheikhpour R, Sarram MA, Gharaghani S, Chahooki MAZ (2017) A survey on semi-supervised feature selection methods. Pattern Recogn 64:141–158CrossRefMATH Sheikhpour R, Sarram MA, Gharaghani S, Chahooki MAZ (2017) A survey on semi-supervised feature selection methods. Pattern Recogn 64:141–158CrossRefMATH
9.
Zurück zum Zitat Che J, Yang Y, Li L, Bai X, Zhang S, Deng C (2017) Maximum relevance minimum common redundancy feature selection for nonlinear data. Inf Sci 409:68–86CrossRefMATH Che J, Yang Y, Li L, Bai X, Zhang S, Deng C (2017) Maximum relevance minimum common redundancy feature selection for nonlinear data. Inf Sci 409:68–86CrossRefMATH
10.
Zurück zum Zitat Ang JC, Mirzal A, Haron H, Hamed HNA (2015) Supervised, unsupervised, and semi-supervised feature selection: a review on gene selection. IEEE/ACM Trans Comput Biol Bioinf 13(5):971–989CrossRef Ang JC, Mirzal A, Haron H, Hamed HNA (2015) Supervised, unsupervised, and semi-supervised feature selection: a review on gene selection. IEEE/ACM Trans Comput Biol Bioinf 13(5):971–989CrossRef
11.
Zurück zum Zitat Sun X, Liu Y, Xu M, Chen H, Han J, Wang K (2013) Feature selection using dynamic weights for classification. Knowl-Based Syst 37:541–549CrossRef Sun X, Liu Y, Xu M, Chen H, Han J, Wang K (2013) Feature selection using dynamic weights for classification. Knowl-Based Syst 37:541–549CrossRef
12.
Zurück zum Zitat Zeng Z, Zhang H, Zhang R, Yin C (2015) A novel feature selection method considering feature interaction. Pattern Recogn 48(8):2656–2666CrossRef Zeng Z, Zhang H, Zhang R, Yin C (2015) A novel feature selection method considering feature interaction. Pattern Recogn 48(8):2656–2666CrossRef
13.
Zurück zum Zitat Hu L, Gao W, Zhao K, Zhang P, Wang F (2018) Feature selection considering two types of feature relevancy and feature interdependency. Expert Syst Appl 93:423–434CrossRef Hu L, Gao W, Zhao K, Zhang P, Wang F (2018) Feature selection considering two types of feature relevancy and feature interdependency. Expert Syst Appl 93:423–434CrossRef
14.
Zurück zum Zitat Wang L, Jiang S, Jiang S (2021) A feature selection method via analysis of relevance, redundancy, and interaction. Expert Syst Appl 183:115365CrossRef Wang L, Jiang S, Jiang S (2021) A feature selection method via analysis of relevance, redundancy, and interaction. Expert Syst Appl 183:115365CrossRef
15.
Zurück zum Zitat Herman G, Zhang B, Wang Y, Ye G, Chen F (2013) Mutual information-based method for selecting informative feature sets. Pattern Recogn 46(12):3315–3327CrossRef Herman G, Zhang B, Wang Y, Ye G, Chen F (2013) Mutual information-based method for selecting informative feature sets. Pattern Recogn 46(12):3315–3327CrossRef
16.
Zurück zum Zitat Sun Y, Todorovic S, Goodison S (2009) Local-learning-based feature selection for high-dimensional data analysis. IEEE Trans Pattern Anal Mach Intell 32(9):1610–1626 Sun Y, Todorovic S, Goodison S (2009) Local-learning-based feature selection for high-dimensional data analysis. IEEE Trans Pattern Anal Mach Intell 32(9):1610–1626
18.
Zurück zum Zitat Vergara JR, Estévez PA (2014) A review of feature selection methods based on mutual information. Neural Comput Appl 24(1):175–186CrossRef Vergara JR, Estévez PA (2014) A review of feature selection methods based on mutual information. Neural Comput Appl 24(1):175–186CrossRef
19.
Zurück zum Zitat Jakulin A, Bratko I (2004) Testing the significance of attribute interactions. In: Proceedings of the Twenty-first International Conference on Machine Learning, p. 52 Jakulin A, Bratko I (2004) Testing the significance of attribute interactions. In: Proceedings of the Twenty-first International Conference on Machine Learning, p. 52
20.
Zurück zum Zitat Jakulin A (2005) Machine learning based on attribute interactions. Univerza v Ljubljani, Thesis Jakulin A (2005) Machine learning based on attribute interactions. Univerza v Ljubljani, Thesis
21.
Zurück zum Zitat Dash M, Liu H (1997) Feature selection for classification. Intell Data Anal 1(1–4):131–156CrossRef Dash M, Liu H (1997) Feature selection for classification. Intell Data Anal 1(1–4):131–156CrossRef
22.
Zurück zum Zitat Zhao J, Zhou Y, Zhang X, Chen L (2016) Part mutual information for quantifying direct associations in networks. Proc Natl Acad Sci 113(18):5130–5135CrossRef Zhao J, Zhou Y, Zhang X, Chen L (2016) Part mutual information for quantifying direct associations in networks. Proc Natl Acad Sci 113(18):5130–5135CrossRef
23.
Zurück zum Zitat Lewis DD (1992) Feature selection and feature extraction for text categorization. In: HLT ’91: Proceedings of the Workshop on Speech and Natural Language, pp. 212–217 Lewis DD (1992) Feature selection and feature extraction for text categorization. In: HLT ’91: Proceedings of the Workshop on Speech and Natural Language, pp. 212–217
24.
Zurück zum Zitat Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550CrossRef Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550CrossRef
25.
Zurück zum Zitat Peng H, Long F, Ding C (2005) Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238CrossRef Peng H, Long F, Ding C (2005) Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238CrossRef
26.
Zurück zum Zitat Fleuret F (2004) Fast binary feature selection with conditional mutual information. J Mach Learn Res 5(9):1531–1555MathSciNetMATH Fleuret F (2004) Fast binary feature selection with conditional mutual information. J Mach Learn Res 5(9):1531–1555MathSciNetMATH
27.
Zurück zum Zitat Bennasar M, Hicks Y, Setchi R (2015) Feature selection using joint mutual information maximisation. Expert Syst Appl 42(22):8520–8532CrossRef Bennasar M, Hicks Y, Setchi R (2015) Feature selection using joint mutual information maximisation. Expert Syst Appl 42(22):8520–8532CrossRef
28.
Zurück zum Zitat Qu G, Hariri S, Yousif M (2005) A new dependency and correlation analysis for features. IEEE Trans Knowl Data Eng 17(9):1199–1207CrossRef Qu G, Hariri S, Yousif M (2005) A new dependency and correlation analysis for features. IEEE Trans Knowl Data Eng 17(9):1199–1207CrossRef
29.
Zurück zum Zitat Gao W, Hu L, Zhang P (2018) Class-specific mutual information variation for feature selection. Pattern Recogn 79:328–339CrossRef Gao W, Hu L, Zhang P (2018) Class-specific mutual information variation for feature selection. Pattern Recogn 79:328–339CrossRef
30.
Zurück zum Zitat Zhang P, Gao W (2020) Feature selection considering uncertainty change ratio of the class label. Appl Soft Comput 95:106537CrossRef Zhang P, Gao W (2020) Feature selection considering uncertainty change ratio of the class label. Appl Soft Comput 95:106537CrossRef
31.
Zurück zum Zitat Yu L, Liu H (2004) Efficient feature selection via analysis of relevance and redundancy. J Mach Learn Res 5:1205–1224MathSciNetMATH Yu L, Liu H (2004) Efficient feature selection via analysis of relevance and redundancy. J Mach Learn Res 5:1205–1224MathSciNetMATH
32.
Zurück zum Zitat Kurgan LA, Cios KJ (2004) Caim discretization algorithm. IEEE Trans Knowl Data Eng 16(2):145–153CrossRef Kurgan LA, Cios KJ (2004) Caim discretization algorithm. IEEE Trans Knowl Data Eng 16(2):145–153CrossRef
Metadaten
Titel
Feature selection using max dynamic relevancy and min redundancy
verfasst von
Kexin Yin
Junren Zhai
Aifeng Xie
Jianqi Zhu
Publikationsdatum
17.02.2023
Verlag
Springer London
Erschienen in
Pattern Analysis and Applications / Ausgabe 2/2023
Print ISSN: 1433-7541
Elektronische ISSN: 1433-755X
DOI
https://doi.org/10.1007/s10044-023-01138-y

Weitere Artikel der Ausgabe 2/2023

Pattern Analysis and Applications 2/2023 Zur Ausgabe

Premium Partner