Skip to main content

2019 | OriginalPaper | Buchkapitel

Using Clustering for Supervised Feature Selection to Detect Relevant Features

verfasst von : Christoph Lohrmann, Pasi Luukka

Erschienen in: Machine Learning, Optimization, and Data Science

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In many applications in machine learning, large quantities of features and information are available, but these can be of low quality. A novel filter method for feature selection for classification termed COLD is presented that uses class-wise clustering to reduce the dimensionality of the data. The idea behind this approach is that if a relevant feature would be removed from the set of features, the separation of clusters belonging to different classes will deteriorate. Four artificial examples and two real-world data sets are presented on which COLD is compared with several popular filter methods. For the artificial examples, only COLD is capable to consistently rank the features according to their contribution to the separation of the classes. For the real-world Dermatology and Arrhythmia dataset, COLD demonstrates the ability to remove a large number of features and improve the classification accuracy or, at a minimum, not degrade the performance considerably.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Bishop, C.M.: Pattern Recognition and Machine Learning. Springer ScienceBusiness Media, New York (2006)MATH Bishop, C.M.: Pattern Recognition and Machine Learning. Springer ScienceBusiness Media, New York (2006)MATH
2.
Zurück zum Zitat Caruana, R., Freitag, D.: Greedy attribute selection. In: Cohen, W., Hirsh, H. (eds.) Proceedings of the 11th International Conference on Machine Learning (ICML 1994), pp. 28–36. Morgan Kaufmann, New Brunswick (1994) Caruana, R., Freitag, D.: Greedy attribute selection. In: Cohen, W., Hirsh, H. (eds.) Proceedings of the 11th International Conference on Machine Learning (ICML 1994), pp. 28–36. Morgan Kaufmann, New Brunswick (1994)
3.
4.
Zurück zum Zitat Chormunge, S., Jena, S.: Correlation based feature selection with clustering for highdimensional data. J. Electr. Syst. Inf. Technol. 5, 542–549 (2018) Chormunge, S., Jena, S.: Correlation based feature selection with clustering for highdimensional data. J. Electr. Syst. Inf. Technol. 5, 542–549 (2018)
5.
Zurück zum Zitat Cover, T.M.: The best two independent measurements are not the two best. IEEE Trans. Syst. Man Cybern. 4(1), 116–117 (1974)CrossRef Cover, T.M.: The best two independent measurements are not the two best. IEEE Trans. Syst. Man Cybern. 4(1), 116–117 (1974)CrossRef
6.
Zurück zum Zitat Dessì, N., Pes, B.: Similarity of feature selection methods. An empirical study across data intensive classification tasks. Expert Syst. Appl. 42(10), 4632–4642 (2015)CrossRef Dessì, N., Pes, B.: Similarity of feature selection methods. An empirical study across data intensive classification tasks. Expert Syst. Appl. 42(10), 4632–4642 (2015)CrossRef
7.
Zurück zum Zitat Duda, R., Hart, P., Stork, D.: Pattern Classification. John Wiley and Sons, New York (2012)MATH Duda, R., Hart, P., Stork, D.: Pattern Classification. John Wiley and Sons, New York (2012)MATH
8.
Zurück zum Zitat Elashoff, J.E., Elashoff, R.M., Goldman, G.E.: On the choice of variables in classification problems with dichotomous variables. Biometrika 54(3), 668–670 (1967)MathSciNetCrossRef Elashoff, J.E., Elashoff, R.M., Goldman, G.E.: On the choice of variables in classification problems with dichotomous variables. Biometrika 54(3), 668–670 (1967)MathSciNetCrossRef
10.
Zurück zum Zitat Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)MATH Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)MATH
12.
Zurück zum Zitat He X., Cai, D., Niyogi, P.: Laplacian score for feature selection. In: Proceedings of the 18th International Conference on Neural Information Processing Systems (NIPS 2005), pp. 507–514. MIT Press, Cambridge (2005) He X., Cai, D., Niyogi, P.: Laplacian score for feature selection. In: Proceedings of the 18th International Conference on Neural Information Processing Systems (NIPS 2005), pp. 507–514. MIT Press, Cambridge (2005)
13.
Zurück zum Zitat Kittler, J., Mardia, K.V.: Statistical pattern recognition in image analysis. J. Appl. Stat. 21(1–2), 61–75 (1994)CrossRef Kittler, J., Mardia, K.V.: Statistical pattern recognition in image analysis. J. Appl. Stat. 21(1–2), 61–75 (1994)CrossRef
15.
Zurück zum Zitat Kononenko, I., Simec, E., Robnik-Sikonja, M.: Overcoming the myopia of inductive learning algorithms with RELIEFF. Appl. Intell. 7, 39–55 (1997)CrossRef Kononenko, I., Simec, E., Robnik-Sikonja, M.: Overcoming the myopia of inductive learning algorithms with RELIEFF. Appl. Intell. 7, 39–55 (1997)CrossRef
17.
Zurück zum Zitat Lohrmann, C., Luukka, P., Jablonska-Sabuka, M., Kauranne, T.: Supervised feature selection with a combination of fuzzy similarity measures and fuzzy entropy measures. Expert Syst. Appl. 110, 216–236 (2018)CrossRef Lohrmann, C., Luukka, P., Jablonska-Sabuka, M., Kauranne, T.: Supervised feature selection with a combination of fuzzy similarity measures and fuzzy entropy measures. Expert Syst. Appl. 110, 216–236 (2018)CrossRef
18.
Zurück zum Zitat Luukka, P.: Feature selection using fuzzy entropy measures with similarity classifier. Expert Syst. Appl. 38, 4600–4607 (2011)CrossRef Luukka, P.: Feature selection using fuzzy entropy measures with similarity classifier. Expert Syst. Appl. 38, 4600–4607 (2011)CrossRef
19.
Zurück zum Zitat Mitra, P., Murthy, C.A., Pal, S.K.: Unsupervised feature selection using feature similarity. IEEE Trans. Pattern Anal. Mach. Intell. 24(3), 301–312 (2002)CrossRef Mitra, P., Murthy, C.A., Pal, S.K.: Unsupervised feature selection using feature similarity. IEEE Trans. Pattern Anal. Mach. Intell. 24(3), 301–312 (2002)CrossRef
20.
Zurück zum Zitat Robnik-Šikonja, M., Kononenko, I.: Theoretical and empirical analysis of ReliefF and RReliefF. Appl. Intell. 53(1–2), 23–69 (2003)MATH Robnik-Šikonja, M., Kononenko, I.: Theoretical and empirical analysis of ReliefF and RReliefF. Appl. Intell. 53(1–2), 23–69 (2003)MATH
21.
Zurück zum Zitat Rousseeuw, P.J.: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20, 53–65 (1987)CrossRef Rousseeuw, P.J.: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20, 53–65 (1987)CrossRef
22.
Zurück zum Zitat Sahu, B., Dehuri, S., Jagadev, A.K.: Feature selection model based on clustering and ranking in pipeline for microarray data. Inf. Med. Unlocked 9, 107–122 (2017)CrossRef Sahu, B., Dehuri, S., Jagadev, A.K.: Feature selection model based on clustering and ranking in pipeline for microarray data. Inf. Med. Unlocked 9, 107–122 (2017)CrossRef
23.
Zurück zum Zitat Sammut, C., Webb, G.I.: Encyclopedia of Machine Learning and Data Mining, 2017th edn. Springer Science+Business Media, New York (2017)CrossRef Sammut, C., Webb, G.I.: Encyclopedia of Machine Learning and Data Mining, 2017th edn. Springer Science+Business Media, New York (2017)CrossRef
24.
Zurück zum Zitat Sotoca, J.M., Pla, F.: Supervised feature selection by clustering using conditional mutual information-based distances. Pattern Recogn. 43, 2068–2081 (2010)CrossRef Sotoca, J.M., Pla, F.: Supervised feature selection by clustering using conditional mutual information-based distances. Pattern Recogn. 43, 2068–2081 (2010)CrossRef
25.
Zurück zum Zitat Toussaint, G.T.: Note on optimal selection of independent binary-valued features for pattern recognition. IEEE Trans. Inf. Theory 17(5), 618 (1971) Toussaint, G.T.: Note on optimal selection of independent binary-valued features for pattern recognition. IEEE Trans. Inf. Theory 17(5), 618 (1971)
26.
Zurück zum Zitat Warton, D.I.: Penalized normal likelihood and ridge regularization of correlation and covariance matrices. J. Am. Stat. Assoc. 103(481), 340–349 (2008)MathSciNetCrossRef Warton, D.I.: Penalized normal likelihood and ridge regularization of correlation and covariance matrices. J. Am. Stat. Assoc. 103(481), 340–349 (2008)MathSciNetCrossRef
Metadaten
Titel
Using Clustering for Supervised Feature Selection to Detect Relevant Features
verfasst von
Christoph Lohrmann
Pasi Luukka
Copyright-Jahr
2019
DOI
https://doi.org/10.1007/978-3-030-37599-7_23