Skip to main content
Erschienen in: Artificial Intelligence Review 1/2019

27.03.2019

A comparative study of neural-network feature weighting

verfasst von: Tongfeng Sun, Shifei Ding, Pin Li, Wei Chen

Erschienen in: Artificial Intelligence Review | Ausgabe 1/2019

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Many feature weighting methods have been proposed to evaluate feature saliencies in recent years. Neural-network (NN) feature weighting, as a supervised method, is founded upon the mapping from input features to output decisions, and implemented by evaluating the sensitivity of network outputs to its inputs. Through training on sample data, NN implicitly embodies the saliencies of input features. The partial derivatives of the outputs with respect to the inputs in the trained NN are calculated to measure their sensitivities to input features, which means that implicit feature weighting of the NN is transformed into explicit feature weighting. The purpose of this paper is to further probe into the principle of NN feature weighting, and evaluate its performance through a comparative study between NN feature weighting method and state-of-art weighting methods in the same working conditions. The motivation of this study is inspired by the lack of direct and comprehensive comparison studies of NN feature weighting method. Experiments in UCI repository data sets, face data sets and self-built data sets show that NN feature weighting method achieves superior performance in different conditions and has promising prospects. Compared with the other existing methods, NN feature weighting method can be used in more complex conditions, provided that NN can work in those conditions. As decision data, output data can be labels, reals or integers. Especially, feature weights can be calculated without the discretization of outputs in the condition of continuous outputs.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Bartlett MS, Movellan JR, Sejnowski TJ (2002) Face recognition by independent component analysis. IEEE Trans Neural Netw Publ IEEE Neural Netw Counc 13(6):1450–1464CrossRef Bartlett MS, Movellan JR, Sejnowski TJ (2002) Face recognition by independent component analysis. IEEE Trans Neural Netw Publ IEEE Neural Netw Counc 13(6):1450–1464CrossRef
Zurück zum Zitat Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550CrossRef Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550CrossRef
Zurück zum Zitat Belhumeur PN, Hespanha JP, and Kriegman DJ (1996) Eigenfaces vs. fisherfaces: recognition using class specific linear projection. In: European conference on computer vision, Berlin, Heidelberg, pp 43–58 Belhumeur PN, Hespanha JP, and Kriegman DJ (1996) Eigenfaces vs. fisherfaces: recognition using class specific linear projection. In: European conference on computer vision, Berlin, Heidelberg, pp 43–58
Zurück zum Zitat Cament LA, Castillo LE, Perez JP, Galdames FJ, Perez CA (2014) Fusion of local normalization and Gabor entropy weighted features for face identification. Pattern Recognit 47(2):568–577CrossRef Cament LA, Castillo LE, Perez JP, Galdames FJ, Perez CA (2014) Fusion of local normalization and Gabor entropy weighted features for face identification. Pattern Recognit 47(2):568–577CrossRef
Zurück zum Zitat Chandrashekar G, Sahin F (2014) A survey on feature selection methods. Comput Electr Eng 40(1):16–28CrossRef Chandrashekar G, Sahin F (2014) A survey on feature selection methods. Comput Electr Eng 40(1):16–28CrossRef
Zurück zum Zitat Cover T, Hart P (1967) Nearest neighbor pattern classification. IEEE Trans Inf Theory 13(1):21–27CrossRefMATH Cover T, Hart P (1967) Nearest neighbor pattern classification. IEEE Trans Inf Theory 13(1):21–27CrossRefMATH
Zurück zum Zitat Cristianini N, Shawe-Taylor J (2000) An introduction to support vector machines and other kernel-based learning methods. Cambridge University Press, CambridgeCrossRefMATH Cristianini N, Shawe-Taylor J (2000) An introduction to support vector machines and other kernel-based learning methods. Cambridge University Press, CambridgeCrossRefMATH
Zurück zum Zitat Delchambre L (2014) Weighted principal component analysis: a weighted covariance eigendecomposition approach. Mon Not R Astron Soc 446(4):3545–3555CrossRef Delchambre L (2014) Weighted principal component analysis: a weighted covariance eigendecomposition approach. Mon Not R Astron Soc 446(4):3545–3555CrossRef
Zurück zum Zitat Diakoulaki D, Mavrotas G, Papayannakis L (1995) Determining objective weights in multiple criteria problems: the CRITIC method. Comput Oper Res 22(7):763–770CrossRefMATH Diakoulaki D, Mavrotas G, Papayannakis L (1995) Determining objective weights in multiple criteria problems: the CRITIC method. Comput Oper Res 22(7):763–770CrossRefMATH
Zurück zum Zitat Duan B, Pao Y-H (2006) Iterative feature weighting with neural networks. US Patents Duan B, Pao Y-H (2006) Iterative feature weighting with neural networks. US Patents
Zurück zum Zitat Duda RO, Hart PE, Stork DG (2001) Pattern classification, 2nd edn. Wiley, NewYorkMATH Duda RO, Hart PE, Stork DG (2001) Pattern classification, 2nd edn. Wiley, NewYorkMATH
Zurück zum Zitat Gilad-Bachrach R, Navot A, Tishby N (2004) Margin based feature selection-theory and algorithms. In: Proceedings of the twenty-first international conference on machine learning, pp 43–50 Gilad-Bachrach R, Navot A, Tishby N (2004) Margin based feature selection-theory and algorithms. In: Proceedings of the twenty-first international conference on machine learning, pp 43–50
Zurück zum Zitat Hocke J, Martinetz T (2013) Feature weighting by maximum distance minimization. In: International conference on artificial neural networks, pp 420–425 Hocke J, Martinetz T (2013) Feature weighting by maximum distance minimization. In: International conference on artificial neural networks, pp 420–425
Zurück zum Zitat Hocke J, Martinetz T (2015) Maximum distance minimization for feature weighting. Pattern Recogn Lett 52:48–52CrossRef Hocke J, Martinetz T (2015) Maximum distance minimization for feature weighting. Pattern Recogn Lett 52:48–52CrossRef
Zurück zum Zitat Huang JZ, Ng MK, Rong H, Li Z (2005) Automated variable weighting in k-means type clustering. IEEE Trans Pattern Anal Mach Intell 27(5):657–668CrossRef Huang JZ, Ng MK, Rong H, Li Z (2005) Automated variable weighting in k-means type clustering. IEEE Trans Pattern Anal Mach Intell 27(5):657–668CrossRef
Zurück zum Zitat Huang G-B, Zhu Q-Y, Siew C-K (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501CrossRef Huang G-B, Zhu Q-Y, Siew C-K (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501CrossRef
Zurück zum Zitat Jing L, Ng MK, Huang JZ (2007) An entropy weighting k-means algorithm for subspace clustering of high-dimensional sparse data. IEEE Trans Knowl Data Eng 19(8):1026–1041CrossRef Jing L, Ng MK, Huang JZ (2007) An entropy weighting k-means algorithm for subspace clustering of high-dimensional sparse data. IEEE Trans Knowl Data Eng 19(8):1026–1041CrossRef
Zurück zum Zitat Kaski S (1998) Dimensionality reduction by random mapping: fast similarity computation for clustering. In: Neural networks proceedings, 1998, The 1998 IEEE international joint conference on IEEE world congress on computational intelligence, vol 1, pp 413–418 Kaski S (1998) Dimensionality reduction by random mapping: fast similarity computation for clustering. In: Neural networks proceedings, 1998, The 1998 IEEE international joint conference on IEEE world congress on computational intelligence, vol 1, pp 413–418
Zurück zum Zitat Kira K, Rendell LA (1992) A practical approach to feature selection. Mach Learn Proc 1992:249–256 Kira K, Rendell LA (1992) A practical approach to feature selection. Mach Learn Proc 1992:249–256
Zurück zum Zitat Kononenko I, Šimec E, Robnik-Šikonja M (1997) Overcoming the myopia of inductive learning algorithms with RELIEFF. Appl Intell 7(1):39–55CrossRef Kononenko I, Šimec E, Robnik-Šikonja M (1997) Overcoming the myopia of inductive learning algorithms with RELIEFF. Appl Intell 7(1):39–55CrossRef
Zurück zum Zitat Krizhevsky A, Sutskever I, Hinton GE (2012). Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems, pp 1097–1105 Krizhevsky A, Sutskever I, Hinton GE (2012). Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems, pp 1097–1105
Zurück zum Zitat Li M, Zhang T, Chen Y, Smola AJ (2014) Efficient mini-batch training for stochastic optimization. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining, pp 661–670 Li M, Zhang T, Chen Y, Smola AJ (2014) Efficient mini-batch training for stochastic optimization. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining, pp 661–670
Zurück zum Zitat Nie F, Xiang S, Jia Y, Zhang C, Yan S (2008) Trace ratio criterion for feature selection. Assoc Adv Artif Intell 2:671–676 Nie F, Xiang S, Jia Y, Zhang C, Yan S (2008) Trace ratio criterion for feature selection. Assoc Adv Artif Intell 2:671–676
Zurück zum Zitat Nie F, Huang H, Cai X, Ding CH (2010) Efficient and robust feature selection via joint ℓ2, 1-norms minimization. In: Advances in neural information processing systems, pp 1813–1821 Nie F, Huang H, Cai X, Ding CH (2010) Efficient and robust feature selection via joint ℓ2, 1-norms minimization. In: Advances in neural information processing systems, pp 1813–1821
Zurück zum Zitat Peng L, Zhang H, Zhang H, Yang B (2017) A fast feature weighting algorithm of data gravitation classification. Inf Sci 375:54–78CrossRef Peng L, Zhang H, Zhang H, Yang B (2017) A fast feature weighting algorithm of data gravitation classification. Inf Sci 375:54–78CrossRef
Zurück zum Zitat Powell M (1990) The theory of RBF approximation in 1990, numerical analysis report. University of Cambridge, Cambridge Powell M (1990) The theory of RBF approximation in 1990, numerical analysis report. University of Cambridge, Cambridge
Zurück zum Zitat Robnik-Šikonja M, Kononenko I (2003) Theoretical and empirical analysis of ReliefF and RReliefF. Mach Learn 53(1–2):23–69CrossRefMATH Robnik-Šikonja M, Kononenko I (2003) Theoretical and empirical analysis of ReliefF and RReliefF. Mach Learn 53(1–2):23–69CrossRefMATH
Zurück zum Zitat Romero E, Sopena JM, Navarrete G, Alquézar R (2003) Feature selection forcing overtraining may help to improve performance. Proc Int Jt Conf Neural Netw 3:2181–2186 Romero E, Sopena JM, Navarrete G, Alquézar R (2003) Feature selection forcing overtraining may help to improve performance. Proc Int Jt Conf Neural Netw 3:2181–2186
Zurück zum Zitat Ruck DW, Rogers SK, Kabrisky M (1990) Feature selection using a multilayer perceptron. J Neural Netw Comput 2(2):40–48 Ruck DW, Rogers SK, Kabrisky M (1990) Feature selection using a multilayer perceptron. J Neural Netw Comput 2(2):40–48
Zurück zum Zitat Rumelhart DE, Hinton GE, Williams RJ (1986) Learning internal representations by error propagation. In: Parallel distributed processing: explorations in the microstructure of cognition, vol 1. California Univ San Diego La Jolla Inst for Cognitive Science, pp 318–362 Rumelhart DE, Hinton GE, Williams RJ (1986) Learning internal representations by error propagation. In: Parallel distributed processing: explorations in the microstructure of cognition, vol 1. California Univ San Diego La Jolla Inst for Cognitive Science, pp 318–362
Zurück zum Zitat Samaria FS, Harter AC (1994) Parameterisation of a stochastic model for human face identification. In: Proceedings of the second IEEE workshop on applications of computer vision, pp 138–142 Samaria FS, Harter AC (1994) Parameterisation of a stochastic model for human face identification. In: Proceedings of the second IEEE workshop on applications of computer vision, pp 138–142
Zurück zum Zitat Sarikaya R, Hinton GE, Deoras A (2014) Application of deep belief networks for natural language understanding. IEEE/ACM Trans Audio Speech Lang Process (TASLP) 22(4):778–784CrossRef Sarikaya R, Hinton GE, Deoras A (2014) Application of deep belief networks for natural language understanding. IEEE/ACM Trans Audio Speech Lang Process (TASLP) 22(4):778–784CrossRef
Zurück zum Zitat Wei G-W (2011) Grey relational analysis method for 2-tuple linguistic multiple attribute group decision making with incomplete weight information. Expert Syst Appl 38(5):4824–4828CrossRef Wei G-W (2011) Grey relational analysis method for 2-tuple linguistic multiple attribute group decision making with incomplete weight information. Expert Syst Appl 38(5):4824–4828CrossRef
Zurück zum Zitat Weinberger KQ, Saul LK (2009) Distance metric learning for large margin nearest neighbor classification. J Mach Learn Res 10(Feb):207–244MATH Weinberger KQ, Saul LK (2009) Distance metric learning for large margin nearest neighbor classification. J Mach Learn Res 10(Feb):207–244MATH
Zurück zum Zitat Wold S, Esbensen K, Geladi P (1987) Principal component analysis. Chemometr Intell Lab Syst 2(1–3):37–52CrossRef Wold S, Esbensen K, Geladi P (1987) Principal component analysis. Chemometr Intell Lab Syst 2(1–3):37–52CrossRef
Zurück zum Zitat Xia B, Bao C (2013). Speech enhancement with weighted denoising auto-encoder. In: Proceedings of INTERSPEECH, pp 3444–3448 Xia B, Bao C (2013). Speech enhancement with weighted denoising auto-encoder. In: Proceedings of INTERSPEECH, pp 3444–3448
Zurück zum Zitat Yan H, Yang J (2016) Sparse discriminative feature weights learning. Neurocomputing 173:1936–1942CrossRef Yan H, Yang J (2016) Sparse discriminative feature weights learning. Neurocomputing 173:1936–1942CrossRef
Zurück zum Zitat Yingming W (1997) Using the method of maximizing deviation to make decision for multiindices. J Syst Eng Electr 8(3):21–26 Yingming W (1997) Using the method of maximizing deviation to make decision for multiindices. J Syst Eng Electr 8(3):21–26
Zurück zum Zitat Yu Z-J, Hu X-P, Mao Q (2009) Novel credit rating method under electronic commerce. Control Decis 11(24):1668–1672 Yu Z-J, Hu X-P, Mao Q (2009) Novel credit rating method under electronic commerce. Control Decis 11(24):1668–1672
Zurück zum Zitat Zhu L, Miao L, Zhang D (2012) Iterative Laplacian score for feature selection. CCPR 2012. Commun Comput Inf Sci 321:80–87 Zhu L, Miao L, Zhang D (2012) Iterative Laplacian score for feature selection. CCPR 2012. Commun Comput Inf Sci 321:80–87
Metadaten
Titel
A comparative study of neural-network feature weighting
verfasst von
Tongfeng Sun
Shifei Ding
Pin Li
Wei Chen
Publikationsdatum
27.03.2019
Verlag
Springer Netherlands
Erschienen in
Artificial Intelligence Review / Ausgabe 1/2019
Print ISSN: 0269-2821
Elektronische ISSN: 1573-7462
DOI
https://doi.org/10.1007/s10462-019-09700-z

Weitere Artikel der Ausgabe 1/2019

Artificial Intelligence Review 1/2019 Zur Ausgabe

Premium Partner