Skip to main content
Top
Published in: Neural Computing and Applications 7/2011

01-10-2011 | Original Article

Accelerating the kernel-method-based feature extraction procedure from the viewpoint of numerical approximation

Authors: Yong Xu, David Zhang

Published in: Neural Computing and Applications | Issue 7/2011

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

The kernel method suffers from the following problem: the computational efficiency of the feature extraction procedure is inversely proportional to the size of the training sample set. In this paper, from a novel viewpoint, we propose a very simple and mathematically tractable method to produce the computationally efficient kernel-method-based feature extraction procedure. We first address the issue that how to make the feature extraction result of the reformulated kernel method well approximate that of the naïve kernel method. We identify these training samples that statistically contribute much to the feature extraction results and exploit them to reformulate the kernel method to produce the computationally efficient kernel-method-based feature extraction procedure. Indeed, the proposed method has the following basic idea: when one training sample has little effect on the feature extraction result and statistically has the high correlation with regard to all the training samples, the feature extraction term associated with this training sample can be removed from the feature extraction procedure. The proposed method has the following advantages: First, it proposes, for the first time, to improve the kernel method through formal and reasonable evaluation on the feature extraction term. Second, the proposed method improves the kernel method at a low extra cost and thus has a much more computationally efficient training phase than most of the previous improvements to the kernel method. The experimental comparison shows that the proposed method performs well in classification problems. This paper also intuitively shows the geometrical relation between the identified training samples and other training samples.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Muller KR, Mika S, Ratsch G et al (2001) An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw 12(2):181–201CrossRef Muller KR, Mika S, Ratsch G et al (2001) An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw 12(2):181–201CrossRef
2.
go back to reference Vapnik VN (1995) The nature of statistical learning theory. Springer, BerlinMATH Vapnik VN (1995) The nature of statistical learning theory. Springer, BerlinMATH
3.
go back to reference Xu J, Zhang X, Li Y (2001) Kernel MSE algorithm: a unified framework for KFD, LS-SVM and KRR. In: Proceedings of the international joint conference on neural networks (IJCNN-2001), Washington, DC, pp 1486–1491 Xu J, Zhang X, Li Y (2001) Kernel MSE algorithm: a unified framework for KFD, LS-SVM and KRR. In: Proceedings of the international joint conference on neural networks (IJCNN-2001), Washington, DC, pp 1486–1491
4.
go back to reference Muller K-R, Mika S, Rätsch G, Tsuda K, Schölkopf B (2001) An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw 12(1):181–201CrossRef Muller K-R, Mika S, Rätsch G, Tsuda K, Schölkopf B (2001) An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw 12(1):181–201CrossRef
5.
go back to reference Girolami M (2002) Mercer kernel based clustering in feature space. IEEE Trans Neural Netw 13(4):669–688 Girolami M (2002) Mercer kernel based clustering in feature space. IEEE Trans Neural Netw 13(4):669–688
6.
go back to reference Shawe-Taylor J, Cristianini N (2004) Kernel methods for pattern analysis. Cambridge University Press, CambridgeCrossRef Shawe-Taylor J, Cristianini N (2004) Kernel methods for pattern analysis. Cambridge University Press, CambridgeCrossRef
7.
go back to reference Yang J, Frangi AF, Yang J-Y, Zhang D, Jin Z (2005) KPCA Plus LDA: A complete kernel fisher discriminant framework for feature extraction and recognition. IEEE Trans Pattern Anal Mach Intell 27(2):230–244CrossRef Yang J, Frangi AF, Yang J-Y, Zhang D, Jin Z (2005) KPCA Plus LDA: A complete kernel fisher discriminant framework for feature extraction and recognition. IEEE Trans Pattern Anal Mach Intell 27(2):230–244CrossRef
8.
go back to reference Schölkopf B, Smola A (2002) Learning with kernels. MIT Press, Cambridge Schölkopf B, Smola A (2002) Learning with kernels. MIT Press, Cambridge
9.
go back to reference Xu Y, Yang JY, Lu JF, Yu DJ (2004) An efficient renovation on kernel Fisher discriminant analysis and face recognition experiments. Pattern Recogn 37:2091–2094CrossRef Xu Y, Yang JY, Lu JF, Yu DJ (2004) An efficient renovation on kernel Fisher discriminant analysis and face recognition experiments. Pattern Recogn 37:2091–2094CrossRef
10.
go back to reference Scholkopf B, Mika S, Burges C, Knirsch P, Muller KR, Ratsch G, Smola A (1999) Input space versus feature space in kernel-based methods. IEEE Trans Neural Netw 10(5):1000–1017CrossRef Scholkopf B, Mika S, Burges C, Knirsch P, Muller KR, Ratsch G, Smola A (1999) Input space versus feature space in kernel-based methods. IEEE Trans Neural Netw 10(5):1000–1017CrossRef
11.
go back to reference Xu Y, Yang JY, Yang J (2004) A reformative Fisher discriminant analysis. Pattern Recogn 37:1299–1302MATHCrossRef Xu Y, Yang JY, Yang J (2004) A reformative Fisher discriminant analysis. Pattern Recogn 37:1299–1302MATHCrossRef
12.
go back to reference Schölkopf B, Smola A, Muller KR (1998) Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput 10(5):1299–1319CrossRef Schölkopf B, Smola A, Muller KR (1998) Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput 10(5):1299–1319CrossRef
13.
go back to reference Mika S, Rätsch G, Weston J, Schölkopf B, Müller K-R (1999) Fisher discriminant analysis with kernels. In: Neural networks for signal processing IX. IEEE, New York, pp 41–48 Mika S, Rätsch G, Weston J, Schölkopf B, Müller K-R (1999) Fisher discriminant analysis with kernels. In: Neural networks for signal processing IX. IEEE, New York, pp 41–48
14.
go back to reference Xu Y, Yang JY, Lu JF (2005) An efficient kernel-based nonlinear regression method for two-class classification. Proceedings of the fourth international conference on machine learning and cybernetics, Guangzhou, China, 18–21 August 2005, pp 4442–4445 Xu Y, Yang JY, Lu JF (2005) An efficient kernel-based nonlinear regression method for two-class classification. Proceedings of the fourth international conference on machine learning and cybernetics, Guangzhou, China, 18–21 August 2005, pp 4442–4445
15.
go back to reference Steve B, Kian L (2002) Nonlinear Fisher discriminant analysis using a minimum squared error cost function and the orthogonal least squares algorithm. Neural Netw 15(1):263–270 Steve B, Kian L (2002) Nonlinear Fisher discriminant analysis using a minimum squared error cost function and the orthogonal least squares algorithm. Neural Netw 15(1):263–270
16.
go back to reference Xu Y, Zhang D, Jin Z, Li M, Yang J-Y (2006) A fast kernel-based nonlinear discriminant analysis for multi-class problems. Pattern Recogn 39(6):1026–1033MATHCrossRef Xu Y, Zhang D, Jin Z, Li M, Yang J-Y (2006) A fast kernel-based nonlinear discriminant analysis for multi-class problems. Pattern Recogn 39(6):1026–1033MATHCrossRef
17.
go back to reference Tipping ME (2000) Sparse kernel principal component analysis. In: Leen TK, Dietterich TG, Tresp V (eds) NIPS 2000: Neural information processing systems. MIT Press, Cambridge, pp 633–639 Tipping ME (2000) Sparse kernel principal component analysis. In: Leen TK, Dietterich TG, Tresp V (eds) NIPS 2000: Neural information processing systems. MIT Press, Cambridge, pp 633–639
18.
go back to reference Cawley GC, Talbot NLC (2003) Efficient leave-one-out cross-validation of kernel fisher discriminant classifiers. Pattern Recogn 36(11):2585–2592MATHCrossRef Cawley GC, Talbot NLC (2003) Efficient leave-one-out cross-validation of kernel fisher discriminant classifiers. Pattern Recogn 36(11):2585–2592MATHCrossRef
19.
go back to reference Kim S-W, Oommen BJ (2008) On using prototype reduction schemes to optimize kernel-based fisher discriminant analysis. IEEE Trans Syst Man Cybern Part B Cybern 38(2):564–570CrossRef Kim S-W, Oommen BJ (2008) On using prototype reduction schemes to optimize kernel-based fisher discriminant analysis. IEEE Trans Syst Man Cybern Part B Cybern 38(2):564–570CrossRef
20.
go back to reference Burges CJC, Scholkopf B (1997) Improving the accuracy and speed of support vector learning machines. In: Mozer M, Jordan M, Petsche T (eds) Advances in neural information processing systems, vol 9. MIT Press, Cambridge, pp 375–381 Burges CJC, Scholkopf B (1997) Improving the accuracy and speed of support vector learning machines. In: Mozer M, Jordan M, Petsche T (eds) Advances in neural information processing systems, vol 9. MIT Press, Cambridge, pp 375–381
21.
go back to reference Mika S, Ratsch G, Müller K-R (2001) A mathematical programming approach to the kernel Fisher algorithm. Adv Neural Inf Process Syst 13:591–597 Mika S, Ratsch G, Müller K-R (2001) A mathematical programming approach to the kernel Fisher algorithm. Adv Neural Inf Process Syst 13:591–597
22.
go back to reference Xu Y, Zhang D, Song F et al (2007) A method for speeding up feature extraction based on KPCA. Neurocomputing 70:1056–1061 Xu Y, Zhang D, Song F et al (2007) A method for speeding up feature extraction based on KPCA. Neurocomputing 70:1056–1061
Metadata
Title
Accelerating the kernel-method-based feature extraction procedure from the viewpoint of numerical approximation
Authors
Yong Xu
David Zhang
Publication date
01-10-2011
Publisher
Springer-Verlag
Published in
Neural Computing and Applications / Issue 7/2011
Print ISSN: 0941-0643
Electronic ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-011-0534-5

Other articles of this Issue 7/2011

Neural Computing and Applications 7/2011 Go to the issue

Premium Partner