Skip to main content
Top

2015 | OriginalPaper | Chapter

Anomaly Detection Through Comparison of Heterogeneous Machine Learning Classifiers vs KPCA

Authors : Goverdhan Reddy Jidiga, P. Sammulal

Published in: Security in Computing and Communications

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

The anomaly detection is applicable to wide range of critical infrastructure elements due to frequent change in anomaly occurrences and make sure to avoid all threats identified in regular. In this perception, we have to identify the abnormal patterns in applications and to model them by using a new adorned machine learning classifiers. In this paper we are investigating the performance by comparison of heterogeneous machine learning classifiers: ICA (Independent Component Analysis), LDA (Linear Discriminant Analysis), PCA (Principal Component Analysis), Kernel PCA and other learning classifiers. The Kernel PCA (KPCA) is a non-linear extension to PCA used to classify the data and detect anomalies by orthogonal transformation of input space into (usually high dimensional) feature space. The KPCA use kernel trick by extract the principal components from set of corresponding eigenvectors and use kernel width as performance parameter to determine rate of classification. The KPCA is implemented on taking two UCI machine learning repository sets and one real bank dataset. The KPCA implemented with classic Gaussian kernel internally. Finally KPCA performance compared with projection methods (ICA, LDA, PLSDA and PCA), other kernel (SVM-K) and non-kernel techniques (ID3, C4.5, Rule C4.5, k-NN and NB) applied on same datasets using training and test set combinations.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Denning, D.E.: An intrusion detection model. IEEE Trans. Softw. Eng. 13(2), 222–232 (1987)CrossRef Denning, D.E.: An intrusion detection model. IEEE Trans. Softw. Eng. 13(2), 222–232 (1987)CrossRef
2.
go back to reference Axelsson, S.: Intrusion Detection Systems: A Survey and Taxonomy. Chalmers University, Technical report, vol. 99(15), March 2000 Axelsson, S.: Intrusion Detection Systems: A Survey and Taxonomy. Chalmers University, Technical report, vol. 99(15), March 2000
3.
go back to reference Lee, W., Stolfo, S.J.: Data mining approaches for intrusion detection. In: 7th USENIX Security Symposium, pp. 79–94, Berkeley (1998) Lee, W., Stolfo, S.J.: Data mining approaches for intrusion detection. In: 7th USENIX Security Symposium, pp. 79–94, Berkeley (1998)
4.
go back to reference Lane, T., Brodley, C.E.: An application of machine learning to anomaly detection. In: Proceedings of the 20th National Information Systems Security Conference, pp. 366–377, October 1997 Lane, T., Brodley, C.E.: An application of machine learning to anomaly detection. In: Proceedings of the 20th National Information Systems Security Conference, pp. 366–377, October 1997
6.
go back to reference Russell, S., Norvig, P.: Artificial Intelligence: A Modern Approach, 3rd edn. Prentice Hall, Saddle River (2009) Russell, S., Norvig, P.: Artificial Intelligence: A Modern Approach, 3rd edn. Prentice Hall, Saddle River (2009)
9.
go back to reference Hotelling, H.: Analysis of a complex statistical variable into principal components. J. Educ. Psychol. 24, 417–441 (1933)CrossRef Hotelling, H.: Analysis of a complex statistical variable into principal components. J. Educ. Psychol. 24, 417–441 (1933)CrossRef
10.
go back to reference Bai, Z.-J., Chan, R.H., Luk, F.T.: Principal component analysis for distributed data sets with updating. In: Cao, J., Nejdl, W., Xu, M. (eds.) APPT 2005. LNCS, vol. 3756, pp. 471–483. Springer, Heidelberg (2005)CrossRef Bai, Z.-J., Chan, R.H., Luk, F.T.: Principal component analysis for distributed data sets with updating. In: Cao, J., Nejdl, W., Xu, M. (eds.) APPT 2005. LNCS, vol. 3756, pp. 471–483. Springer, Heidelberg (2005)CrossRef
11.
go back to reference Lakhina, A., Crovella, M., Diot, C.: Diagnosing network-wide traffic anomalies. In: Proceedings of ACM Conference, Special Interest Group on Data Communication (2004) Lakhina, A., Crovella, M., Diot, C.: Diagnosing network-wide traffic anomalies. In: Proceedings of ACM Conference, Special Interest Group on Data Communication (2004)
12.
go back to reference Eskin, E., Arnold, A., Prerau, M.: A Geometric framework for unsupervised anomaly detection: detecting intrusions in unlabeled data. In: Applications of Data Mining in Computer Security (2003) Eskin, E., Arnold, A., Prerau, M.: A Geometric framework for unsupervised anomaly detection: detecting intrusions in unlabeled data. In: Applications of Data Mining in Computer Security (2003)
13.
go back to reference Lin, C.H., Chun Liu, J.. Ho, C.H.: Anomaly Detection Using LibSVM Training Tools. IEEE, Tunghai University, Taiwan. doi:10.1109/ISA.2008.12, ISBN-978-0-7695-3126-7/08 2008 Lin, C.H., Chun Liu, J.. Ho, C.H.: Anomaly Detection Using LibSVM Training Tools. IEEE, Tunghai University, Taiwan. doi:10.​1109/​ISA.​2008.​12, ISBN-978-0-7695-3126-7/08 2008
14.
go back to reference Kleinbaum, D.G.: Applied Regression Analysis and Multivariable Methods, 3rd edn. Brooks/Cole Publishing Company, Pacific Grove (1998) Kleinbaum, D.G.: Applied Regression Analysis and Multivariable Methods, 3rd edn. Brooks/Cole Publishing Company, Pacific Grove (1998)
15.
go back to reference Hoffmann, H.: Kernel PCA for novelty detection. Pattern Recogn. 40(3), 863–874 (2006)CrossRef Hoffmann, H.: Kernel PCA for novelty detection. Pattern Recogn. 40(3), 863–874 (2006)CrossRef
16.
go back to reference Genton, M.G.: Classes of kernels for machine learning: a statistics perspective. J. Mach. Learn. Res. 2, 299–312 (2001)MathSciNet Genton, M.G.: Classes of kernels for machine learning: a statistics perspective. J. Mach. Learn. Res. 2, 299–312 (2001)MathSciNet
17.
go back to reference Scholkopf, B., Smola, A.J., Muller, K.R.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 10, 1299–1319 (1998)CrossRef Scholkopf, B., Smola, A.J., Muller, K.R.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 10, 1299–1319 (1998)CrossRef
18.
go back to reference Jidiga, G.R., Sammulal, P.: Foundations of IDS: focus on role of anomaly detection using machine learning. In: ICACM-2013 Elsevier 2nd International Conference. August 2013. ISBN No: 9789351071495 Jidiga, G.R., Sammulal, P.: Foundations of IDS: focus on role of anomaly detection using machine learning. In: ICACM-2013 Elsevier 2nd International Conference. August 2013. ISBN No: 9789351071495
19.
go back to reference Quinlan, J.R.: Induction of decision trees. Mach. Learn. 1(1), 81–106 (1986) Quinlan, J.R.: Induction of decision trees. Mach. Learn. 1(1), 81–106 (1986)
20.
go back to reference Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, LosAltos (1993) Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, LosAltos (1993)
21.
go back to reference Quinlan, J.R., Rivest, R.L.: Inferring decision trees using the minimum description length principle. Inf. Comput. 80, 227–248 (1989)MathSciNetCrossRefMATH Quinlan, J.R., Rivest, R.L.: Inferring decision trees using the minimum description length principle. Inf. Comput. 80, 227–248 (1989)MathSciNetCrossRefMATH
22.
go back to reference Polat, K., Güne, S.: A novel hybrid intelligent method based on C4.5 decision tree classifier and one against all approach for multi-class classification problems. Expert Syst. Appl. 36, 1587–1592 (2009)CrossRef Polat, K., Güne, S.: A novel hybrid intelligent method based on C4.5 decision tree classifier and one against all approach for multi-class classification problems. Expert Syst. Appl. 36, 1587–1592 (2009)CrossRef
23.
go back to reference Yu, M., Ai, T.H.: Study of RS data classification based on rough sets and C4.5 algorithms. In: Proceedings of the SPIE Conference Series (2009) Yu, M., Ai, T.H.: Study of RS data classification based on rough sets and C4.5 algorithms. In: Proceedings of the SPIE Conference Series (2009)
24.
go back to reference Prema, R., Kannan, A.: An active rule approach for network intrusion detection with enhanced C4.5 Algorithm. In: IJCNS, pp. 285–385 (2008) Prema, R., Kannan, A.: An active rule approach for network intrusion detection with enhanced C4.5 Algorithm. In: IJCNS, pp. 285–385 (2008)
27.
go back to reference Daniel, L., Davis, J.: Improving Markov network structure learning using decision trees. J. Mach. Learn. Res. 15, 501–532 (2014) Daniel, L., Davis, J.: Improving Markov network structure learning using decision trees. J. Mach. Learn. Res. 15, 501–532 (2014)
29.
go back to reference Fukunaga, K.: Statistical Pattern Recognition. Academic Press, New York (1989) Fukunaga, K.: Statistical Pattern Recognition. Academic Press, New York (1989)
30.
go back to reference Haeb’h, R., Ney, H.: Linear discriminant analysis for improved large vocabulary speech recognition. In: Proceedings of ICASSP 1992, pp. 13–16, San Francisco, March 1992 Haeb’h, R., Ney, H.: Linear discriminant analysis for improved large vocabulary speech recognition. In: Proceedings of ICASSP 1992, pp. 13–16, San Francisco, March 1992
31.
go back to reference Hyvarinen, A.: Fast and robust fixed-point algorithms for independent component analysis. IEEE Trans. Neural Netw. 10(3), 626–634 (1999)CrossRef Hyvarinen, A.: Fast and robust fixed-point algorithms for independent component analysis. IEEE Trans. Neural Netw. 10(3), 626–634 (1999)CrossRef
32.
33.
go back to reference Kocsor, A., Tóth, L., Paczolay, D.: A nonlinearized discriminant analysis and its application to speech impediment therapy. In: Matoušek, V., Mautner, P., Mouček, R., Tauser, K. (eds.) TSD 2001. LNCS (LNAI), vol. 2166, pp. 249–257. Springer, Heidelberg (2001)CrossRef Kocsor, A., Tóth, L., Paczolay, D.: A nonlinearized discriminant analysis and its application to speech impediment therapy. In: Matoušek, V., Mautner, P., Mouček, R., Tauser, K. (eds.) TSD 2001. LNCS (LNAI), vol. 2166, pp. 249–257. Springer, Heidelberg (2001)CrossRef
34.
go back to reference Kocsor, A., Csirik, J.A.: Fast independent component analysis in kernel feature spaces. In: Pacholski, L., Ružička, P. (eds.) SOFSEM 2001. LNCS, vol. 2234, pp. 271–281. Springer, Heidelberg (2001)CrossRef Kocsor, A., Csirik, J.A.: Fast independent component analysis in kernel feature spaces. In: Pacholski, L., Ružička, P. (eds.) SOFSEM 2001. LNCS, vol. 2234, pp. 271–281. Springer, Heidelberg (2001)CrossRef
35.
go back to reference Scholkopf, B., Smola, A.J., Muller, K.R.: Kernel Principal Component Analysis in Advances in Kernel Methods - Support Vector Learning, pp. 327–352. MIT Press, Cambridge (1999) Scholkopf, B., Smola, A.J., Muller, K.R.: Kernel Principal Component Analysis in Advances in Kernel Methods - Support Vector Learning, pp. 327–352. MIT Press, Cambridge (1999)
36.
go back to reference Siohan, O.: On the robustness of linear discriminant analysis as a preprocessing step for noisy speech recognition. In: Proceedings of ICASSP 1995, pp. 125–128, Detroit, May 1995 Siohan, O.: On the robustness of linear discriminant analysis as a preprocessing step for noisy speech recognition. In: Proceedings of ICASSP 1995, pp. 125–128, Detroit, May 1995
37.
go back to reference Beveridge, J.R., She, K., Draper, B., Givens, G.H.: A nonparametric statistical comparison of principal component and linear discriminant subspaces for face recognition. In: Proceedings of the IEEE Conference on CVPR, pp. 535–542, USA, December 2001 Beveridge, J.R., She, K., Draper, B., Givens, G.H.: A nonparametric statistical comparison of principal component and linear discriminant subspaces for face recognition. In: Proceedings of the IEEE Conference on CVPR, pp. 535–542, USA, December 2001
38.
go back to reference Martinez, A., Kak, A.: PCA versus LDA. IEEE Trans. Pattern Anal. Mach. Intell. 23(2), 228–233 (2001)CrossRef Martinez, A., Kak, A.: PCA versus LDA. IEEE Trans. Pattern Anal. Mach. Intell. 23(2), 228–233 (2001)CrossRef
39.
go back to reference Hang, Z., Greenberg, A., Roughan, M.: Network anomography. In: Proceedings of Internet Measurement Conference (IMC) (2005) Hang, Z., Greenberg, A., Roughan, M.: Network anomography. In: Proceedings of Internet Measurement Conference (IMC) (2005)
Metadata
Title
Anomaly Detection Through Comparison of Heterogeneous Machine Learning Classifiers vs KPCA
Authors
Goverdhan Reddy Jidiga
P. Sammulal
Copyright Year
2015
DOI
https://doi.org/10.1007/978-3-319-22915-7_44

Premium Partner