Skip to main content
Top
Published in: Knowledge and Information Systems 3/2019

11-06-2018 | Regular Paper

A hybrid isotonic separation training algorithm with correlation-based isotonic feature selection for binary classification

Authors: B. Malar, R. Nadarajan, J. Gowri Thangam

Published in: Knowledge and Information Systems | Issue 3/2019

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Isotonic separation is a classification technique which constructs a model by transforming the training set into a linear programming problem (LPP). It is computationally expensive to solve large-scale LPPs using traditional methods when data set grows. This paper proposes a hybrid binary classification algorithm, meta-heuristic isotonic separation with particle swarm optimization and convergence criterion (MeHeIS–CPSO), in which a particle swarm optimization-based meta-heuristic is embedded in the training phase to find a solution for LPP. The proposed framework formulates the LPP as a directed acyclic graph (DAG) and arranges decision variables using topological sort. It obtains a new threshold value from training set and sets up a convergence criterion using this threshold. It also deploys a new correlation coefficient-based supervised feature selection technique to select isotonic features and improves predictive accuracy of the classifier. Experiments are conducted on publicly available data sets and synthetic data set. Theoretical, empirical, and statistical analyses show that MeHeIS–CPSO is superior to its predecessors in terms of training time and predictive ability on large data sets. It also outperforms state-of-the-art machine learning and isotonic classification techniques in terms of predictive performance on small- and large-scale data sets.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Jacob V, Krishnan R, Ryu YU (2007) Internet content filtering using isotonic separation on content category ratings. ACM Trans Internet Technol 7(1):1–19CrossRef Jacob V, Krishnan R, Ryu YU (2007) Internet content filtering using isotonic separation on content category ratings. ACM Trans Internet Technol 7(1):1–19CrossRef
2.
go back to reference Ryu YU, Yue WT (2005) Firm bankruptcy prediction; experimental comparison of isotonic separation and other classification approaches. IEEE Trans Syst Man Cybern Part A Syst Hum 35(5):727–737CrossRef Ryu YU, Yue WT (2005) Firm bankruptcy prediction; experimental comparison of isotonic separation and other classification approaches. IEEE Trans Syst Man Cybern Part A Syst Hum 35(5):727–737CrossRef
3.
go back to reference Ryu YU, Chandrasekaran R, Jacob VS (2007) Breast cancer detection using the isotonic separation technique. Eur J Oper Res 181:842–854CrossRefMATH Ryu YU, Chandrasekaran R, Jacob VS (2007) Breast cancer detection using the isotonic separation technique. Eur J Oper Res 181:842–854CrossRefMATH
4.
go back to reference Ryu YU, Chandrasekaran R, Jacob VS (2004) Prognosis using an isotonic prediction technique. Inf J Manag Sci 50(6):777–785 Ryu YU, Chandrasekaran R, Jacob VS (2004) Prognosis using an isotonic prediction technique. Inf J Manag Sci 50(6):777–785
6.
go back to reference Cano JR, Aljohani NR, Abbasi RA, Alowidbi JS, Garcia S (2017) Prototype selection to improve monotonic nearest neighbor. EAAI 60:128–135 Cano JR, Aljohani NR, Abbasi RA, Alowidbi JS, Garcia S (2017) Prototype selection to improve monotonic nearest neighbor. EAAI 60:128–135
7.
go back to reference Gonzalez S, Herrera F, Garcia S (2015) Monotonic random forest with an ensemble pruning mechanism based on the degree of monotonicity. New Gener Comput 33(4):367–388CrossRef Gonzalez S, Herrera F, Garcia S (2015) Monotonic random forest with an ensemble pruning mechanism based on the degree of monotonicity. New Gener Comput 33(4):367–388CrossRef
8.
go back to reference Ahuja RK, Magnanti TL, Orlin JB (1993) Network flows. Prentice-Hall, Englewood CliffsMATH Ahuja RK, Magnanti TL, Orlin JB (1993) Network flows. Prentice-Hall, Englewood CliffsMATH
9.
go back to reference Goldberg AV (1998) Recent developments in maximum flow algorithms. In: Proceedings of the 1998 Scandinavian workshop on algorithm theory, Springer, London, UK Goldberg AV (1998) Recent developments in maximum flow algorithms. In: Proceedings of the 1998 Scandinavian workshop on algorithm theory, Springer, London, UK
10.
go back to reference Dantzig GB, Thapa MN (1997) Linear programming 1: introduction. Springer, New YorkMATH Dantzig GB, Thapa MN (1997) Linear programming 1: introduction. Springer, New YorkMATH
11.
go back to reference Monteiro R, Adler I (1989) Interior Path following primal-dual algorithms. Part II: convex quadratic programming. Math Program 44:43–66CrossRefMATH Monteiro R, Adler I (1989) Interior Path following primal-dual algorithms. Part II: convex quadratic programming. Math Program 44:43–66CrossRefMATH
12.
go back to reference Deb K (2001) Multiobjective optimization using evolutionary algorithm. Wiley, New YorkMATH Deb K (2001) Multiobjective optimization using evolutionary algorithm. Wiley, New YorkMATH
13.
go back to reference Kennedy J, Eberhart R (1995) Particle swarm optimization. In: Proceedings of the IEEE international conference on neural networks, vol 4, pp 1942–1948, Piscataway Kennedy J, Eberhart R (1995) Particle swarm optimization. In: Proceedings of the IEEE international conference on neural networks, vol 4, pp 1942–1948, Piscataway
14.
go back to reference Kalousis A, Prados J, Hilario M (2007) A stability of feature selection algorithms: a study on high dimensional spaces. Knowl Inf Syst 12(1):95–116CrossRef Kalousis A, Prados J, Hilario M (2007) A stability of feature selection algorithms: a study on high dimensional spaces. Knowl Inf Syst 12(1):95–116CrossRef
15.
go back to reference Robertson T, Wright FT, Dykstra RL (1988) Order restricted statistical inference. Wiley, New YorkMATH Robertson T, Wright FT, Dykstra RL (1988) Order restricted statistical inference. Wiley, New YorkMATH
16.
go back to reference Dorigo M (1992) Optimization, learning and natural algorithms. In: PhD thesis, Dipartimento di Elettronica, Politecnico di Milano, Milan, Italy Dorigo M (1992) Optimization, learning and natural algorithms. In: PhD thesis, Dipartimento di Elettronica, Politecnico di Milano, Milan, Italy
17.
go back to reference Malar B, Nadarajan R (2013) Evolutionary Isotonic separation for classification: theory and experiments. Knowl Inf Syst 37(3):531–553CrossRef Malar B, Nadarajan R (2013) Evolutionary Isotonic separation for classification: theory and experiments. Knowl Inf Syst 37(3):531–553CrossRef
18.
go back to reference Goldberg DE (1989) Genetic algorithms for search, optimization and machine learning. Addision Wesley, BostonMATH Goldberg DE (1989) Genetic algorithms for search, optimization and machine learning. Addision Wesley, BostonMATH
19.
go back to reference Majid A, Lee CH, Mahmood MT et al (2012) Impulse noise filtering based on noise free pixels using genetic programming. Knowl Inf Syst 32(3):505–526CrossRef Majid A, Lee CH, Mahmood MT et al (2012) Impulse noise filtering based on noise free pixels using genetic programming. Knowl Inf Syst 32(3):505–526CrossRef
20.
go back to reference Duivesteijn W, Feelders A (2008) Nearest neighbor classification with monotonicity constraints. ECML/PKDD 1:301–316 Duivesteijn W, Feelders A (2008) Nearest neighbor classification with monotonicity constraints. ECML/PKDD 1:301–316
21.
go back to reference García J, Fardoun HM, Algazzawi DM, Cano JR, Garcia S (2017) MoNGEL: monotonic nested generalized exemplar learning. Pattern Anal Appl 20:441–452MathSciNetCrossRef García J, Fardoun HM, Algazzawi DM, Cano JR, Garcia S (2017) MoNGEL: monotonic nested generalized exemplar learning. Pattern Anal Appl 20:441–452MathSciNetCrossRef
22.
go back to reference Sousa RG, Cardoso JS (2011) Ensemble of decision trees with GLBAL constraints for ordinal classification, In: Proceedings of 11th international conference on intelligent systems design and applications, pp 1164–1169 Sousa RG, Cardoso JS (2011) Ensemble of decision trees with GLBAL constraints for ordinal classification, In: Proceedings of 11th international conference on intelligent systems design and applications, pp 1164–1169
23.
go back to reference Daniels H, Velikova M (2010) Monotone and partially monotone neural networks. IEEE Trans Neural Netw 21(6):906–917CrossRef Daniels H, Velikova M (2010) Monotone and partially monotone neural networks. IEEE Trans Neural Netw 21(6):906–917CrossRef
24.
go back to reference Eberhart RC, Shi Y (2001). Particle swarm optimization: developments, applications, and resources. In: Proceedings of the 2001 congress on evolutionary computation 2001. pp 81–86 Eberhart RC, Shi Y (2001). Particle swarm optimization: developments, applications, and resources. In: Proceedings of the 2001 congress on evolutionary computation 2001. pp 81–86
25.
go back to reference Shi Y, Eberhart R (1998) A modified particle swarm optimizer. In: Proceedings of the IEEE international conference on evolutionary computation, IEEE Press, Piscataway, pp 69–73 Shi Y, Eberhart R (1998) A modified particle swarm optimizer. In: Proceedings of the IEEE international conference on evolutionary computation, IEEE Press, Piscataway, pp 69–73
26.
go back to reference Eberhart RC, Simpson PK, Dobbins RW (1996) Computational intelligence PC tools. AP Professional, Boston Eberhart RC, Simpson PK, Dobbins RW (1996) Computational intelligence PC tools. AP Professional, Boston
27.
go back to reference Kennedy J, Eberhart R (1997) A discrete binary version of the Particle Swarm algorithm. In: Proceedings of the international conference on systemics, cybernatics and informatics, Orlando, FL, vol 5, pp 4104–4109 Kennedy J, Eberhart R (1997) A discrete binary version of the Particle Swarm algorithm. In: Proceedings of the international conference on systemics, cybernatics and informatics, Orlando, FL, vol 5, pp 4104–4109
28.
go back to reference Shen Q, Jiang JH, Jiao CX, Shen GL, Yu RQ (2004) Modified particle swarm optimization algorithm for variable selection in MLR and PLS modeling: QSAR studies of antagonism of angiotensin II antagonists. Eur J Pharm Sci 22(2–3):145–152CrossRef Shen Q, Jiang JH, Jiao CX, Shen GL, Yu RQ (2004) Modified particle swarm optimization algorithm for variable selection in MLR and PLS modeling: QSAR studies of antagonism of angiotensin II antagonists. Eur J Pharm Sci 22(2–3):145–152CrossRef
29.
go back to reference Wang L, Wang X, Fu J, Zhen L (2008) A novel probability binary particle swarm optimization algorithm and its application. J Softw 3(9):28–35CrossRef Wang L, Wang X, Fu J, Zhen L (2008) A novel probability binary particle swarm optimization algorithm and its application. J Softw 3(9):28–35CrossRef
30.
go back to reference Poli R (2008) Analysis of the publications on the applications of particle swarm optimization. J Artif Evolut Appl 2008:1–10 Poli R (2008) Analysis of the publications on the applications of particle swarm optimization. J Artif Evolut Appl 2008:1–10
31.
go back to reference Cormen TH, Leiserson CE, Rivest RL, Stein C (2001) Introduction to algorithms. MIT Press and McGrawHill, New York, pp 549–552MATH Cormen TH, Leiserson CE, Rivest RL, Stein C (2001) Introduction to algorithms. MIT Press and McGrawHill, New York, pp 549–552MATH
32.
go back to reference Merz CJ, Murphy PM (1998) UCI repository of machine learning databases. Department of information and computer sciences, University of California, Irvine Merz CJ, Murphy PM (1998) UCI repository of machine learning databases. Department of information and computer sciences, University of California, Irvine
33.
go back to reference Castillo C, Donato D, Becchetti L, Boldi P, Leonardi S, Santini M, Vigna S (2006) A reference collection for web spam. SIGIR Forum 40(2):11–24CrossRef Castillo C, Donato D, Becchetti L, Boldi P, Leonardi S, Santini M, Vigna S (2006) A reference collection for web spam. SIGIR Forum 40(2):11–24CrossRef
35.
go back to reference Gutierrez PA, Garcia S (2016) Current prospects on ordinal and monotonic classification, prog, artificial intelligence. Springer, Heidelberg Gutierrez PA, Garcia S (2016) Current prospects on ordinal and monotonic classification, prog, artificial intelligence. Springer, Heidelberg
36.
go back to reference Poli R, Kennedy J, Blackwell T (2007) Particle swarm optimization: an overview. Swarm Intell 1(1):33–57CrossRef Poli R, Kennedy J, Blackwell T (2007) Particle swarm optimization: an overview. Swarm Intell 1(1):33–57CrossRef
37.
go back to reference Klotz JH (2006) A computational approach to statistics, department of statistics, University of Wisconsin at Madison Klotz JH (2006) A computational approach to statistics, department of statistics, University of Wisconsin at Madison
38.
go back to reference Dawson RJM (1997) Turning the tables: a t-table for today. J Stat Edu 5(2):1–6 Dawson RJM (1997) Turning the tables: a t-table for today. J Stat Edu 5(2):1–6
40.
go back to reference Wu X, Vipin Kumar J, Quinlan R, Ghosh J, Yang Q et al (2008) Top 10 algorithms in data mining. Knowl Inf Syst 14(1):1–37CrossRef Wu X, Vipin Kumar J, Quinlan R, Ghosh J, Yang Q et al (2008) Top 10 algorithms in data mining. Knowl Inf Syst 14(1):1–37CrossRef
41.
go back to reference Quinlan JR (1993) C4.5: programs for machine learning. Morghan Kaufman, San Mateo Quinlan JR (1993) C4.5: programs for machine learning. Morghan Kaufman, San Mateo
42.
go back to reference Watters CB, Shepherd M (2003) Support vector machines for text categorization. In: Proceedings of the Hawaii 2003 international conference on system sciences, IEEE computer science society Watters CB, Shepherd M (2003) Support vector machines for text categorization. In: Proceedings of the Hawaii 2003 international conference on system sciences, IEEE computer science society
43.
go back to reference Eberhart RC, Kennedy J (1995) A new optimizer using particle swarm theory. In: Sixth international symposium on micro machine and human science, Nagoya, pp 39–43 Eberhart RC, Kennedy J (1995) A new optimizer using particle swarm theory. In: Sixth international symposium on micro machine and human science, Nagoya, pp 39–43
44.
go back to reference Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH (2009) The WEKA data mining software: an update. SIGKDD Explor 11(1):10–18CrossRef Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH (2009) The WEKA data mining software: an update. SIGKDD Explor 11(1):10–18CrossRef
45.
go back to reference Han J (2005) Datamining concepts and techniques. Morgan Kaufmann Publishers Inc., San Francisco Han J (2005) Datamining concepts and techniques. Morgan Kaufmann Publishers Inc., San Francisco
46.
go back to reference Joachims T (1998) Text categorization with support vector machines: learning with many relevant features. In: Proceedings of 1998 European conference on machine learning (ECML) Joachims T (1998) Text categorization with support vector machines: learning with many relevant features. In: Proceedings of 1998 European conference on machine learning (ECML)
48.
go back to reference Ntoulas A, Najork M, Manasse M, Fetterly D, (2006) Detecting spam web pages through content analysis. In: Proceedings of international conference on World Wide Web Ntoulas A, Najork M, Manasse M, Fetterly D, (2006) Detecting spam web pages through content analysis. In: Proceedings of international conference on World Wide Web
Metadata
Title
A hybrid isotonic separation training algorithm with correlation-based isotonic feature selection for binary classification
Authors
B. Malar
R. Nadarajan
J. Gowri Thangam
Publication date
11-06-2018
Publisher
Springer London
Published in
Knowledge and Information Systems / Issue 3/2019
Print ISSN: 0219-1377
Electronic ISSN: 0219-3116
DOI
https://doi.org/10.1007/s10115-018-1226-6

Other articles of this Issue 3/2019

Knowledge and Information Systems 3/2019 Go to the issue

Premium Partner