Skip to main content

2015 | OriginalPaper | Buchkapitel

A Holistic Classification Optimization Framework with Feature Selection, Preprocessing, Manifold Learning and Classifiers

verfasst von : Fabian Bürger, Josef Pauli

Erschienen in: Pattern Recognition: Applications and Methods

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

All real-world classification problems require a carefully designed system to achieve the desired generalization performance. Developers need to select a useful feature subset and a classifier with suitable hyperparameters. Furthermore, a feature preprocessing method (e.g. scaling or pre-whitening) and a dimension reduction method (e.g. Principal Component Analysis (PCA), Autoencoders or other manifold learning algorithms) may improve the performance. The interplay of all these components is complex and a manual selection is time-consuming. This paper presents an automatic optimization framework that incorporates feature selection, several feature preprocessing methods, multiple feature transforms learned by manifold learning and multiple classifiers including all hyperparameters. The highly combinatorial optimization problem is solved with an evolutionary algorithm. Additionally, a multi-classifier based on the optimization trajectory is presented which improves the generalization. The evaluation on several datasets shows the effectiveness of the proposed framework.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Fußnoten
1
Hyperparameters control the learning algorithm itself – e.g. the number of hidden layers in a neural network.
 
2
The average memory consumption of the proposed system is below 8 GB.
 
3
The statlogheart dataset is a medical application in which diagnoses are correlated with absence of presence of serious heart diseases.
 
4
The glass dataset is a forensic application in which glass is classified by oxide contents with the goal to identify the origins of the glass.
 
5
The popular iris dataset correlates variants of the iris plant with leaf dimensions.
 
Literatur
1.
Zurück zum Zitat Wolpert, D.H.: The lack of a priori distinctions between learning algorithms. Neural Comput. 8, 1341–1390 (1996)CrossRef Wolpert, D.H.: The lack of a priori distinctions between learning algorithms. Neural Comput. 8, 1341–1390 (1996)CrossRef
2.
Zurück zum Zitat Jain, A.K., Duin, R.P.W., Mao, J.: Statistical pattern recognition: a review. IEEE Trans. Pattern Anal. Mach. Intell. 22, 4–37 (2000)CrossRef Jain, A.K., Duin, R.P.W., Mao, J.: Statistical pattern recognition: a review. IEEE Trans. Pattern Anal. Mach. Intell. 22, 4–37 (2000)CrossRef
3.
Zurück zum Zitat Juszczak, P., Tax, D., Duin, R.: Feature scaling in support vector data description. In: Proceedings ASCI, pp. 95–102. Citeseer (2002) Juszczak, P., Tax, D., Duin, R.: Feature scaling in support vector data description. In: Proceedings ASCI, pp. 95–102. Citeseer (2002)
4.
Zurück zum Zitat Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intelli. 35, 1798–1828 (2013)CrossRef Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intelli. 35, 1798–1828 (2013)CrossRef
5.
Zurück zum Zitat Bürger, F., Pauli, J.: Representation optimization with feature selection and manifold learning in a holistic classification framework. In: De Marsico, M., Fred, A., eds.: Proceedings of the International Conference on Pattern Recognition Applications and Methods ICPRAM 2015, vol. 1, pp. 35–44. INSTICC, SCITEPRESS, Lisbon (2015) Bürger, F., Pauli, J.: Representation optimization with feature selection and manifold learning in a holistic classification framework. In: De Marsico, M., Fred, A., eds.: Proceedings of the International Conference on Pattern Recognition Applications and Methods ICPRAM 2015, vol. 1, pp. 35–44. INSTICC, SCITEPRESS, Lisbon (2015)
6.
Zurück zum Zitat Thornton, C., Hutter, F., Hoos, H.H., Leyton-Brown, K.: Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms. In: Proceedings of KDD-2013, pp. 847–855 (2013) Thornton, C., Hutter, F., Hoos, H.H., Leyton-Brown, K.: Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms. In: Proceedings of KDD-2013, pp. 847–855 (2013)
7.
Zurück zum Zitat Bengio, Y.: Gradient-based optimization of hyperparameters. Neural Comput. 12, 1889–1900 (2000)CrossRef Bengio, Y.: Gradient-based optimization of hyperparameters. Neural Comput. 12, 1889–1900 (2000)CrossRef
8.
Zurück zum Zitat Bergstra, J., Bardenet, R., Bengio, Y., Kégl, B., et al.: Algorithms for hyper-parameter optimization. In: 25th Annual Conference on Neural Information Processing Systems (NIPS 2011) (2011) Bergstra, J., Bardenet, R., Bengio, Y., Kégl, B., et al.: Algorithms for hyper-parameter optimization. In: 25th Annual Conference on Neural Information Processing Systems (NIPS 2011) (2011)
9.
Zurück zum Zitat Bergstra, J., Bengio, Y.: Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13, 281–305 (2012)MathSciNetMATH Bergstra, J., Bengio, Y.: Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13, 281–305 (2012)MathSciNetMATH
10.
Zurück zum Zitat Huang, C.L., Wang, C.J.: A GA-based feature selection and parameters optimization for support vector machines. Expert Syst. Appl. 31, 231–240 (2006)CrossRef Huang, C.L., Wang, C.J.: A GA-based feature selection and parameters optimization for support vector machines. Expert Syst. Appl. 31, 231–240 (2006)CrossRef
11.
Zurück zum Zitat Huang, H.L., Chang, F.L.: ESVM: evolutionary support vector machine for automatic feature selection and classification of microarray data. Biosyst. 90, 516–528 (2007)CrossRef Huang, H.L., Chang, F.L.: ESVM: evolutionary support vector machine for automatic feature selection and classification of microarray data. Biosyst. 90, 516–528 (2007)CrossRef
12.
Zurück zum Zitat Åberg, M., Wessberg, J.: Evolutionary optimization of classifiers and features for single trial EEG discrimination. Biomed. Eng. Online 6, 32 (2007)CrossRef Åberg, M., Wessberg, J.: Evolutionary optimization of classifiers and features for single trial EEG discrimination. Biomed. Eng. Online 6, 32 (2007)CrossRef
13.
Zurück zum Zitat Hutter, F., Hoos, H.H., Leyton-Brown, K., Stützle, T.: ParamILS: an automatic algorithm configuration framework. J. Artif. Intell. Res. 36, 267–306 (2009)MATH Hutter, F., Hoos, H.H., Leyton-Brown, K., Stützle, T.: ParamILS: an automatic algorithm configuration framework. J. Artif. Intell. Res. 36, 267–306 (2009)MATH
14.
Zurück zum Zitat Hutter, F., Hoos, H.H., Leyton-Brown, K.: Sequential model-based optimization for general algorithm configuration. In: Coello, C.A.C. (ed.) LION 2011. LNCS, vol. 6683, pp. 507–523. Springer, Heidelberg (2011) CrossRef Hutter, F., Hoos, H.H., Leyton-Brown, K.: Sequential model-based optimization for general algorithm configuration. In: Coello, C.A.C. (ed.) LION 2011. LNCS, vol. 6683, pp. 507–523. Springer, Heidelberg (2011) CrossRef
15.
Zurück zum Zitat Ansótegui, C., Sellmann, M., Tierney, K.: A gender-based genetic algorithm for the automatic configuration of algorithms. In: Gent, I.P. (ed.) CP 2009. LNCS, vol. 5732, pp. 142–157. Springer, Heidelberg (2009) CrossRef Ansótegui, C., Sellmann, M., Tierney, K.: A gender-based genetic algorithm for the automatic configuration of algorithms. In: Gent, I.P. (ed.) CP 2009. LNCS, vol. 5732, pp. 142–157. Springer, Heidelberg (2009) CrossRef
16.
Zurück zum Zitat Buschmann, F., Meunier, R., Rohnert, H., Sommerlad, P., Stal, M., Stal, M.: Pattern-Oriented Software Architecture: A System of Patterns, vol. 1, Wiley, New York (1996) Buschmann, F., Meunier, R., Rohnert, H., Sommerlad, P., Stal, M., Stal, M.: Pattern-Oriented Software Architecture: A System of Patterns, vol. 1, Wiley, New York (1996)
17.
Zurück zum Zitat Bishop, C.M., Nasrabadi, N.M.: Pattern recognition and machine learning. vol. 1, Springer, New York (2006) Bishop, C.M., Nasrabadi, N.M.: Pattern recognition and machine learning. vol. 1, Springer, New York (2006)
18.
Zurück zum Zitat Van der Maaten, L., Postma, E., Van Den Herik, H.: Dimensionality reduction: a comparative review. J. Mach. Learn. Res. 10, 1–41 (2009) Van der Maaten, L., Postma, E., Van Den Herik, H.: Dimensionality reduction: a comparative review. J. Mach. Learn. Res. 10, 1–41 (2009)
20.
Zurück zum Zitat Ma, Y., Fu, Y.: Manifold Learning Theory and Applications. CRC Press, Boca Raton (2011) Ma, Y., Fu, Y.: Manifold Learning Theory and Applications. CRC Press, Boca Raton (2011)
21.
Zurück zum Zitat Tenenbaum, J.B., De Silva, V., Langford, J.C.: A global geometric framework for nonlinear dimensionality reduction. Science 290, 2319–2323 (2000)CrossRef Tenenbaum, J.B., De Silva, V., Langford, J.C.: A global geometric framework for nonlinear dimensionality reduction. Science 290, 2319–2323 (2000)CrossRef
22.
Zurück zum Zitat Bengio, Y., Paiement, J.f., Vincent, P., Delalleau, O., Roux, N.L., Ouimet, M.: Out-of-sample extensions for LLE, Isomap, MDS, eigenmaps, and spectral clustering. In: Advances in Neural Information Processing Systems (2003) Bengio, Y., Paiement, J.f., Vincent, P., Delalleau, O., Roux, N.L., Ouimet, M.: Out-of-sample extensions for LLE, Isomap, MDS, eigenmaps, and spectral clustering. In: Advances in Neural Information Processing Systems (2003)
23.
Zurück zum Zitat Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: theory and applications. Neurocomputing 70, 489–501 (2006)CrossRef Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: theory and applications. Neurocomputing 70, 489–501 (2006)CrossRef
24.
Zurück zum Zitat Darwin, C.: On the Origins of Species by Means of Natural Selection. Murray, London (1859) Darwin, C.: On the Origins of Species by Means of Natural Selection. Murray, London (1859)
25.
Zurück zum Zitat Bäck, T.: Evolutionary algorithms in theory and practice. Oxford UniversityPress (1996) Bäck, T.: Evolutionary algorithms in theory and practice. Oxford UniversityPress (1996)
27.
Zurück zum Zitat Müller, M.: Ein Entwurfsmuster für die multikriterielle Parameteradaption mit Evolutionsstrategien in der Bildverarbeitung. VDI-Verlag (2012) Müller, M.: Ein Entwurfsmuster für die multikriterielle Parameteradaption mit Evolutionsstrategien in der Bildverarbeitung. VDI-Verlag (2012)
28.
Zurück zum Zitat Ranawana, R., Palade, V.: Multi-classifier systems: review and a roadmap for developers. Int. J. Hybrid Intell. Syst. 3, 35–61 (2006)MATH Ranawana, R., Palade, V.: Multi-classifier systems: review and a roadmap for developers. Int. J. Hybrid Intell. Syst. 3, 35–61 (2006)MATH
Metadaten
Titel
A Holistic Classification Optimization Framework with Feature Selection, Preprocessing, Manifold Learning and Classifiers
verfasst von
Fabian Bürger
Josef Pauli
Copyright-Jahr
2015
DOI
https://doi.org/10.1007/978-3-319-27677-9_4

Premium Partner