Skip to main content
Top

2010 | OriginalPaper | Chapter

6. Classification with EEC, Divergence Measures, and Error Bounds

Authors : Deniz Erdogmus, Dongxin Xu, Kenneth Hild II

Published in: Information Theoretic Learning

Publisher: Springer New York

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

The previous chapters provided extensive coverage of the error entropy criterion (EEC) especially in regard to minimization of the error entropy (MEE) for linear and nonlinear filtering (or regression) applications. However, the spectrum of engineering applications of adaptive systems is much broader than filtering or regression. Even looking at the subclass of supervised applications we have yet to deal with classification, which is an important application area for learning technologies. All of the practical ingredients are here to extend EEC to classification inasmuch as Chapter 5 covered the integration of EEC with the backpropagation algorithm (MEE-BP). Hence we have all the tools needed to train classifiers with MEE. We show that indeed this is the case and that the classifiers trained with MEE have performances normally better than MSE-trained classifiers. However, there are still no mathematical foundations to ascertain under what conditions EEC is optimal for classification, and further work is necessary.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Aczél J., Daróczy Z., On measures of information and their characterizations, Mathematics in Science and Engineering, vol. 115, Academic Press, New York, 1975. Aczél J., Daróczy Z., On measures of information and their characterizations, Mathematics in Science and Engineering, vol. 115, Academic Press, New York, 1975.
2.
go back to reference Ahmad I., Lin P., A nonparametric estimation of the entropy for absolutely continuous distributions, IEEE Trans. on Inf. Theor., 22:372–375, 1976.CrossRefMATHMathSciNet Ahmad I., Lin P., A nonparametric estimation of the entropy for absolutely continuous distributions, IEEE Trans. on Inf. Theor., 22:372–375, 1976.CrossRefMATHMathSciNet
21.
go back to reference Battiti R., Using mutual information for selecting features in supervised neural net learning, IEEE Trans. Neural Netw., 5(4):537–550, July 1994.CrossRef Battiti R., Using mutual information for selecting features in supervised neural net learning, IEEE Trans. Neural Netw., 5(4):537–550, July 1994.CrossRef
37.
go back to reference Biem A., Katagiri S., Juang B., Pattern recognition using discriminative feature extraction, IEEE Trans. Signal Process., 45(2):500–504, Feb. 1997.CrossRef Biem A., Katagiri S., Juang B., Pattern recognition using discriminative feature extraction, IEEE Trans. Signal Process., 45(2):500–504, Feb. 1997.CrossRef
38.
go back to reference Bishop C., Neural Networks for Pattern Recognition, Clarendon Press, Oxford, 1995. Bishop C., Neural Networks for Pattern Recognition, Clarendon Press, Oxford, 1995.
68.
go back to reference Deco G., Obradovic D., An Information-Theoretic Approach to Neural Computing, Springer, New York, 1996.CrossRefMATH Deco G., Obradovic D., An Information-Theoretic Approach to Neural Computing, Springer, New York, 1996.CrossRefMATH
80.
go back to reference Duda R., Hart P., Stork D., Pattern Classification and Scene Analysis. John Wiley & Sons, New York, 2nd edition, 2001. Duda R., Hart P., Stork D., Pattern Classification and Scene Analysis. John Wiley & Sons, New York, 2nd edition, 2001.
86.
go back to reference Erdogmus D., Information theoretic learning: Renyi’s entropy and its applications to adaptive systems training, Ph.D. Dissertation, University of Florida, Gainesville, 2002. Erdogmus D., Information theoretic learning: Renyi’s entropy and its applications to adaptive systems training, Ph.D. Dissertation, University of Florida, Gainesville, 2002.
92.
go back to reference Erdogmus D., Principe J., Lower and upper bounds for misclassification probability based on Renyi’s information, J. VLSI Signal Process. Syst., 37(2/3):305–317, 2004.CrossRefMATH Erdogmus D., Principe J., Lower and upper bounds for misclassification probability based on Renyi’s information, J. VLSI Signal Process. Syst., 37(2/3):305–317, 2004.CrossRefMATH
97.
go back to reference Fano R., Transmission of Information: A Statistical Theory of Communications, MIT Press, New York, 1961. Fano R., Transmission of Information: A Statistical Theory of Communications, MIT Press, New York, 1961.
101.
go back to reference Fisher R., The use of multiple measurements in taxonomic problems, Ann. Eugenics 7; 170–188, Wiley, New York, 1950. Fisher R., The use of multiple measurements in taxonomic problems, Ann. Eugenics 7; 170–188, Wiley, New York, 1950.
108.
go back to reference Fukunaga K., An Introduction to Statistical Pattern Recognition, Academic Press, New York, 1972 Fukunaga K., An Introduction to Statistical Pattern Recognition, Academic Press, New York, 1972
145.
go back to reference Hertz J., Krogh A., and Palmer R., Introduction to the Theory of Neural Computation, Addison Wesley, Readings, MA, 1991. Hertz J., Krogh A., and Palmer R., Introduction to the Theory of Neural Computation, Addison Wesley, Readings, MA, 1991.
149.
go back to reference Hild II K., Blind source separation of convolutive mixtures using Renyi’s divergence, University of Florida, Gainesville, Fall 2003. Hild II K., Blind source separation of convolutive mixtures using Renyi’s divergence, University of Florida, Gainesville, Fall 2003.
151.
go back to reference Hild II K., Erdogmus D., Torkkola K., and Principe J., Feature extraction using information-theoretic learning, IEEE Trans. Pat. Anal. Mach. Intell. 28(9):1385–1392, 2006CrossRef Hild II K., Erdogmus D., Torkkola K., and Principe J., Feature extraction using information-theoretic learning, IEEE Trans. Pat. Anal. Mach. Intell. 28(9):1385–1392, 2006CrossRef
195.
go back to reference LeCun Y., Bottou L., Bengio Y., Haffner P., Gradient-based learning applied to document recognition, Proc. IEEE, 86(11):2278–2324, Nov. 1998.CrossRef LeCun Y., Bottou L., Bengio Y., Haffner P., Gradient-based learning applied to document recognition, Proc. IEEE, 86(11):2278–2324, Nov. 1998.CrossRef
223.
go back to reference Morejon R., An information theoretic approach to sonar automatic target recognition, Ph.D. dissertation, University of Florida, Spring 2003 Morejon R., An information theoretic approach to sonar automatic target recognition, Ph.D. dissertation, University of Florida, Spring 2003
253.
go back to reference Principe J., Euliano N., Lefebvre C., Neural Systems: Fundamentals through Simulations, CD-ROM textbook, John Wiley, New York, 2000. Principe J., Euliano N., Lefebvre C., Neural Systems: Fundamentals through Simulations, CD-ROM textbook, John Wiley, New York, 2000.
254.
go back to reference Principe J., Xu D., Zhao Q., Fisher J. Learning from examples with information theoretic criteria, VLSI Signal Process. Syst., 26:61–77, 2001.CrossRef Principe J., Xu D., Zhao Q., Fisher J. Learning from examples with information theoretic criteria, VLSI Signal Process. Syst., 26:61–77, 2001.CrossRef
268.
go back to reference Ripley B., Pattern Recognition and Neural Networks, Cambridge University Press, New York, 1996CrossRefMATH Ripley B., Pattern Recognition and Neural Networks, Cambridge University Press, New York, 1996CrossRefMATH
283.
go back to reference Santos J., Alexandre L., Sa J., The error entropy minimization algorithm for neural network classification, in A. Lofti (Ed.), Int Conf. Recent Advances in Soft Computing, pp. 92–97, 2004. Santos J., Alexandre L., Sa J., The error entropy minimization algorithm for neural network classification, in A. Lofti (Ed.), Int Conf. Recent Advances in Soft Computing, pp. 92–97, 2004.
298.
go back to reference Silva L., Felgueiras C., Alexandre L., Sa J., Error entropy in classification problems: a univariate data analysis, Neural comput., 18(9):2036–2061, 2006.CrossRefMATHMathSciNet Silva L., Felgueiras C., Alexandre L., Sa J., Error entropy in classification problems: a univariate data analysis, Neural comput., 18(9):2036–2061, 2006.CrossRefMATHMathSciNet
299.
go back to reference Silva L., Neural networks with error density risk functionals for data classification, Ph.D. Thesis, Faculdade de Engenharia, University of Porto, 2008. Silva L., Neural networks with error density risk functionals for data classification, Ph.D. Thesis, Faculdade de Engenharia, University of Porto, 2008.
318.
go back to reference Torkkola K., Visualizing class structure in data using mutual information, Proceedings of NNSP X, pp. 376–385, Sydney, Australia, 2000 Torkkola K., Visualizing class structure in data using mutual information, Proceedings of NNSP X, pp. 376–385, Sydney, Australia, 2000
323.
324.
go back to reference Velten V., Ross T., Mossing J., Worrell S., Bryant M., Standard SAR ATR evaluation experiments using the MSTAR public release data set, research report, Wright State University, 1998. Velten V., Ross T., Mossing J., Worrell S., Bryant M., Standard SAR ATR evaluation experiments using the MSTAR public release data set, research report, Wright State University, 1998.
334.
go back to reference Wittner B. Denker J., Strategies for teaching layered networks classification tasks, in Neural Inf. Proc. Syst. (Ed Anderson), 850–859, Ame. Inst. Phys. 1987. Wittner B. Denker J., Strategies for teaching layered networks classification tasks, in Neural Inf. Proc. Syst. (Ed Anderson), 850–859, Ame. Inst. Phys. 1987.
340.
go back to reference Xu D., Energy, Entropy and Information Potential for Neural Computation, PhD Dissertation, University of Florida, Gainesville, 1999 Xu D., Energy, Entropy and Information Potential for Neural Computation, PhD Dissertation, University of Florida, Gainesville, 1999
349.
go back to reference Zhao Q., Principe J., Brennan V., Xu D., Wang Z., Synthetic aperture radar automatic target recognition with three strategies of learning and representation, Opt. Eng., 39(5):1230–1244, 2000.CrossRef Zhao Q., Principe J., Brennan V., Xu D., Wang Z., Synthetic aperture radar automatic target recognition with three strategies of learning and representation, Opt. Eng., 39(5):1230–1244, 2000.CrossRef
Metadata
Title
Classification with EEC, Divergence Measures, and Error Bounds
Authors
Deniz Erdogmus
Dongxin Xu
Kenneth Hild II
Copyright Year
2010
Publisher
Springer New York
DOI
https://doi.org/10.1007/978-1-4419-1570-2_6

Premium Partner