Skip to main content

2013 | OriginalPaper | Buchkapitel

5. Multi-Layer Perceptrons

verfasst von : Rudolf Kruse, Christian Borgelt, Frank Klawonn, Christian Moewes, Matthias Steinbrecher, Pascal Held

Erschienen in: Computational Intelligence

Verlag: Springer London

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Having described the structure, the operation and the training of (artificial) neural networks in a general fashion in the preceding chapter, we turn in this and the subsequent chapters to specific forms of (artificial) neural networks. We start with the best-known and most widely used form, the so-called multi-layer perceptron (MLP), which is closely related to the networks of threshold logic units we studied in a previous chapter. They exhibit a strictly layered structure and may employ other activation functions than a step at a crisp threshold.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
Conservative logic is a mathematical model for computations and computational powers of computers, in which the fundamental physical principles that govern computing machines are explicitly taken into account. Among these principles are, for instance, that the speed with which information can travel as well as the amount of information that can be stored in the state of a finite system are both finite (Fredkin and Toffoli 1982).
 
2
In the following we assume implicitly that the output function of all neurons is the identity. Only the activation functions are exchanged.
 
3
Note that this approach is not easily transferred to functions with multiple arguments. For this to be possible, the influences of the two or more inputs have to be independent in a certain sense.
 
4
Note, however, that with this approach the sum of squared errors is minimized in the transformed space (coordinates x′=lnx and y′=lny), but this does not imply that it is also minimized in the original space (coordinates x and y). Nevertheless this approach usually yields very good results or at least an initial solution that may then be improved by other means.
 
5
Note again that with this procedure the sum of squared errors is minimized in the transformed space (coordinates x and \(z = \ln (\frac{Y-y}{y} )\)), but this does not imply that it is also minimized in the original space (coordinates x and y), cf. the preceding footnote.
 
6
Unless the output function is not differentiable. However, we usually assume (implicitly) that the output function is the identity and thus does not introduce any problems.
 
7
In order to avoid this factor right from the start, the error of an output neuron is sometimes defined as \(e_{u}^{(l)} = \frac{1}{2} (o_{u}^{(l)} - \operatorname{out}_{u}^{(l)} )^{2}\). In this way the factor 2 simply cancels in the derivation.
 
8
Note that the bias value θ u is already contained in the extended weight vector.
 
Literatur
Zurück zum Zitat S.E. Fahlman. An Empirical Study of Learning Speed in Backpropagation Networks. In: Touretzky et al. (1988) S.E. Fahlman. An Empirical Study of Learning Speed in Backpropagation Networks. In: Touretzky et al. (1988)
Zurück zum Zitat E. Fredkin and T. Toffoli. Conservative Logic. International Journal of Theoretical Physics 21(3/4):219–253. Plenum Press, New York, NY, USA, 1982 MathSciNetMATHCrossRef E. Fredkin and T. Toffoli. Conservative Logic. International Journal of Theoretical Physics 21(3/4):219–253. Plenum Press, New York, NY, USA, 1982 MathSciNetMATHCrossRef
Zurück zum Zitat R.A. Jakobs. Increased Rates of Convergence Through Learning Rate Adaption. Neural Networks 1:295–307. Pergamon Press, Oxford, United Kingdom, 1988 CrossRef R.A. Jakobs. Increased Rates of Convergence Through Learning Rate Adaption. Neural Networks 1:295–307. Pergamon Press, Oxford, United Kingdom, 1988 CrossRef
Zurück zum Zitat A. Pinkus. Approximation Theory of the MLP Model in Neural Networks. Acta Numerica 8:143–196. Cambridge University Press, Cambridge, United Kingdom, 1999 MathSciNetCrossRef A. Pinkus. Approximation Theory of the MLP Model in Neural Networks. Acta Numerica 8:143–196. Cambridge University Press, Cambridge, United Kingdom, 1999 MathSciNetCrossRef
Zurück zum Zitat M. Riedmiller and H. Braun. Rprop—A Fast Adaptive Learning Algorithm. Technical Report, University of Karlsruhe, Karlsruhe, Germany, 1992 M. Riedmiller and H. Braun. Rprop—A Fast Adaptive Learning Algorithm. Technical Report, University of Karlsruhe, Karlsruhe, Germany, 1992
Zurück zum Zitat M. Riedmiller and H. Braun. A Direct Adaptive Method for Faster Backpropagation Learning: The RPROP Algorithm. Int. Conf. on Neural Networks (ICNN-93, San Francisco, CA), 586–591. IEEE Press, Piscataway, NJ, USA, 1993 CrossRef M. Riedmiller and H. Braun. A Direct Adaptive Method for Faster Backpropagation Learning: The RPROP Algorithm. Int. Conf. on Neural Networks (ICNN-93, San Francisco, CA), 586–591. IEEE Press, Piscataway, NJ, USA, 1993 CrossRef
Zurück zum Zitat D.E. Rumelhart, G.E. Hinton and R.J. Williams. Learning Representations by Back-Propagating Errors. Nature 323:533–536, 1986 CrossRef D.E. Rumelhart, G.E. Hinton and R.J. Williams. Learning Representations by Back-Propagating Errors. Nature 323:533–536, 1986 CrossRef
Zurück zum Zitat T. Tollenaere. SuperSAB: Fast Adaptive Backpropagation with Good Scaling Properties. Neural Networks 3:561–573, 1990 CrossRef T. Tollenaere. SuperSAB: Fast Adaptive Backpropagation with Good Scaling Properties. Neural Networks 3:561–573, 1990 CrossRef
Zurück zum Zitat D. Touretzky, G. Hinton and T. Sejnowski (eds.) Proc. of the Connectionist Models Summer School (Carnegie Mellon University). Morgan Kaufman, San Mateo, CA, USA, 1988 D. Touretzky, G. Hinton and T. Sejnowski (eds.) Proc. of the Connectionist Models Summer School (Carnegie Mellon University). Morgan Kaufman, San Mateo, CA, USA, 1988
Zurück zum Zitat P.J. Werbos. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Ph.D. Thesis, Harvard University, Cambridge, MA, USA, 1974 P.J. Werbos. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Ph.D. Thesis, Harvard University, Cambridge, MA, USA, 1974
Metadaten
Titel
Multi-Layer Perceptrons
verfasst von
Rudolf Kruse
Christian Borgelt
Frank Klawonn
Christian Moewes
Matthias Steinbrecher
Pascal Held
Copyright-Jahr
2013
Verlag
Springer London
DOI
https://doi.org/10.1007/978-1-4471-5013-8_5

Premium Partner