1993 | OriginalPaper | Buchkapitel
A Scalable Neural Architecture Combining Unsupervised and Suggestive Learning
verfasst von : R. B. Lambert, W. P. Cockshott, R. J. Fryer
Erschienen in: Artificial Neural Nets and Genetic Algorithms
Verlag: Springer Vienna
Enthalten in: Professional Book Archive
Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.
Wählen Sie Textabschnitte aus um mit Künstlicher Intelligenz passenden Patente zu finden. powered by
Markieren Sie Textabschnitte, um KI-gestützt weitere passende Inhalte zu finden. powered by
Multi-layered perceptions are, in theory, capable of solving a wide range of problems. However, as the scale of many problems is increased, or requirements change, multi-layered perceptrons fail to learn or become impractical to implement. Self-organizing networks are not so limited by scale, but require a-priori information, typically in the form of preset weights or suitable control parameters to achieve a good categorization of a data set.Based on research into the behaviour of biological neurons during learning, a new self-organizing neural network has been devised. Moving away from the traditional McCulloch and Pitts model, each neuron stores several independent patterns, each capable of initiating a neuron output. By structuring such neurons into a network, a rapid and equal distribution of data across competitive nodes is possible.This paper introduces the new network, known as a Master-Slave architecture, and learning paradigm. By using competitive and suggestive learning, inputs are distributed across all available classification units, without the need for a-priori knowledge. Two experiments are described, highlighting the potential of the master-slave architecture as a building block for larger networks.