Skip to main content

1995 | OriginalPaper | Buchkapitel

A Transformation for Implementing Efficient Dynamic Backpropagation Neural Networks

verfasst von : George L. Rudolph, Tony R. Martinez

Erschienen in: Artificial Neural Nets and Genetic Algorithms

Verlag: Springer Vienna

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Most Artificial Neural Networks (ANNs) have a fixed topology during learning, and often suffer from a number of short-comings as a result. Variations of ANNs that use dynamic topologies have shown ability to overcome many of these problems. This paper introduces Location-Independent Transformations (LITs) as a general strategy for implementing distributed feedforward networks that use dynamic topologies (dynamic ANNs) efficiently in parallel hardware. A LIT creates a set of location-independent nodes, where each node computes its part of the network output independent of other nodes, using local information. This type of transformation allows efficient support for adding and deleting nodes dynamically during learning. In particular, this paper presents an LIT for standard Backpropagation with two layers of weights, and shows how dynamic extensions to Backpropagation can be supported.

Metadaten
Titel
A Transformation for Implementing Efficient Dynamic Backpropagation Neural Networks
verfasst von
George L. Rudolph
Tony R. Martinez
Copyright-Jahr
1995
Verlag
Springer Vienna
DOI
https://doi.org/10.1007/978-3-7091-7535-4_13

Neuer Inhalt