Skip to main content

1995 | OriginalPaper | Buchkapitel

Massively Parallel Training of Multi Layer Perceptrons With Irregular Topologies

verfasst von : D. Koll, M. Riedmiller, H. Braun

Erschienen in: Artificial Neural Nets and Genetic Algorithms

Verlag: Springer Vienna

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

In this paper we present an approach to the training of feed forward neural networks on massively parallel SIMD-architectures. In order to cover a wide field of applications we focus our attention on the flexibility of the load balancing routines. Our approach is characterized by three important properties: 1. All four types of parallelism inherent in the training phase are used. 2. In a preprocessing step neural networks are transformed into equivalent topologies, more suited for parallel computation. 3. Each learning task can be parallelized in a number of different ways, the best of which is chosen according to estimations of the computing efficiency.Following these concepts we developed PINK2, a massively parallel simulator kernel for the MasPar MP1216. In contrast to most known approaches, efficient only for special topologies, it achieves good computing performance on a broad range of differing benchmark problems.

Metadaten
Titel
Massively Parallel Training of Multi Layer Perceptrons With Irregular Topologies
verfasst von
D. Koll
M. Riedmiller
H. Braun
Copyright-Jahr
1995
Verlag
Springer Vienna
DOI
https://doi.org/10.1007/978-3-7091-7535-4_77

Neuer Inhalt