Abstract | ||
---|---|---|
The author presents a survey of the basic theory of the backpropagation neural network architecture covering architectural design, performance measurement, function approximation capability, and learning. The survey includes previously known material, as well as some new results, namely, a formulation of the backpropagation neural network architecture to make it a valid neural network (past formulations violated the locality of processing restriction) and a proof that the backpropagation mean-squared-error function exists and is differentiable. Also included is a theorem showing that any L/sub 2/ function from (0, 1)/sup n/ to R/sup m/ can be implemented to any desired degree of accuracy with a three-layer backpropagation neural network. The author presents a speculative neurophysiological model illustrating how the backpropagation neural network architecture might plausibly be implemented in the mammalian brain for corticocortical learning between nearby regions of the cerebral cortex.<> |
Year | DOI | Venue |
---|---|---|
1988 | 10.1109/IJCNN.1989.118638 | Neural networks for perception (Vol. 2) |
Keywords | DocType | Volume |
backpropagation neural network | Journal | 1 |
Issue | ISSN | ISBN |
Supplement-1 | Neural Networks | 0-12-741252-2 |
Citations | PageRank | References |
301 | 47.33 | 0 |
Authors | ||
2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Robert Hecht-Nielsen | 1 | 448 | 110.50 |
Hecht-Nielsen, R. | 2 | 301 | 47.33 |