A statistical analysis of neural computation
- Creators
- Cortese, John A.
- Goodman, Rodney M.
Abstract
This paper presents an architecture and learning algorithm for a feedforward neural network implementing a two pattern (image) classifier. By considering the input pixels to be random variables, a statistical binary hypothesis (likelihood ratio) test is implemented. A linear threshold separates p[X|H_0] and p[X|H_1], minimizing a risk function. In this manner, a single neuron is considered as a BSC with the pdf error tails probability ε. A Single layer of neurons is viewed as a parallel bank of independent BSC's, which is equivalent to a single effective BSC representing that layer's hypothesis testing performance. A multiple layer network is viewed as a cascade of BSC channels, and which again collapses into a single effective BSC.
Additional Information
© 1994 IEEE.
Attached Files
Published - 00394753.pdf
Files
Name | Size | Download all |
---|---|---|
md5:7db496720c6e0be391b532d4e8bdabb7
|
75.5 kB | Preview Download |
Additional details
- Eprint ID
- 93886
- DOI
- 10.1109/isit.1994.394753
- Resolver ID
- CaltechAUTHORS:20190315-142359458
- Created
-
2019-03-15Created from EPrint's datestamp field
- Updated
-
2021-11-16Created from EPrint's last_modified field