Optimal Brain Damage
LeCun, Yann, Denker, John S., Solla, Sara A.
–Neural Information Processing Systems
We have used information-theoretic ideas to derive a class of practical andnearly optimal schemes for adapting the size of a neural network. By removing unimportant weights from a network, several improvementscan be expected: better generalization, fewer training examples required, and improved speed of learning and/or classification. The basic idea is to use second-derivative information tomake a tradeoff between network complexity and training set error. Experiments confirm the usefulness of the methods on a real-world application. 1 INTRODUCTION Most successful applications of neural network learning to real-world problems have been achieved using highly structured networks of rather large size [for example (Waibel, 1989; Le Cun et al., 1990a)]. As applications become more complex, the networks will presumably become even larger and more structured.
Neural Information Processing Systems
Dec-31-1990
- Country:
- Europe > Switzerland > Zürich > Zürich (0.14)
- Industry:
- Health & Medicine > Therapeutic Area > Neurology (0.43)
- Technology: