English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

On the Asymptotic Information Storage Capacity of Neural Networks

MPS-Authors
/persons/resource/persons225707

Palm,  G
Former Department Structure and Function of Natural Nerve-Net , Max Planck Institute for Biological Cybernetics, Max Planck Society;
Max Planck Institute for Biological Cybernetics, Max Planck Society;

Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available
Citation

Palm, G. (1989). On the Asymptotic Information Storage Capacity of Neural Networks. In R. Eckmiller, & C. von der Malsburg (Eds.), Neural Computers (pp. 271-280). Berlin, Germany: Springer.


Cite as: https://hdl.handle.net/21.11116/0000-0006-9F67-4
Abstract
Neural networks can be useful and economic as associative memories, even in technical applications. The asymptotic information storage capacity of such neural networks is defined and then calculated and compared for various local synaptic rules. It turns out that among these rules the simple Hebb rule is optimal in terms of its storage capacity. Furthermore the capacity of the clipped Hebb rule (C = In 2) is even higher than the capacity of the unclipped Hebb rule (C = 1/(8-ln 2)).