
It is well-known that the computing power of natural and artificial neural networks arises from massive parallelism of simple "computing elements" [l]. Both natural neural networks and future molecular devices are "analog" computers subject to inaccuracies and some randomness, in contrast to present digital simulations. In this presentation, we show that trained artificial neural networks can be made of relatively inaccurate molecular components without adversely affecting their input-output behavior. The redundancy offered by parallelism in these networks overcomes inaccuracies in "computations" performed by any single component and yields accurate input-output behavior. This complements previous results on parallelism.
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 1 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
