Some comments on Raymond G. Van De Walker's epistle:
> Let me show you the numbers: The human brain has 10 billion
Most recent estimtes put it at 100 billion.
> Now, I thought about ways to reduce this by editing the system,
> but they won't work. Like most real computing systems, the majority of
> the logic (>95%) by weight or volume is I/O. (The cerebrum, cerebellum,
> gyrii, and most of the encephalon) Neural networks are great for I/O:
> they're robust and compact compared to the digital systems they replace.
> You would not want to use anything else to construct the phenomenal
> systems of a robot.
Hmm. One way of estimating the brain's processing power is like this:
The human brain contains about 10^11 neurons. Each neuron has about 5*10^3 synapses, and signals are transmitted along these synapses at an average frequency of about 10^2 Hz. Each signal contains, say, 5 bits. This equals 10^17 ops.
Another method, used by Moravec, is to look at one part of the nervous system whose function we can replicate on computers today (the retina). Then we multiply the resources needed for this computation by the factor by which the total brain is larger than the retina. This gives us the figure 10^14 ops for the brain, three magnitudes less than the first estimate.
The second estimate presupposes that we can make some optimizations. Maybe intelligent design and new computational elements enable us to do several orders of magnitude better than mother nature with the same number of ops. Why couldn't the same be possible with regard to memory requirements? There is no evidence that the brain's memory system is not highly redundant, so that with highly reliable artificial (or simulated) neurons one get away with a lot less memory. We simply don't know.
Another problem with the multiplicative approach - multiplying neurons, synapeses per neuron, and resources per synapse - is that it might turn out that a lot of the brain's computing power is in higher-order interactions in the dendritic trees. Signals may not only be added, but multiplied, and with interesting time-integration effects that might be used by the brain. The multiplicative approach neglects these potential effects, and may thus understimate the brain's computing power.
> Using Drexler's estimates for random-access memory
> (20MBytes/cubic micron), we can fit 305 of 64K computers in a cubic
> micron. The computers therefore take roughly 9.8x10^4 cubic microns.
Does this take into account that there might be a need for cooling?
> 11,300 cubic microns is small. It's a cube about 22.5 microns on
> a side, say a quarter-millimeter on a side, about 1/8 the size of a
> crystal of table salt. 17,300 cubic microns (storing synaptic addresses)
> is still small, about 25.9 microns on a side. Even 34,600 cubic microns
> (double everything) is small, maybe 32.6microns on a side, the size of a
> crystal of table salt.
As an estimate of precisely how small advance nanotech could make a human-equivalent computer I think we have to take this grain of salt with a grain of salt so to speak. An uncertainty interval of a couple of orders of magnitude does not seem unreasonable.
http://www.hedweb.com/nickb email@example.com Department of Philosophy, Logic and Scientific Method London School of Economics