The title sounds outrageous. But supporting data comes from
research at the most prestigious Salk Institute. Other researchers had made
enormous storage capacity estimates for brains, but this new estimate is 10
times greater. The estimate is on the order of petabytes, as much as the entire
World Wide Web.
How does anybody come up with such estimates? What is the
basic premise? First, memory is quantified in terms of number of bits of
information that can be stored and recovered. In the case of brain, the
question is how much information can be stored in a synapse, the communicating
junction between neurons. Size and operational strength of a synapse are the
key variables: strength can be measured in bits and strengths correlate with umber
and size synaptic size. Under high magnification, synapses look like a bunch of
beads on a string. The newborn brain, there are relatively few “beads,” but
these increase in number and size as the baby grows and learns things. Sadly,
in old age, many of these beads disappear unless the brain is kept very active.
In the Salk study of rat brains, electron micrographs of the
memory-forming structure, the hippocampus, allowed 3-D reconstruction and
detection of the diameters of synapses, which are the target synaptic
structures in neurons. Synaptic strength correlates with storage capacity, and
the strength is measured by the size of synapses, which appear as a round bead
attached to a stalk or neck to the
supporting neuronal membrane. The size of the bead varies with synaptic
strength because synaptic strength is created by more molecular machinery for
mediating synaptic communication. Thus, spine bead size is a proxy for synaptic
strength and storage capacity.
The investigators found that a small cube of brain tissue
contained 26 different bead sizes, each associated with a distinct synaptic
strength. The authors state that this equates to an approximate 4.7 bits of
information at each synapse. Multiply 4.7 times the trillions of synapses and neurons
in the brain and you get phenomenal storage capacity.
While I marvel at the elegant complexity of these research
methods, I think the interpretations are a bit simplistic. There are some
caveats that the authors overlooked. For one, an assumption is made that that
the number of storage bits equals the logarithm of the number of bead sizes. The
“bit” is a unit of information in digital communications. Information theory
holds that one bit is the probability of a binary random variable that is 0 or
1, or more loosely defined as present or absent. One has to take some liberties
to apply this concept to memory storage in brain, because the brain is not a
digital computer. It is an analog biological computer.
Then there is the problem that the hippocampus deals only
with forming declarative and episodic memories, not procedural memories like
touch typing or playing the piano. Thus, the storage capacity, whatever it is,
is not estimated for procedural memories. Secondly, declarative and episodic
memories are not stored in the hippocampus, but rather stored in a distributed
fashion throughout the brain. Since synaptic measures were made only on
hippocampal tissue, there are no data for the rest of the brain.
But there is a larger issue. How does one know how many bits
it takes to represent different memories? Not all memories are the same and
surely they don't all require the same number of storage bits.
Actually, the exact number of bits of information that brains
can store is rather irrelevant. By any measure, common experience teaches that
nobody utilizes all their memory capacity. Moreover, the amount of information
a given person stores varies profoundly depending on such variables as
motivation to remember things, use of mnemonics, and level of education. The
question that needs answering, given that we have vast amounts of unused
storage capacity, is "Why don't we remember more than we do?" Books like my Memory Power 101 provide some practical answers.
Source:
Bartol, Thomas M. et al. (2015). Neuroconnectomic upper bound on the
variability of synaptic plasticity.
eLIfe. Nov. 30. http://dx.doi.org/10.7554/eLife.10778