CogSci 30PY Project
The IT Cortex
Back to A Hybrid Neural Network
Model of Binocular Rivalry |
"The brain is not an organ of thinking but an organ
of survival, like claws and fangs. It is made in such a way as to
make us accept as truth that which is only advantage. It is an exceptional,
almost pathological constitution one has, if one follows thoughts
logically through, regardless of consequences. Such people make martyrs,
apostles, or scientists, and mostly end on the stake, or in a chair,
electric or academic." -Albert Szent-Gyorgi |
http://www.comp.leeds.ac.uk/ugadmit/cogsci/tech/hopn.htm:
Hopfield nets can remeber 0.14N states stably, where N is the number
of units in the net.
http://www.ida.his.se/ida/kurser/ai_ann/kursmaterial/archive/lectureA.pdf:
Synchronous updates may lead to oscillation and the network may not always
converge.
Is this what we need for binocular rivalry?
http://www.nada.kth.se/~asa/hopfield.ps:
for synchronous updating the network may end up flipping between two (but
not more) states.
--------------------------------------------------------------------------------------
David J.C. MacKay
Information Theory, Inference, and Learning Algorithms
Ch 32/33 Exact
Monte Carlo Sampling / Variational Methods Ch 42 references 33.2, but
it is all maths!
Ch 39 The
Single Neuron as classifier
Ch 40 Capacity
of a Single Neuron
Ch 41 Learning
as inference
Ch 42 Hopfield
Networks - Explanation starts with Hebbian learning -> Hopfield
NNs!
Ch 43 Boltzmann
Machines
Hopfield
Networks / Boltzmann machines
This book
is also available from Amazon
-------------------------------------------------------------------------
C
Files for the book "An Introduction to Neural Networks"
by
James A. Anderson,
Department of Cognitive and Linguistic Sciences,
Brown University, Providence, RI 02912., on MIT Press.
-------------------------------------------------------------------------
Hopfield
example code from "Neural Networks at Your Fingertips"
--------------------------------------------------------------------------
Background
info
Hopfield
Networks at Wolfram Research:
Hopfield networks, or associative networks, are typically used for classification.
Given a distorted input vector, the Hopfield network associates it with
an undistorted pattern stored in the network.
--------------------------------------------------------
NASATech
says: The second architecture is that of autoassociative neural networks,
which, in nature, enable organisms to recall old memories from partial
or noisy stimuli. In an autoassociative network, each neuron is connected
to every other neuron in the network. The excitability of any such neuron
is determined by the state of all the other neurons and the strengths
(weights) of the interconnections between neurons. For any state, the
pattern of activation at the next instant in time is completely determined
by the preset weights. The number of different firing patterns is 2^n,
where n is the number of neurons in the net; because this number is finite,
as the sequence of firing patterns proceeds, cycles are inevitable. Cycles
that consist of repeated instances of the same pattern are called fixed
points, and these fixed points can be chosen by setting the weights to
make the points attract nearby patterns. The fixed points represent memories.
An autoassociative sensor web could be used to search for known gaseous,
biological, or geological signatures, for example.
A
chapter on Attractor Networks. This is a good explanation of some
of the terms (like "attractor basin").
Oxford Lecture
notes on Hopfield Networks and Alzheimer's. Here's a cut from it:
Fusi's PhD
thesis, which I've linked to before, seems relevant:
It is about the IT
cortex;
context-dependent
memory;
recognition but also a no-recognition state (i.e.,
a spontaneous activity state)
-- p. 6, bottom.
Following are some extracts from his thesis.
From the thesis. Page numbers are from the
pdf, not the printed numbers on the page. (My italics):
Another key property of slow learning is that
memory of the class prototypes is
always stronger than memory of any other
members of the same class which
have been shown. This
is essential for the classification
capability and for reproducing the behavior
of the paradigmatic attractor cell. (p. 11)
...the process of learning can be fully described
in terms of a few parameters which are the
transition probabilities for LTP and LTD. (p. 11)
...learning requires many presentations of the same
stimulus... (p. 11) (over 500)
Stochastic learning provides a good tool for studying
quantitatively the process of attractor
formation without specifying the details of the synaptic dynamics. However,
to proceed
further, it is important to have a framework that provides a link between
the transition
probabilities and the neural activity driven synaptic dynamics. This implies
the identification of the source of noise which drives the stochastic
mechanism. The solution we propose is suggested by the analysis of cortical
recordings: the spike trains recorded in vivo are similar to poissonian
processes [Softky Koch 1993] and the variability in the intervals between
the spikes of pre- and post-synaptic neurons is a quantity that can be
\read" by the synapses. What we show in chapter 7.1 is that it is
possible to use this source of
stochasticity to obtain transitions with small probabilities. (p. 13)
Persistent enhanced spike rates have been observed
throughout the delay interval between
two successive visual stimulations in several cortical areas: e.g. they
are encountered in
infero-temporal (IT) cortex [Fuster, 1990, Fuster & Jervey, 1981,
Miyashita & Chang, 1988,
Miyashita, 1988, Sakai & Miyashita, 1991, Miller et al, 1993, Nakamura
& Kubota, 1995],
in pre-frontal (PF) cortex [Funahashi et al, 1989, Wilson et al. 1993,
Miller et al, 1996,
Scalaidhe et al, 1997, Rao et al. 1997] and in posterior parietal cortex
[Koch &Fuster, 1989]
(for a review see [Fuster, 1995]). The phenomenon is detected in single
unit extra-cellular
recordings of spikes following, rather than during, the presentation of
a sensory stimulus. It
has been found in primates trained to perform a delayed response task,
in which the behav-
ing monkey must remember the identity or location of an initial eliciting
stimulus in order
to decide upon its behavioral response. The pattern of delay activity
is considered a neural
correlate of working memory, which is related to the ability of the
animal to actively hold an
item in memory for a short time. In fact, lesions in IT and PF are known
to produce impair-
ments in spatial and associative memory tasks [Murrayet al. 1993, Gutnikov
et al. 1997],
and a ect recognition of an object after a delay in both humans and monkeys.
(p. 15)
The collective aspect is expressed in the mechanism
by which a stimulus that had
been learned previously, has left a synaptic engram of potentiated
excitatory synapses
connecting the cells driven by the stimulus. When this subset of cells
is re-activated by
a sensory event, they cooperate to maintain elevated ring rates in the
selective set of
neurons, via the same set of potentiated synapses, after the stimulus
is removed. In this
way, these cells can provide each other, i.e. each one of the group, with
an a afferent signal
that clearly differentiates the members of the group from other cells
in the same cortical
module. (p. 18)
Attractor dynamics (associative memory) implies that each pattern of
delay activity is
excited by a whole class of stimuli. Each of these stimuli raises the
rates of a sufficient
number of cells belonging to the pattern of activity, so that via the
learned collateral
synaptic matrix they can excite the entire subset of neurons characterizing
the attractor
and maintain them with elevated rates when the stimulus is removed. (p.
25)
The class of stimuli leading to the same persistent distribution is the
basin of attraction of that particular attractor. (pp. 25-26) As one moves
in the space of stimuli, at some point the boundary of the basin
of attraction of the attractor is reached. Moving even slightly beyond
this boundary,
the network will relax either into another learned attractor, or to the
grand attractor of
spontaneous activity. See e.g. [Amit & Brunel, 1997a, Amit, 1995].
(p. 26)
The dynamics of the simplest integrate-and-fire (IF) neuron can
be fully described in terms
of a single internal variable: the depolarization V (t) of the neural
membrane. The neuron
integrates the afferent current, and when the depolarization crosses a
threshold, the neuron
emits a spike. The depolarization is then reset to the hyperpolarization
value H, and the
integration process starts again after an absolute refractory period Tarp.
Formally, the differential equation which governs the dynamics below threshold,
can be
written as follows:
dV (t) / dt = -L(t1V) + I(t),
where I(t) is the net charging current, expressed in units of potential
per unit time, produced by afferent spikes, and L(t) is the leakage term.
Usually it is proportional to the depolarization: L(t) = -V(t)/r, where
r is the integration time constant of the membrane depolarization. A linear
constant decay, not dependent
on V, L(t) = B simplifies the analysis of the collective behavior of the
network without sacrificing those features that allow to reproduce the
phenomenology of cortical recordings [Fusi & Mattia 1999].
(p. 31. has properly formatted equations)
(p. 32)
This neuron he calls the LIF neuron (Linear Integrate-and-Fire neuron).3.1.1
The anatomy of the network module
We consider a local module that is a large network containing a number
of neurons of
the order of 10^4 - 10^5 . The number of connections between neurons of
the same module is
higher than the number of connections between neurons belonging to different
modules.
Each neuron is randomly connected to a subset of neurons of the same module
through the
recurrent collaterals, and to neurons of other modules through longer
range connections
(see Fig.3.1).
In particular, the network module is composed of excitatory (pyramidal)
and inhibitory
(inter-neurons) neurons. Each neuron has some probability (connectivity
level) of a
synaptic contact with any other neuron (excitatory or inhibitory) of the
same module, and
receives external excitatory afferents from pyramidal neurons of other
modules.
The connectivity structure can be fully described in terms of two sets
of variables:
the binary variables cij of having a connection between neuron j (pre-synaptic
neurons)
and neuron j (post-synaptic neuron) and the corresponding synaptic weight
Jij . In what
follows the cij are generated at random and never changed, while the synaptic
couplings
Jij can vary to acquire information about the stimuli that excite part
of the neurons of the
network module. (p. 32)
SD = signal dominated (p. 33)
ND = noise dominated (p. 34)
Each neuron will receive a current Ii(t) (i denotes the index of the
neuron) given by:
(p. 32)
where cij is 1 if there exists a synaptic contact between neuron j and
neuron i, Jij is
the corresponding synaptic strength, and the sum extends over all the
Nall neurons of the
populations (recurrent or external) that are interacting.
"Such being the nature of our subject and such our way of arguing
in our discussion of it, we must be satisfied with a rough outline of
the truth."
-Aristotle
|