Professional Documents
Culture Documents
Reading Assignments:
HBTNN:
I.3 Dynamics and Adaptation in Neural Networks (Arbib)
III. Associative Networks (Anderson)
III. Energy Functions for Neural Networks (Goles)
TMB2:
8.2 Connectionist Models of Adaptive Networks
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 1
Hopfield Networks
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 2
Hopfield Networks
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 3
“Energy” of a Neural Network
E = - ½ Σ ij sisjwij + Σ i siθi
If we pick unit i and the firing rule (previous slide) does not
change its si, it will not change E.
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 4
si: 0 to 1 transition
∆E = - ½ Σj (wijsj + wjisj) + θi
= - (Σ j wijsj - θi) (by symmetry)
≤ 0.
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 5
si: 1 to 0 transition
∆E = Σj wijsj - θi < 0
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 6
Minimizing Energy
We stress that there may be different such states — they are local
minima. Global minimization is not guaranteed.
Basin of
Attraction for C
A
B
D
E
C
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 7
The Symmetry Condition wij = wji is crucial for ∆E ≤ 0
In a control problem: a link wij might express the likelihood that the
action represented by i should precede that represented by j, and thus
wij = wji is normally inappropriate.
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 8
The condition of asynchronous update is crucial
1
Unit "0" 0.5 0.5 Unit "1"
1
Consider the above simple "flip-flop" with constant input 1, and
with w12 = w21 = 1 and θ1 = θ2 = 0.5
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 9
The condition of asynchronous update is crucial
1
Unit "0" 0.5 0.5 Unit "1"
we have E(0,0) = 0
E(0,1) = E(1,0) = 0.5
E(1,1) = 0
and the Hopfield network will converge to the minimum
at (0,0) or (1,1).
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 10
Hopfield Nets and Optimization
Hopfield and Tank 1986 have constructed VLSI chips for such networks which
do indeed settle incredibly quickly to a local minimum of E.
Unfortunately, there is no guarantee that this minimum is an optimal solution to
the traveling salesman problem. Experience shows it will be "a pretty good
approximation," but conventional algorithms exist which yield better
performance.
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 11
The traveling salesman problem 1
exp(1) = 2.72
Let neuron Nij express the decision to go straight from city i to city j.
The cost of this move is simply lij.
We can re-express the "visit a city only once" criterion by saying that,
for city j, there is one and only one city i from which j is directly
approached. Thus (ΣiNij-1)2 can be seen as a measure of the extent to
which this constraint is violated for paths passing on from city j.
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 14
Constraint Optimization Network
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 15
The traveling salesman problem 3
Compare this to the general energy expression (with si now replaced by Nij):
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 17
TSP Network Connections
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 18
Boltzmann Machines
Attraction for C
h A
B
D
E
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 19
Boltzmann’s statistical theory of gases
Hinton, Sejnowski and Ackley’s idea was that this distribution might
also be used to describe neural interactions, where low temperature T
is replaced by a small noise term T (the neural analog of random
thermal motion of molecules).
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 20
Boltzmann Distribution
At thermal equilibrium at temperature T, the
Boltzmann distribution gives the relative
probability that the system will occupy state A vs.
state B as:
P( A) æ E ( A) − E ( B ) ö exp( E ( B ) / T )
= expç − ÷=
P( B) è T ø exp( E ( A) / T )
where E(A) and E(B) are the energies associated with states
A and B.
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 21
Simulated Annealing
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 22
Simulated Annealing in Hopfield Nets
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 23
Statistical Mechanics of Neural Networks
Hertz, J., Krogh. A., and Palmer, R.G., 1991, Introduction to the
Theory of Neural Computation, Santa Fe Institute Studies in the
Sciences of Complexity, Addison-Wesley.
The book is quite mathematical, but has much accessible material, exploiting
the analogy between neuron state and atomic “spins” in a magnet.
Laurent Itti: CS564 - Brain Theory and Artificial Intelligence. Hopfield Networks 24