Professional Documents
Culture Documents
Fall 2006
1
Biological Inspiration for RBFs
The nervous system contains many examples of
neurons with local or tuned receptive fields.
Sigmoidal unit:
y j = tanh
i
w ji x i
Decision boundary is a hyperplane
Gaussian unit:
y j = exp
x j2
2
j
Decision boundary is a hyperellipse
3
RBF = Local Response Function
4
RBF Network
wj
gaussian
RBF units
Output = w jexp
j j2
x
j
2
5
Tiling the Input Space
6
Properties of RBF Networks
7
RBFs and Parzen Windows
9
Training an RBF Network
10
RBF Demo
matlab/rbf/rbfdemo
Regularly
spaced
gaussians
with fixed 2
11
Training Tip
12
Early in Training
13
Training Complete
14
Random Gaussians
15
After Training
16
Locality of Activation
17
Locality of Activation
18
Winning in High Dimensions
19
How to Place RBF Units?
20
k-Means Clustering Algorithm
21
Online Version of k-Means
j x
j
i
23
Using SOFM to Pick RBF Centers
24
Determining the Variance 2
1) P-nearest-neighbor heuristic:
Set each j so that there is a certain amount of
overlap with the P closest neighbors of unit j.
25
Phoneme Clustering (338 points)
Trajectory of
cluster center.
Variances set
for overlap P=2.
26
Phoneme Classification Task
Moody & Darken (1989): classify 10 distinct vowel
sounds based on F1 vs. F2.
338 training points; 333 test points.
27
Defining the Variance
j2
x u
dj =
2j
29
Transforming the Input Space
30
Smoothness Problem
3 4
x
At point x neither RBF unit is very
active, so the output of the network
sags close to zero. Should be 3.5.
31
Assuring Smoothness
y jw j
j
Output =
yj
j
x
3 4
Smooth interpolation
along this line.
No output sag in the
middle.
32
Training RBF Nets with Backprop
E E E
Calculate , , and .
j j wj
33
Summary of RBFs
RBF units provide a new basis set for synthesizing an
output function. The basis functions are not
orthogonal and are overcomplete.
RBFs only work well for smooth functions.
Would not work well for parity.
Overlapped receptive fields give smooth blending of
output values.
Training is much faster than backprop nets: each
weight layer is trained separately.
34
Summary of RBFs
Hybrid learning algorithm: unsupervised learning
sets the RBF centers; supervised learning trains the
hidden to output weights.
RBFs are most useful in high-dimensional spaces.
For a 2D space we could just use table lookup and
interpolation.
In a high-D space, curse of dimensionality important.
OCR: 16 x 16 pixel image = 256 dimensions.
Speech: 5 frames @ 16 values/frame = 80 dimensions.
35
Psychological Model: ALCOVE
36
Category Learning
Train humans on a toy classification problem. Then
measure their generalization behavior on novel
exemplars.
37
ALCOVE Equations
Hiddens: a hid
j
[
= exp c ih ji
j
i
]
in r q / r
Category: a out
k = kj j
w a hid
j
is a mapping constant
(softmax)
38
Dimensional Attention
o x
o x
o x
o x
39
Dimensional Attention
x
x
x
o
o
o
o
Because ALCOVE does not use a full covariance
matrix, it cannot shrink or expand the input space
along directions not aligned with the axes. However,
for cognitive modeling purposes, a diagonal covariance
matrix appears to suffice.
40
Disease Classification Problem
Terrigitis
Midosis
{ } T
Novel
Items:
Test Set
} M
Humans and ALCOVE: N3,N4 > N1,N2 and N5,N6 > N7,N8 41