Professional Documents
Culture Documents
skill acquisition
email: andrew.heathcote@newcastle.edu.au
Abstract
We outline SEEXC, a neural network model of choice response time (RT) based on
leaky competitive integrators. SEEXC is different from extant neutral network
models in that it incorporates the effects of practice by modulating recurrent selfconnection weights. For simplified versions of this model, we provide analytic and
numeric results concerning RTs and the relationship between RT and practice the
Law of Practice that match those observed empirically. We also show that
previous methods of modelling practice in similar systems, which modulate inputs,
are unlikely to successfully match observed data. The simplified versions of the
model analysed are appropriate for modelling the non-stochastic parts of simple RT
and two-choice RT, provide insight into the behaviour of the full version of SEEXC,
and suggest a new form for the Law of Practice.
10
Response time
A ctivation
Time
x1 = I 1 kx1 + f ( x1 ) f ( x 2 )
(1)
x 2 = I 2 kx 2 + f ( x 2 ) f ( x1 )
(2)
Of course real choice RT data contains both decision and non-decision components. We neglect the
x0
x<0
( N ) = k (1 e N )
(3)
x s = I s Sx s
(4)
xd = I d Dxd
(5)
S=k-+
(6)
D=k--
(7)
Note that (4) and (5) hold only until quenching; that is, only while the activation of
both units is strictly positive and hence f(x) x.
The values of S and D determine the properties of the sum and difference
dynamics. Given that in SEEXC k, S > 0 and so the sum of activations converges
to xs = IS/S. Similarly, when k > + , D > 0 and the difference converges to Id/D.
However, as practice increases k, and so D < 0 as > 0. Hence if k > practice
can cause a transition from converging to diverging differences, but if > k the
difference diverges at all levels of practice. Diverging differences guarantee a
response even with very small differences between inputs, and in the stochastic model
a response will occur even with exactly equal inputs. Generally the sum will
converge quickly relative to the difference, as illustrated in Figure 1, because typically
Is > Id.
Equations (4) and (5) are examples of initial value problems of the form
t=
1 A By
B 0;
ln
B A By 0
t=
y y0
A 0, B = 0
A
(8)
From (4), (5) and (8), and using xs(0)=2x0 and xd(0)=0 yields solutions for the sum
(where S > 0 always applies) and difference equations expressed as functions of x:
t=
1 I s Sx s
ln
S I s 2 Sx0
(9)
7
1 I d Dxd
ln
D
Id
t=
D0
t=
xd x0
D=0
Id
(10)
xs =
Is
I
+ 2 x 0 s e St
S
S
=
Id
D
(1
Dt
(11)
D0
xd = I d t D = 0
(12)
Since we have assumed t < tq, we can use x1=(xs+xd) to find an expression
for activation in terms of time:
I
I
I
I
x 1 = s + d + x 0 s e St d e Dt D 0
2D
2S
2D
2S
x1 =
Idt
I
I
+ s + x 0 s e St
2
2S
2S
(13a)
D=0
(13b)
The equation for x2 is the same as (13), except with Id replaced by -Id.
Quenching occurs when x2 = 0. Hence, tq is the value of t for which:
I
I
I
Is
d + x 0 s e St + d e Dt = 0 D 0
2D
2S
2D
2S
Idt
I
I
+ s + x0 s
2
2S
2S
St
= 0
e
D=0
(14a)
(14b)
A solution is available for (14) only when x2(t=) < 0. This condition is guaranteed
when D 0 (14a), but will only be satisfied for D > 0 when I 1 I 2 > (k ) ,
otherwise competition is not sufficiently strong to quench the losing unit, which has
an asymptotic activation x2 (t = ) = (I s S ) (I d D ) > 0 . As noted above, all results
so far are valid only for t< tq.
5
4
3
Activation
2
1
0
Activation
10
20
30
40
50
60
(a)
(b)
10
12
14
Figure 2. Solid lines indicate the unit with larger input, dashed lines the unit with the
smaller input, red lines indicate N = 0 (prior to practice) and green N = 50 (after
extensive practice). Plots are terminated at quenching.
In Figure 2b the difference in inputs is much larger (Id = 0.1), but the other
parameters are the same. Activation changes more quickly and quenching occurs
much earlier (note that the range of t in 2a is four times the range in 2b). The
difference between each units activation becomes evident well before the sum has
converged and the effect of practice is much smaller.
The parameters for Figure 3 are the same as for Figure 2b, except that
k = 0.2444. The much greater leakage slows the accumulation of activation. At N = 0
(red lines) the difference is convergent (D > 0). Quenching does not occur as was
9
Activation
large the effect of practice, which increases from zero to k, is large in this example.
10
20
30
40
Figure 3. Solid lines indicate the unit with larger input, dashed lines the unit with the
smaller input, red lines indicate N = 0, blue N = 1, and green N = 50. Plots are
terminated at quenching, except for N = 0, where quenching does not occur in a finite
time.
RT (assuming I1 > I2) will be the value of t such that x1= , where the response
criterion > x0 (again assuming that is set such that RT<tq). This equation, and
Equation (14) do not have an explicit solution for t, except in some special cases2.
For example, when k = , S = -D, so x1 = and x2 = 0 have the form a = be(ct)+de(-ct), with solutions:
1 a a 2 4bd .
t = ln
c
2b
10
I1 (k ) I 2 > 12 (k ) 2 .
30
0
10
20
RT
40
50
60
10
20
30
40
I ( k )t
x (t ) = x0 +
e
(15)
RT =
( k ) + I
1
ln
k ( k )x0 + I
(16)
Substituting (3) into (16) we obtain the equation for the practice curve:
1 N I kx0 e N
RT ( N ) = e ln
N
k
I ke
(17)
The trade-off between the exponential and logarithmic terms makes this curve
difficult to study analytically. To simplify, we apply the linear Taylor approximation
ln(1+A) A, and make use of the fact that:
I kx0 e N
ke N ( x 0 )
1+
I ke N
I ke N
The fraction on the right hand side is small, given reasonable choices of parameters.
More importantly, it decays asymptotically to zero as N diverges to infinity. Thus, the
13
RT ( N ) =
x0
I k e N
(18)
To derive the RLR of this function, we need its asymptotic value as N diverges:
lim RT ( N ) = lim
( x 0 )
I ke N
( x 0 )
I k lim e N
x0
I
x0
N
I ke
d
= ( x 0 )
I k e N
dN
( x0 )ke N
(I ke )
N 2
(19)
Hence:
RLR ( N ) =
I
I k e N
(20)
The asymptotic value of this RLR function, and a value that the RLR is greater than at
all times, is given by:
I
I
=
=
N
N I ke
I k lim e N
So, the practise function specified by (18) has an RLR that is bounded, decreasing
from the upper bound (at N = 0) to lower bound (at N = ) with practice.
I
> RLR >
I k
I Kx0 1 I Kx
1 Nlim
0
= ln
=0
lim RT = ln
N
K lim I K K I K
N
(21)
Hence, in the limit, practice makes the decision process instantaneous. The derivative
of RT with respect to practice trials (N) is:
x0
dRT
dI
=
(I Kx0 )(I Kx ) dN
dN
(22)
RLR ( N ) =
x0
dI
I Kx0 I K (I Kx0 )(I Kx ) dN
ln
I
K
I
Kx
(23)
x0
I Kx0
= I 2 ln1 + K
I 2 ln
I K
I K
For large N the argument to the log function becomes approximately 1 and so the
logarithm can be accurately approximated by a linear Taylor function ln(1+A) A:
RLR(N ) = I 2 K
x0
I K
(24)
15
1 dI
I dN
Thus, the asymptotic RLR goes to zero unless dI dN increases at a faster-than-linear
rate, and so asymptotically I must increase at a faster than quadratic rate. For a oneunit version of Pages (2000) model, which assumes a slower than linear increase in I
with N, the asymptotic RLR will decrease to zero.
Two-Unit System
0.0
0.1
RLR
0.2
0.3
RLR decreases slightly for small N and approaches in the limit of large N.
10
20
30
40
Figure 5. RLR functions corresponding to Figure 2, and Figure 2 with = 0.1. Values
of are plotted as dashed lines.
16
1.0
0.0
0.5
RLR
1.5
2.0
10
20
30
40
Figure 6. RLR functions corresponding to Figure 3, and Figure 3 with = 0.1. Values
of are plotted as dashed lines.
Discussion
In this paper we described the SEEXC model of skill acquisition and
examined the properties of deterministic versions of the model. For one-unit systems
we showed analytically that learning on self-excitation could produce practice curves
that are close to the form seen in Heathcote et al.s (2000) survey of empirical practice
curves. We also showed analytically that learning on inputs, as suggested by Page
(2000), could only match the data when the increase in the input with practice is
17
18
19
20
RT ( N ) = a +
b
,
1 ce N
b=
x0
,
I
c=
k
I
Note that b > 0, and 0 < c < 1. As N the decision component approaches b (so
RT = a + b), whereas when N = 0, the decision component is b/(1-c). Under this
21
to N = . This form is likely to be more tractable for fitting as the location and scale
parameters are both linear:
(1 c )e N
RT (N ) = A + B
N
1 ce
A = a + b,
B=
bc
1 c
22
References
Brown, S. & Heathcote, A. (2001). From the O-U process to the diffusion model:
Including learning and automaticity in simple decision models. Paper presented
at the 10th Australasian Mathematical Psychology Conference, Newcastle
University, Newcastle, Australia.
Heathcote, A. (1998). Neuromorphic models of response time, Australian Journal of
Psychology, 50, 157-164.
Heathcote, A., Brown, S. (2000a). The law of practice and localist neural network
models. In Page, M., Connectionist modelling in psychology: A localist
Manifesto, Behavioral and Brain Sciences, 23, 479-480.
Heathcote, A., & Brown, S. (2000b) SEEXC: A self-exciting expert competitive
neural network model of speed skill. Paper presented at the 33rd Annual Meeting
of the Society for Mathematical Psychology, Kingston, Canada.
Heathcote, A. & Brown, S. (2002). A Sufficient Architecture for Neuromorphic
Response Time Models. Paper presented at the 6th Conference of the
Australasian Cognitive Science Society, Fremantle, WA, Australia.
Heathcote, A., Brown, S. & Mewhort, D.J.K. (2000) Repealing the power law: The
case for an exponential law of practice. Psychonomic Bulletin and Review, 7,
185-207.
Page, M. (2000). Connectionist modelling in psychology: A localist Manifesto,
Behavioral and Brain Sciences, 23, 443-512.
Ratcliff, R. (1978). A theory of memory retrieval, Psychological Review, 85, 59-108.
Usher, M. & McClelland, J. L. (1995). On the time course of perceptual choice: A
model based on principles of neural computation. Technical Report
PDP.CNS.95.5, Department of Psychology, Carnegie Mellon University.
Usher, M. & McClelland, J. L. (2001). The time course of perceptual choice: The
leaky, competing accumulator model, Psychological Review, 108, 550-592.
23
0.0
0.2
0.4
RLR
0.6
0.8
1.0
10
20
30
40
0.0
0.1
0.2
RLR
0.3
0.4
0.5
10
20
30
40
0.0
0.1
0.2
RLR
0.3
0.4
0.5
2.1
10
20
30
40
2.2
24
0.0
0.1
RLR
0.2
0.3
10
20
30
40
0.0
0.1
RLR
0.2
0.3
2.3
10
20
30
40
0.15
0.10
0.05
0.0
RLR
0.20
0.25
0.30
2.4
10
20
30
40
2.5
25
0.15
0.0
0.05
0.10
RLR
0.20
0.25
0.30
10
20
30
40
0.0
0.1
RLR
0.2
0.3
2.6
10
20
30
40
0.0
0.1
RLR
0.2
0.3
2.7
10
20
30
40
2.8
26
0.0
0.1
RLR
0.2
0.3
10
20
30
40
2.9
27