Escolar Documentos
Profissional Documentos
Cultura Documentos
NETWORKS
OUTLINE
Why use recurrent networks
Typical architectures
A taste of backprop
Delayed response tasks are a standard way of investigating short-term memory (STM). The
agent is typically assumed to remember in some way the necessary information about the
stimulus (for example the side on which a stimulus appeared) during the delay period.
Studies upto now are far away from suggesting a specific model of how animals solve delayed
response tasks (Ziemke). The solution seems to lie in an interdisciplinary approach.
Figure The two situations in the simple T-maze environment. Adapted from Ulbricht (1996).
I4
I2
L1 I1
I5
I6
L2
Repeated T-maze
Multiple T-maze
QuickTime og en
TIFF (LZW)-dekomprimerer
kreves for se dette bildet.
Figure 5.6
12
13
14
15
********************
The hidden units represent an internal reduced
representation of the data in the sequence
that precedes the current input.
16
18
19
ESCN
More specifically, given an input vector ij(t), j = 1n, state unit vector s(t) and output vector
o(t) are usually calculated as follows by the function network:
Where f is the logistic activation function, and W0(t) and WS(t), together referred to as function network weights, are
two-dimensional connection weight matrices dynamically computed in every time step by the context network.
W 0(t) = {wa(t), wb(t), .wn(t) }
// ONE FOR EACH OUTPUT UNIT
W s(t) = {wsa(t), wsb(t), .wsn(t) } // ONE FOR EACH STATE UNIT
20
ESCN
Theres an additional decision unit in ESCN.
Given the state vector sk(t), k = 1m, the decision unit activation d(t) and the function network weight matrices W0(t) and WS(t), are
dynamically computed in every time step t as follows by the ESCNs context network:
Where f is the logistic activation function, Vd is a one-dimensional connection weight matrix, mapping current state to decision unit
activation, and V0 and VS are two-dimensional connection weight matrices mapping the current internal state s(t) to the next
time steps function network weights, if the decision unit is active, i.e. has an activation value of at least 0.5.
21
QuickTime og en
TIFF (LZW)-dekomprimerer
kreves for se dette bildet.
22
Figure 5.4
24
Figure 5.5
25
A recurrent backpropogation
network
A backpropogation network need not be strictly
feedforward and can have recurrent
connections.
A unit can feed to itself, to units in the same or
lower levels.
A recurrent connection feeds back activation that
will affect the output from the network during
subsequent iterations.
For every recurrent network there is a
feedforward network with identical behavior.
26
27
28
29
Rumelhart network
The network has:
5 input units, one for each letter.
3 output units, one for each number.
30 hidden units.
Each hidden unit is connected to itself and
every other hidden unit.
31
Training Steps
T1: Input unit corresponding to the first letter is 1 and other
inputs are off.
T2: Input unit corresponding to the second letter is 1 and
other inputs are off.
T3: The target output is the third character in the sequence.
T4: The target output is the fourth character in the
sequence.
T5: The target output is the fifth character in the sequence.
T6: The target output is the sixth character in the sequence.
32
34
An example:
Counting to three.
35
QuickTime and a
decompressor
are needed to see this picture.
36