08 Ann PDF
08 Ann PDF
08 Ann PDF
Synapse Dendrites
Axon
Axon
Soma Soma
Dendrites
Synapse
Middle Layer
Input Layer Output Layer
x1
Y
w1
x2
w2
Neuron Y Y
wn
Y
xn
Y Y Y Y
+1 +1 +1 +1
0 X 0 X 0 X 0 X
-1 -1 -1 -1
Inputs
x1 Linear Hard
w1 Combiner Limiter
Output
∑ Y
w2
θ
x2
Threshold
Class A 1
1
2
1
x1
Class A 2 x1
x1w 1 + x2w 2 − θ = 0 x1 w 1 + x 2 w 2 + x 3 w 3 − θ = 0
x3
(a) Two-input perceptron. (b) Three-input perceptron.
e( p) = Yd ( p) −Y ( p) where p = 1, 2, 3, . . .
1 1 1
x1 x1 x1
0 1 0 1 0 1
First Second
Input hidden hidden Output
layer layer layer layer
i wij j wjk
xi k yk
m
n l yl
xn
Input Hidden Output
layer layer layer
Error signals
θ3
w13 −1
x1 1 3 w35
w23 θ5
5 y5
w24
x2 2 4 w45
w24
Input θ4 Output
layer layer
−1
Hidden layer
Negnevitsky, Pearson Education, 2011 37
I The effect of the threshold applied to a neuron in the
hidden or output layer is represented by its weight, θ,
connected to a fixed input equal to −1.
I The initial weights and threshold levels are set
randomly as follows:
w13 = 0.5, w14 = 0.9, w23 = 0.4, w24 = 1.0, w35 = −1.2,
w45 = 1.1, θ3 = 0.8, θ4 = −0.1 and θ5 = 0.3.
θ 5 = θ 5 + ∆ θ 5 = 0 .3 + 0 .0127 = 0 .3127
0
10
Sum-Squared Error
-1
10
-2
10
-3
10
-4
10
0 50 100 150 200
Epoch
+1.5
−1
+1.0
x1 1 3 −2.0 +0.5
+1.0
5 y5
+1.0
x2 2 +1.0
4
+1.0
+0.5
−1
Negnevitsky, Pearson Education, 2011 45
Decision boundaries
x2 x2 x2
x 1 + x 2 – 1.5 = 0 x 1 + x 2 – 0.5 = 0
1 1 1
x1 x1 x1
0 1 0 1 0 1
tan h 2a
Y = −bX
−a
1+ e
where a and b are constants.
Suitable values for a and b are:
a = 1.716 and b = 0.667
Negnevitsky, Pearson Education, 2011 47
I We also can accelerate training by including a
momentum term in the delta rule:
∆w jk ( p ) = β ⋅ ∆w jk ( p − 1) + α ⋅ y j ( p ) ⋅ δ k ( p )
1
Learning Rate
0.5
-0.5
-1
0 20 40 60 80 100 120 140
Epoch
0.8
Learning Rate
0.6
0.4
0.2
0
0 20 40 60 80 100 120
Epoch
2
Learning Rate
1.5
0.5
0
0 10 20 30 40 50 60 70 80 90
Epoch
Output Signals
Input Signals
x2 2 y2
xi i yi
xn n yn
+1, if X > 0
sign
Y = −1, if X < 0
Y , if X = 0
y1
y
Y = 2
M
y n
m=1
(−1, 1, 1) (1, 1, 1)
y1
0
1 −1
x1(p) 1 x1(p+1) 1
1 y1(p) 1 y1(p)
x2(p) 2 x2(p+1) 2
2 y2(p) 2 y2(p)
xi(p)
j yj(p) j yj(p)
i xi(p+1) i
m ym(p) m ym(p)
xn(p) n xn(p+1) n
Input Output Input Output
layer layer layer layer