38 Backpropagation
38 Backpropagation
Backpropagation
DR. ABHISHEK SARKAR
ASSISTANT PROFESSOR
MECHANICAL ENGG., BITS
Backpropagation
• Based on gradient-descent algorithm
8
Back-Propagation Network
10
1 1 w01
v01 w0k
v0p w0m
v0j
t1
x1 v11 Y1 y1
x1
v1j w11
v1p
Z1 w1k
vi1 w1m
xi t1
x2i vij Y
x2k yk
wj1
vip
Zj wjk
wjm
vn1
xn t1
xn vnj
wp1
Ym ym
vnp
wpk
Zp wpm
12
Architecture
1
• The activation function could be any function v01
v0p
which increases monotonically and is also x1 v11 v0j
differentiable. x1
n v1j
v1p
Z1
• The net input to zj is zinj =v 0 j + ∑ x i v ij
i=1 xi vi1
and the output is z j =f ( z inj ) x2i vij
• Similarly, p
vip
Zj
the net input to yk is y ink =w 0 k + ∑ z j w jk vn1
i=1
xn xn vnj
and the output is y k =f ( y ink ) vnp
13 Zp
Architecture
14
Training Algorithm
15
Feed-forward phase (Phase I)
• Step 4: Each hidden unit zj(j = 1 to p) sums its weighted input signals
n
to calculate net input:
zinj =v 0 j + ∑ x i v ij
i=1
16
Feed-forward phase (Phase I)
• Step 5: For each output unit yk (k = 1 to m), calculate the net input:
p
y ink =w 0 k + ∑ z j w jk
j=1
17
Back-propagation of error (Phase ll)
• Step 7: Each hidden unit (zj, j = 1 to p) sums its delta inputs from the
output units: m
δ inj = ∑ δ k w jk
k=1
• The term δinj gets multiplied with thee derivative of f(zinj) to calculate
the error term: δ j =δ inj f ' ( zinj )
[The derivative f'(zinj) can be calculated as before.]
• On the basis of the calculated δj, update the change in weights and
bias: Δ v ij =α δ j x i ; Δ v 0 j =α δ j
19
Weight and bias updation (Phase Ill):
1 v01= 0.3 1
w01= -0.2
24
Example