DL Notes B Div
DL Notes B Div
b Nitesh Khapra
Modue-4 0
NPTEL
The imput to Hhe Nettoork is an y-dime
VectoT
h the netwoTk Contains L- hiclden layers
havinq mneuTOMS each
ftmalty there is one oun layer
Contal ning k neurons
h each neuron im the hidden ayer
OutputJayer can be split Iito tuD
a parts PTe-acivastion k acivation
he impuk aye1 can be called O
Layey
hL
The pre-athvction at Layer IS
The activatiom at the ven by ailx) bi + wihi-i(x)]
=
Outputs
Beal Valuu Probabsilihieg
Outpus Acivaion Linear Softmax -F)
ossumchion Squared Cvoss Entropy
WL
hay
od(o) Df()ýaLi
Wel
FOY a Simgft tyainimg exomple, Backpropagatiom: - algortm
0 35
b Layer
Jarqef Value
Hfnal 1Le
te-
UT 0.01 H = 0.59326
0 o 11 X, XWyt b
H , xw3 t
0.0S XO 25-+ 0-1oX0 30t0 3S
H2 0.3925
S9moid - 5968
H fina 14eH
=HXWs t fh64 b
Y Y 0.5932 xO 400.5q688 AO.45+0-60
as T= O 0J
Y 110590
Sigmoid fun Tavgeti cdo not
Ya 0.727
Total Emon
2(tcmet -0utpti)
ETotul
Ct-8fiad t-fimal) 2
(.1-0.751 (0.99 O 42
- 0.41 0- 2173 Sme
O.294S + O.02360
weg ate
0.298109
Backoard pal8 atthe output Layey Evo7 E t ot a l - a
final
(T4 =e(yf
1
Ifined
e-1-Yfne
final
inal x( finad finalx - fina - (8
final Os13 x (1- O751 36)
oina =
0 1868
Unit- 2
Sie neurom lwith 3 mputs -
n a c h vatian
a-7 umitim
ofp
layer a =ga:
lossuncion
Classificahioh
IRegyessiOn
Cross e n t r o p s
msE Simar CYosS eHtroPy
MAE Categonca
Lossunction A erroY f u n i s f o r Single
loss unciom
training exomple /input
overthe entire
Coss -fun:. A CosEun is he average loss
trainimg dotaset
Kegresio
H S f (mean Squortd eo)
MAE (mear Aosolule Eror
Clascation
msethe
Binar
Calegoical
diference
erdolo
CrosS- entropy
nS entrop
CrossEntror
betueen he atual value model
detase
pTediction, squqr t average it across the whole
A
MSE N
(t-
MAEThe diftrencebwthe actual value k model prediction
average it aCTOSS the whole olatset
MAE 1yi -
Ni
Hupekparameker S tumima
HuperqTamelers Crethe variable which delerrmincg
the
4the hehworkstucturt Ceq no ohidden uni)
Varnables urhic% deter mins howKe mehwork 13
troumed Ceadeaming ae).
Huper paraneersareset eore tvainina befor
opttmizinq the weight b i a
GTadient dejecent it tpes -
Snitialise wb
LDBadk qao
olren dien
Cast Cos
w w
A
Rms p (
Thoment eshimation)
(Adaptive
SGD
Adom
Aeavnin g a
dagrad - modies he qeneya à Ssumalated
Parametrs unm
highey pqramekrs unimg
deermine
vriables uhich
fetweY ayers -
anis hing
Exploclin q
r CdiertEE)
kadiai
ealent
adent
/1
4 L l3 LLa L3 0P
Layer
Slope
The derivative o slope The devivotive olope
will
ill get SmaUey qe lavgey lorgRY
SmalleY
Due tu weiqht, not oz
Depenepende On wejghts the atation un ho
n0 Om activaton
fumdion
exist u i natkpropagation
Remedics foY Vanishing Yadiey7t heblem
fov 4ve i
RELUfumchon f) Devivartive dRELU
f) = *i,when «yo
O, when x{ )
Model. complie
Model add
mode
Reqyession (L2 Regulavizahio
Ridge
+ Regularized term
Ridge Reqvession= LOssunchon
RsSridpe (w,b) = Z ( i - (o,xi tb} *E u
Vaidaton set
Aalidatiom
set
Tyaning s e
aThy Numbey
iterations
. Spliltine of datq set e
foldYoss ylidatioy) e e e
yalidating You approac uin9 fold Vaildafieon
TData, l i t mo pvitio
Vadi,dalvnStn|
d olidahol yinino (oinna
old VvalialatioValidation"Tvaining
faAs1 Yekuke [Training Validakian
Caty Dogs dalaset Image classiication Ueing NN
fo Tmoyemen kevval
padding asdimg 2exo n boun davy
3 cale
paddi mg Volid
Poelina Layers
Toredute ho ondihion ality
To letect eda a,eyes, n8es
Mlax pooling ma o t
Averaaa ooung