ANN Theory
ANN Theory
playhr
X
Sleephr LosS
funcion -9)
NewtoD Outpuut
Cwe run hdue
any humbe
Newnon
Hidden lay ens)
Dadaptt(Binany clasificaton)
Stu dy ply cleep Pans
h hr hr fa
3
5
3
HL 1 Fotw d
Propogaton
Bias
w 2
Exi i Act(9))| /1
Bic
BLck
=DwT 2
propotiON
uwerghs are th ual valuA th a re
attathtd to each i/P/fcatune and fhuy ton vey
thu mpo1 tance th t torespon cding featne
pre elichin he final op
wetghts Convey followtng Thinh9s
) Importen ce Jeatune. in predicHng pvailue.
werght 7 #eutune value N.
Sigmoid Acivation
funct on for binuy classificuton
Siqmoid=- , or 1
t e Xiwi+b)
<o O
O forwond nuopogcdio)
bias-
Tnput scign weights >
InputX we)ght +
Actua valul
output Acfivationni
fo
witb)
Bia
X
O
U2
2
lo-Cy-9) J
g-Ecox1+b
Z= C ) sigmoid fwd
Aitivaction
Propogation
Ue cen Solve function
hon- (ineuy problemns
uwinqthi&
we need to updahe uts win9 back propogution,
(oss (yy) W
x3
bwd propoyutionn
(-9)
Dod t LoS
when we are at A
we want to feuch W for
lobal inim4 -Ve slope
/
bnew=botd
2
O 2
d
e W4 old ) J oss
u ol cd.
X chan ule
id
uo neo 1d 9
old
SL -xO So chain uude
1old
O2we
ple
ON
031 loss(9-9)
uInew u oid
hew Iod 022
SLOs
IdSOgI SO So
Juule
SO O deni vaives.
LO O O d wol d
021 Oy
031 Loss (y-9)
new
old-9 JL SEMSE
TL Os1Oa O
Uotd O O So Iold
SLqmoid Acttvation
2
O4dsigmoid)e oar
denivcive 1 deivadtue
Sigmoid w e .
ConclHon.
S
pld
O 2x0 IS x O:0 y O'oS O'o2
deivative Vauu decreases
Veny Smal valul.
Veny tm ay
wnew old
not git
he wt uul
lowy
pdated or vey ven
p dae d
This icalled kinq gtadi ent proble m
vanir
Tan h
Re1
Leaky Rely
Prelu
O sgmeid Otn)
oS Ss,Od0tx)0'2
gradicnt vanirliney
o-0-2. Pro blem.
Sigmoid
Not 2eMo w t upduion i
enteed wmve
ey eus y ih
Adv (ase 2eno
Centud Cuves
Smooth qracien, plevens jumps i0
Pvaluly
2 ince ofp blo o and , Jf
normalisey olp each nenon.
3 lean preelichon veay c(ose fo o r
Disdelu uwe cunnsr creuAP
Prone to 9radient anishn q.edeep ntunas
nckwoR.
Funchfen op not 2eo
(enied., ot updation
) Sigmoid has ex SSuL
ponntRa oppenction, wch
S lower for (ompukUs fo
Caleudase
ten b funtio0 hypen bolic tangent funcion)
Hanh () : e'- e*|
eMte-
Ctan ho))=D
(O Oto4
fo2
tan h(7)-D to 1
S
Ad ) zno tentric output
Disad) stil (an CYecte Vauishing talienA
Pioblem
2 oS+
- 202 4 - 2
2
ReLolx) d ReLUa)
not diffenenkahle
a Ze4O
Re LU max(orx) d
not Zeno centic
*when Relu
du no vanishing Mgdic
neunon
2) fcunt caluletlon eacd.
3) when
Thi's a problemn.
input s tve, no
aolitn
Satuncution aob1en.
Disadu )not a zeno (enthic funtion Sin ce op is
ethm 0 oT
2) whun input is
-ve, ReLU inachve completely.
D ReLU d e d for inputk= O.
grddient = Zeno Same problem
Cigmoi d amd atan h.
41 Leaky ReLU function fcx)= max (o-te, «)
d (ft/dx
2 2
ale-1), othenwis)
dfcx)
S
o
o:2
-IO-55
T4olues problems feLU.
Adu: ) No lead Relu istu
2) Zeno (crted.
3) Bet/ tht Relo (proved
Di'sadu t)
tomputaion uavy Cduu fo
Cdut to
cxponunHa tulh
CXponentw tale hn)
8). SwTSH ( se/f garec) funcron
fen)= Mx Sigmeid Cx)
gaing in LsT
9) May o u
huppein e
PRel0, ELV
Reession
we linCu) a ctivution f
Also thene is
Sepaae
loss function.
-
C
any vaiutlon
ReLO
LosSfunctonu
DL (ANN)
Biny
Regression .cAaiin
MSE
clcnsfi ccutton
Ocro Binun
entopyCE
2) MAE
ctcgoT rc
CE
3)Hub joss
nuutie le
* LoSS VIS
cost funtlon
100 eCS in
fwd rop
LOSSA (9-9)*
uadruHC Eq" a d i e
Adu
Jeseent
3f i differentia ble
) + has any one lotalglobal nainimet:
3) 4 (onvenge feste
0iS ddv
not robuot to outi ens. we ul Squning fm error)
AeroF penctlsing
M e c t 0 tbsolue Error
Sub grudic
LF:19-9) cF 2 = y-9)
)Ro bt to outlies
disadv
not dif ad =O
2) Timt conjunuing lut to Sub adien
10SS =
yx log (9) -
TCZ) = e
e ZJ
, Prob
e
tolwey
onL
0'6
HL LoS
Concusion C C.E
ultic las> Ae LU, Joftmax.
8inny C6
Bin CM R e L U , sigmoid
weight
W
w2
w
wR
9
6
osbL
anch 2h i
(9-9)
we we optimi z ED 1D
batk propogfon to upcdate
ot So that we (an
loss /ey roY
Epocs
Disa clv
)dt oun4e extenAiVe technigul
more tam ugired)
Epocb
3Man thaing newnul netwo7k uith futl dataseh
for one ycle t Fod and bwd uopoguHon)
EX date
ooo 00 polrn
o9oo0-
Epoch.
1ogs
foation 1
updak
wt
2 record
1ss I t aton 2
updat
0isadu
tonuenge vey s lo uw
2) Ti me complexi fy hgh.
Adu:
low memory uguurcc.
3 Min - Batch S D
Bactch size= (o00 re rS.
1 nillton e coreld=D No
Epoth 1 tution Million
eooe O0Ttnation
coSth Ttatioo
ooofcend
9 Ttnaton MiM
batch
AdU
2)
Less Ronge intensi ve
3)
Betle
Betiu[impmoved
0ag T4uuto )
ConvnGence
AAA .Globa
m
T i complex1 y
SGD
Graclier -iu-batch
descent
SGrD
Noist
Globus
MNMn
noise =D sD mini-baheh sD
MO1P c s No ise
nois
To Temove noise
use mome rtum.
4) SGD uth momentum
aini btth momentum Ubed to eliminak
V
noise sing moving vgs.
EXponental weiyhted Avea982 Cxponential.
This is clso wsed in Iinu
Saies
Wneo Wod-
ne Dold
boid
VEh B Vtp-
+C1-)Cn
This
pp lied n dl
Noo
wt
whure
Vao Br Vau, +(158) xd L
Smoothening
Loss/costfh Adv
> Smoothenin g noSe
r e d ,
a n d q u u c ke n
noise
> e
C o n v e n
g e n c e .
Reducinq th noise
. Quicken convengance
weqhtr
AdagucLd=D=D Adaptive Gradient oleor ent
wt wL--2 L fixcd in a abouc
dwE- optini2e).
Rcplace uith 2
- L
Reqwire nent:
Tniti dl4 lagR
Hate
eotning
uull make quitko6 Small no. to
Conven gence iutially, but avold 2en0
o
wapproa h gtobul ivisron
Cro
numo Jeoarng Aate
rlul 8hoUld udule
do uith eveny
fhet e do not nss H ie
global minima. epoch t incyeccocs.
CUs h T iDCreaE will n ereccse
wt
ll decreanc c we uull u a ch
hea to gtoha minima.
be - -9 Vop Sdw+e