Final PPT DataMining
Final PPT DataMining
Mihir Mohite
Jeet Kulkarni
Rituparna Bhise
Shrinand Javadekar
http://www.csse.uwa.edu.au/teaching/units/233.407/lecture
Notes/Lect4-UWA.pdf
http://www.csse.uwa.edu.au/teaching/units/233.407/lecture
Notes/Lect4-UWA.pdf
http://www.comp.glam.ac.uk/digimaging/neural.htm
http://www.nbb.cornell.edu/neurobio/linster/lecture4.pdf
src:http://www.nbb.cornell.edu/neurobio/linster/lecture4.pdf
x1-x2= 1
x1
Bias as an input
X0= w0
o/p
+1
class
w1
x1 ∑ f
wn
Activation func
xn Summing func
A Multilayer Feed-Forward
Neural Network
Output Class
Ok
Output nodes
w jk
Oj
Hidden nodes
wij - weights
Input nodes
Network is fully connected
Input Record : xi
Inputs to a Neural Network
INPUT: records without class attribute with normalized attributes
values.
HIDDEN LAYER – the number of nodes in the hidden layer and the
number of hidden layers depends on implementation.
Net Weighted Input
• Given a unit j in a hidden or output layer, the
net input is
I j wij Oi j
i
Oj = 1 if lj>T
Oj= 0 if lj<=T
data
Using Error Correction
Basic Elements
A set of neurons that are all same except synaptic
weight distribution
respond differently to a given set of input pattern
A mechanism to compete to respond to a given input
The winner that wins the competition is called“winner-
takes-all”
Competitive Learning
Src: http://www.csse.uwa.edu.au/teaching/units/233.407/lectureNotes/Lect4-UWA
Single Discrete Perceptron Training
Algorithm (SDPTA)
Src: http://www.csse.uwa.edu.au/teaching/units/233.407/lectureNotes/Lect4-UWA
Given are P training pairs
{X1,d1,X2,d2....Xp,dp}, where
Xi is (n*1)
di is (1*1)
i=1,2,...P
Yi= Augmented input pattern( obtained by appending 1 to the input
vector)
i=1,2,…P
In the following, k denotes the training step and p denotes the step
counter within the training cycle
Step 1: c>0 , Emin is chosen,
Step 2: Weights are initialized at w at small values, w is (n+1)*1.
Counters and error are initialized.
k=1,p=1,E=0
Step 3: The training cycle begins here. Input is presented and
output computed:
Y=Yp, d=dp
O=f(net) net=wtY.
SCPTA contd..
Src: http://www.csse.uwa.edu.au/teaching/units/233.407/lectureNotes/Lect4-UWA
Algorithm
Given are P training pairs
{X1,d1,X2,d2....Xp,dp}, where
Xi is (n*1)
di is (n*1)
No of Categories=R.
i=1,2,...P
Yi= Augmented input pattern( obtained by appending 1 to the input
vector)
i=1,2,…P
In the following, k denotes the training step and p denotes the step
counter within the training cycle
Step 1: c>0 , Emin is chosen,
Step 2: Weights are initialized at w at small values, w is (n+1)*1.
Counters and error are initialized.
k=1,p=1,E=0
Step 3: The training cycle begins here. Input is presented and
output computed:
Y=Yp, d=dp
Oi=f(wtY) for i=1,2,….R
RDPTA contd..
Type:
Feedforward
Neuron layers:
1 input layer
1 or more hidden layers
1 output layer
Learning Method:
Supervised
Notation:
x = input training vector
t = Output target vector.
δk = portion of error correction weight for wjk that is due
to an error at output unit Yk; also the information about
the error at unit Yk that is propagated back to the hidden
units that feed into unit Yk
δj = portion of error correction weight for vjk that is due to
the backpropagation of error information from the output
layer to the hidden unit Zj
α = learning rate.
voj = bias on hidden unit j
wok = bias on output unit k
EBPTA contd..
Generalisation
Once trained, weights are held constant, and
input patterns are applied in feedforward.
mode. - Commonly called “recall mode”.
We wish network to “generalize”, i.e. to make
sensible choices about input vectors which are
not in the training set.
Commonly we check generalization of a
3) http://www.codeproject.com/dotnet/simple_ocr.asp
Steps for Classification
Binarisation
Preprocessing
Segmentation using
heuristic algorithm
Digitized input
representation
** source http://www.codeproject.com/dotnet/simple_ocr.asp
Preprocessing
Neural Network
Present extracted
segmentation points to ANN n - inputs 1 - output
Network Pruning
Rule Extraction
Network construction and training
Construct and train a neural network
Network Pruning
Aims at removing redundant links and units without
increasing the classification error rate
Small number of units and links are left in the network
Rule Extraction
Extracts classification rules from the pruned network
(a1 θ v1) ^ (a2 θ v2) ^ … (an θ vn) then Cj
Rule Extraction Algorithm**
Input nodes, Hidden nodes, Output node
Activation values
**http://en.wikipedia.org/wiki/Image:Neuralnetwork.png
1. Enumerate hidden node activation values
E.g.
H = {0,0,1,1,0}
E.g.
(H1 = 0) ^ (H2 = 0) ^ (H3 = 1) ^ (H4 = 1) ^ (H5 = 0) then O
3. For each hidden unit, enumerate the input
values that lead to them
E.g.
For H1, I = {0,0}
For H2, I = {0,1}
For H3, I = {1,0}
For H4, I = {1,1}
For H5, I = {-1,-1}
4. Generate rules that describe the hidden unit
activation value in terms of inputs
E.g.
(I1 = 0) ^ (I2 = 0) then H1
(I1 = 0) ^ (I2 = 1) then H2
(I1 = 1) ^ (I2 = 0) then H3
(I1 = 1) ^ (I2 = 1) then H4
(I1 =-1) ^ (I2 =-1) then H5
Incremental training