What Actions Can Human Brain Do?: Trained
What Actions Can Human Brain Do?: Trained
Human brain
• The brain is a highly complex, non-linear, parallel
information processing system.
• It performs tasks like pattern recognition, perception,
motor control, many times faster than the fastest
digital computers.
• It characterizes by
– Robust and fault tolerant
– Flexible – can adjust to new environment by learning
– Can deal with fuzzy, probabilistic, noisy or inconsistent
information
– highly parallel
– small, compact and requires little power.
2
Human Brain VS Von Neuman Computer
3
Biological Neuron structure
• Brain consists of approximately 1011 elements
called neurons.
• Components of Biological Neurons:
► Dendrite: Receives signals from other neurons
► Soma: Processes the information
► Axon: Transmits the output of this neuron
► Synapse: Point of connection to other neurons
4
How Biological neurons work
• Electrical signal is generated by the neuron, passes down
the axon, and is received by the synapses that join onto
other neurons dendrites.
5
• Neurons are arranged in a hierarchical fashion – Layers ,
each layer has its own role and responsibility
6
Learning in networks of neurons
• Knowledge is represented in neural networks by the
strength of the synaptic connections between neurons
(hence “connectionism”)
7
Artificial Neural network (ANN)
► An Artificial neural network is an information-processing
system that has certain performance characteristics in
common with biological neural networks.
BNN VS ANN
8
How ANN Works ?
1) Information processing occurs at many simple elements called neurons.
9
Learning in ANN
Adding Bias
10
ANN with Bias
11
Example-1:
Calculate net input
net input = 3
12
Example-2
net input = 3
13
Feature Extraction
Characteristics of ANN
1. Architecture (Structure): the pattern of nodes and connections
between them
2. Training, or learning, algorithm: method of determining the weights
on the connections
3. Activation function: A function that produces an output based on
the input values received by node
14
Architecture
1- Feed Forward NN
► fully connected
B- Multi-Layer network :
► A net with one or more layers of neurons
► Hidden neurons between input units and output layers.
► Solve more complicated problems than can single layer nets.
15
2- Feed-back (Recurrent) NN :
3- Associative Network:
16
Training – Learning algorithm.
17
1- Supervised Learning
• Presented with inputs together with the target (teacher signal).
• NN produce output close to target by adjusting the weights.
• “error correction method”,
– Least Mean Square (LMS)
– Back Propagation
2- Un-Supervised Learning
Define: Define:
Each of the training patterns Each of the training patterns
associated with target output not associated with target
vector output vector
Data: Data:
(Input, desired output) (Different input)
Problems: Problems:
Classification , regression Clustering , data reduction
Pattern recognition
NN models:
NN models: Self-organizing maps (SOM)
18
perceptron Hopfield
Heb
Activation function
19
3- Bipolar step function – Symmetric hard Limit
20
5- hyperbolic tangent (‘‘tanh’’)
• Squashes the neuron’s pre-activation between -1 and 1
• Can be positive or negative
• Bounded
• Strictly increasing
21
7- Radial basis activation function
22
Artificial Neural Network Development Process
23
Linearly Separable Function
§ The function is capable of assigning all inputs to two categories.
§ Decision boundary : line that partitions the plane into two decision
regions
b + åi =1 x i w i = 0
n
§ Decision boundary has equation
b + x1w1 + x2 w 2 ³ 0
o Negative region : decision boundary for output -1 with
equation
b + x1w1 + x2 w 2 < 0
§ Example on linear separable problems
24
§ If two classes of patterns can be separated by a decision boundary
then they are said to be linearly separable
§ If such a decision boundary does not exist, then the two classes
are said to be linearly inseparable (Non-Linear Separable)
25
Capacity of single neuron
• Could do binary classification (two outputs):
• also known as logistic regression classifier
o if greater than 0.5,
predict class 1
o otherwise, predict
class 0
26
u Can’t solve non-linearly separable problems
McCulloch-Pitts network
• Each neuron has a fixed threshold value T such that if the net
input to the neuron is greater than the threshold, the neuron fires.
•
27
Architecture
• McCulloch-Pitts neuron receive signals from any number of neurons.
Algorithm
• The weights for a McCulloch-Pitts neuron are set, together with the
threshold for the neuron's activation function,
28
Example -1 : And Function
Example -2 : OR Function
29
Example -3 : And Not (y = X1’ X2 )
30
Example -5 : XOR Function
31
ý Applications of neural Network
☺ Financial modelling – predicting stocks, currency exchange rates
☺ Other time series prediction – climate, weather,
☺ Computer games – intelligent agents, backgammon
☺ Control systems – autonomous adaptable robotics
☺ Pattern recognition – speech recognition, hand-writing recognition,
☺ Data analysis – data compression, data mining
☺ Noise reduction – ECG noise reduction
☺ Bioinformatics – DNA sequencing
ý Advantage of Neural Network
o ANN are powerful computation system: consists of many neurons
o Generalization:
§ can learn from training data and generalize to new one.
§ using responses to prior input patterns to determine the
response to a novel input
o fault tolerance:
§ able to recognizes a pattern that has some noises
§ Still works when part of the net fails
o Massive parallel processing:
§ process more than one pattern at same time using same
set of weights
o distributed memory representation
o Adaptability:
§ increase network ability of recognition by more training
o low energy consumption
o useful for brain modeling
o used pattern recognition
o Able to learn any complex non-linear mapping
o Learning instead of programming
o Robust: Can deal with incomplete and/or noisy data
32
ý Disadvantage of Neural Network
o Need training to operate
o High processing time for training
o require specialized HW and SW
o lack of understanding the behavior of NN
o convergence not guaranteed ( to reach solution is not
guaranteed )
o no mathematical proof for the learning process
o Difficult to design
o They are no clear design rules for arbitrary applications
o Learning process can be very time consuming
o Can over-fit the training data, becoming useless for
generalization
33
ý Who is concerned with NNs?
34
Kohonen Self Organizing Map (SOM)
ý Training algorithm used for clustering
o Euclidean distance
d (i, j) = (| x - x |2 + | x - x |2 +...+ | x - x |2 )
i1 j1 i2 j 2 ip jp
9
ý Number of neurons in output layer equal to number of clusters
R=1
R=2
o Rectangular
R=0
R=1
R=2
o Hexagonal
R=0
R=1
R=2
10
ý SOM algorithm in high level
o Initialize weight
parameters.
D(J) = ∑ (wij – xi )2
the weights:
ý Stopping condition
11
ý Example
Place (1 1 0 0), (0 0 0 1), (1 0 0 0), (0 0 1 1) into 2 clusters with no
topological structure (assumes no neighborhood) R = 0
Initial α
α (0) = 0.6
decrease α
α (t+1) = 0.5 * a(t)
Solution
W1 of cluster 1 W2 of cluster 2
.2 .8
.6 .4
.5 .7
.9 .3
Compute distance
D(1) = (W1 - P1) = (.2 – 1)2 + (.6 – 1)2 + (.5 – 0)2 + (.9 – 0)2 = 1.86
D(2) = (W2 - P1) = (.8 – 1)2 + (.4 – 1)2 + (.7 – 0)2 + (.3 – 0)2 = .98
Find winner
12
New weight
13
Third Pattern P3= (1 0 0 0)
14
ÞApplication Examples of Kohonen Network
o Data compression
15