Lesson 01
Lesson 01
1 Introduction
1.1 What is Neural Net?
An Artificial Neural Network (ANN) or Neural Network have been developed as generalizations
of mathematical models of human cognition or neural biology, based on following assumptions:
• A neural net consists of a large number of simple processing elements called neurons, units,
cells, or nodes.
• Each neuron is connected to other neurons by means of directed communication links, each
with an associated weight.
• Each neuron has an internal state, called its activation or activity level, which is a function of
the inputs it has received.
• The activation 𝑦 of neuron 𝑌 is given by some function of its net input, 𝑦 = 𝑓(𝑦$% ).
• The several common activation functions are given in the following table.
Table 1.1
• Now suppose further that neuron 𝑌 is connected to neuron 𝑍! and 𝑍" , with weights 𝑣! and
𝑣" respectively as shown in Figure below.
• Neuron 𝑌 sends its signal 𝑦 to each of these units. In general, the values received by neurons
𝑍! and 𝑍" will be different.
• The arrangement of the units (the architecture of the net) and the method of training the net
(find optimal value for the weight) are discussed further in the next section.
A Single-Layer Net
A Multilayer Net
• A multilayer is a net with one or more layers of nodes between the input units and the output
units.
• Multilayer networks are more powerful than single-layer networks. For instance, a two-layer
network having a sigmoid first layer and a linear second layer can be trained to approximate
most functions arbitrarily well.
Supervised Training
Unsupervised Training
Table 1.2
If the connection weights for a neural net are stored in a matrix 𝑾 where
𝑤!! ⋯ 𝑤!"
𝑾=# ⋮ ⋱ ⋮ (,
𝑤#! ⋯ 𝑤#"
(1.2)
the net input to 𝑌& (with no bias on unit 𝑗) is simply the dot product of the vectors 𝒙 =
(𝑥! , 𝑥" , … … … , 𝑥$ , … … … , 𝑥% ) and 𝒘.𝒋 (the jth column of the weight matrix):
𝑦$%! = 𝒙 ∙ 𝒘.𝒋
= ∑%$)! 𝑥$ 𝑤$&
(1.3)
Bias
Examples 1.1
(a) Given a two-input neuron with the following parameters: 𝑏 = 1.2, 𝑾 = [3 2] and 𝒙 =
[−5 6]+ ,calculate the neuron output for the following transfer functions:
(i) A symmetrical hard limit transfer function
(ii) A saturating linear transfer function
(iii) A hyperbolic tangent sigmoid (tansig) transfer function
(b) Given a two-input neuron with the following weight matrix and input vector: 𝑾 =
[3 2] and 𝒙 = [−5 7]+ , we would like to have an output of 0.5. Do you suppose that
there is a combination of bias and transfer function that might allow this?
(i) Is there a transfer function from Table 1.1 that will do the job if the bias is zero?
(ii) Is there a bias that will do the job if the linear transfer function is used? If yes, what
is it?
(iii) Is there a bias that will do the job if a log-sigmoid transfer function is used? Again,
if yes, what is it?
(iv) Is there a bias that will do the job if a symmetrical hard limit transfer function is
used? Again, if yes, what is it?