0% found this document useful (0 votes)
40 views

Artificial Intelligence, An Introductory Course

The document provides an overview of artificial neural networks and learning strategies. It discusses supervised and unsupervised learning. Supervised learning uses external feedback to correct the network by reducing error, while unsupervised learning discovers patterns in unlabeled data through self-organization. The general algorithm for network learning iteratively updates weights according to a learning rule to minimize error between actual and desired outputs. Delta learning is a common supervised rule that adjusts weights proportionally to the error and activation function derivative.

Uploaded by

praveenvarma26
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
40 views

Artificial Intelligence, An Introductory Course

The document provides an overview of artificial neural networks and learning strategies. It discusses supervised and unsupervised learning. Supervised learning uses external feedback to correct the network by reducing error, while unsupervised learning discovers patterns in unlabeled data through self-organization. The general algorithm for network learning iteratively updates weights according to a learning rule to minimize error between actual and desired outputs. Delta learning is a common supervised rule that adjusts weights proportionally to the error and activation function derivative.

Uploaded by

praveenvarma26
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Artificial Intelligence,

An Introductory Course

The International Congress for global Science and Technology


www.icgst.com

Instructor: Ashraf Aboshosha, Dr. rer. nat.


Engineering Dept., Atomic Energy Authority,
8th Section, Nasr City, Cairo, P.O. Box. 29
E-mail: aboshosha@icgst.com,
Tel.: 012-1804952

Lecture (3): Artificial Neural Networks

Course Syllabus:

• An introduction to artificial intelligence and


machine learning
• Artificial Neural Networks
• Intelligent search techniques
• Neural programming

This is free educational material


6. Network learning strategies
Under the notation of learning in a network, we will consider a process
of forcing a network to yield a particular response to a specific input. A
particular response may or may not be specified to provide external
correction. learning is necessary when the information about inputs/ outputs is
unknown or incomplete a priori, so that no design of a network can be
performed in advance. The majority of the networks requires training in a
supervised or unsupervised learning mode. Some of the networks, however,
can be designed without incremental training. They are designed by batch
learning rather than stepwise training.

Batch learning takes place when the network weights are adjusted in
a single training step. In this mode of learning, the complete set of input/output
training data is needed to determine the weights, and feedback information
produced by the network itself is not involved in developing the network. This
learning technique is also called recording. Learning with feedback either
from the teacher or from the environment rather than a teacher, however, is
more typical for neural networks. Such learning is called incremental and is
usually performed in steps. The concept of feedback plays a central role in
learning. The concept is highly elusive and somewhat paradoxical. In a broad
sense it can be understood as an introduction of a pattern of relationships into
the cause-and-effect path. We will distinguish two different types of learning.

1-)supervised learning 2-)unsupervised learning

1
F(W,X)

X O
Network Input Network output
W

Network d
Learning
Algorithm

Learning
Signal

fig. (9) Supervised Learning Rule

In supervised learning fig. (9) we assume that at each instant of time


when the inputs applied, the desired response d of the system is provided by
the teacher, the distance between the actual and desired response serves as
error measure and is used to correct network parameters externally. Sense
we assume adjustable weights, the teacher may implement a reward-and-
punishment scheme to adapt the network's weight matrix W. For instance, in
learning classifications of input patterns or situations with known response,
the error can be used to modify weights so that the error decreases. This
mode of learning is very pervasive. Also, it is used in many situations of
natural learning. A set of input and output patterns called a training set is
required for this learning mode. Typically, supervised learning rewards
accurate classifications or associations and punishes those which yield
inaccurate response. The teacher estimates the negative error gradient
direction and reduces the error accordingly. In many situations, the inputs,
outputs and the computed gradient are deterministic, however, the
minimization of error proceeds over all its random realizations. As a result,
most supervised learning algorithms reduce to stochastic minimization of error
in multi-dimensional weight space. In learning without supervision, the
desired response is not known , thus , explicit error information cannot be
used to improve network behavior. Since no information is available as to
correctness or incorrectness of responses, learning must somehow be
accomplished based on observations of responses to inputs that we have
marginal or no knowledge about. for example unsupervised learning can
easily result in finding the boundary between classes of input patterns.
2
F(W,X)

X O
Network Input Network output
W

Fig. (10) Unsupervised Learning Rule

Unsupervised learning fig. (10) algorithms use patterns that are


typically redundant raw data having no labels regarding their class
membership, or association. in this mode of learning the network must
discover for itself any possibly existing patterns, regularities, separating
properties, etc. while discovering these, the network undergoes change of its
parameters, which is called self-organization. the technique of unsupervised
learning is often used to perform clustering as the unsupervised classification
of objects without providing information about the actual classes. This kind of
learning corresponding to minimal a priori information available. Some
information about the number of clusters, or similarity versus dissimilarity of
patterns, can be helpful for this mode of learning. Unsupervised learning is
sometimes called learning without teacher. this terminology is not the most
appropriate, because learning without teacher is not possible at all. although,
the teacher does not have to be involved in every training step, he has to set
goals even in an unsupervised learning mode. We may think of the following
analogy. learning with supervision corresponds to classroom learning with the
teacher's questions answered by students and corrected, if needed by the
teacher. Learning without supervision corresponds to learning the subject
from a videotape lecture covering the material but not including any other
teacher's involvement. Therefor, the student cannot get explanations of
unclear questions, check answers and become fully informed.

6.1. The general algorithm of learning


3
Neural networks have a different types and every type has its own
learning rule. All the methods of learning have the same general algorithm,
this algorithm mainly change the network parameters according to its learning
rule to accommodate the network’s characteristics to its desired pattern. In
general for the neuron I and its input j the weight vector
wi = [ wi1 wi 2 ............ win ]t increases in proportion to the product of input x and

learning signal r. The learning signal r is in general a function of Wi , X , and

sometimes of the teacher’s signal di . We thus have for the network shown in

fig. (11):

r = f ( wi , x , d i ) ..........................................(16)

The increment of the weight vector Wi produced by the learning step at time t
according to the general learning rule is

∆ w i ( t ) = cr [ w i ( t ), x ( t ), d i ( t )] x ( t ) .................... (17)

Where c is a positive number called the learning constant that determines the rate of
learning. The weight vector adapted at time t becomes at the next instant, or learning
step,
w i ( t + 1) = w i ( t ) + c r [ w i ( t ), x ( t ), d i ( t ) ] x ( t ) ....... (18)

The superscript convention will be used in this text to index the discerete-time
training steps as in equ. (18). For the k’th step we thus have from (18) using this
convention.

w ik + 1 = w ik + cr ( w ik , x k , d ik ) x k ......................... (19)

The learning in (18,19) assumes the form of a sequence of discerte- time weight
modifications. Continuous-time learning can be expressed as :
4
dwi ( t )
= crx ( t ) ........................................................ (20)
dt

w1
x1

w2
x2
O=f(net)
f(net)

Wn
f’(net)
Xn ∆W
-
r
d
X X X + +

Fig. (11) Neural Networks learning algorithm

3.6.2. Delta learning rule


The Delta learning rule is valid for continuos activation functions, and
in the supervised training mode. The learning signal for this rule is called delta
and is defined as follows

r ≅ [ d i - f ( w it X )] f '( w it X ) ...................(21)

The term f '(Wi t X ) is the derivative of the activation function f(net)

computed for net= wit X . The explanation of the delta learning rule is shown in
fig. (3.11) this learning rule can be readily derived from the condition of least
square error between oi and d i . Calculating the gradient vector with respect
to wi of the square error defined as

E ≅ 0 .5 (d i - o i ) 2 ..............................(22)

Which is equivalent to

5
E = 0.5 [ d i - f ( w it X )] 2 ...........................(23)

We Obtain the error gradient vector are

∇ E = - ( d i - o i ) f '( w it X ) X ......................(24)

The component of the gradient vector are

∂ E / ∂ W ij = - ( d i - o i ) f '(W i t X ) X i .......(25)

Since the minimization of the error requires the weight changes to be in the
negative gradient direction, we take

∆Wi=−c ∇Ε ........................................(26)

Where η is a positive constant. We then obtain from equations.

∆wi=c(di−oi)f’(neti)x .....................(27)

Οr, for the single weight the adjustment becomes

∆wij=c(di-oi)f’(neti)xj ....................(28)

Νote that the weight adjustment as in equation is computed based on


minimization of the squared error. considering the use of the general learning
rule and plugging in the learning signal as defined in equation , the weight
adjustment becomes

∆wi=c(di-oi)f’(neti)xj ...................(29)

Science c has been assumed to be arbitrary constant. The weights are


initialized at any values for this method of training. The delta rules was
introduced only recently for neural networks training (McClelland and
Rumelhart 1986). this rule parallels the discrete perceptron training rule. It
also can be called the continuous perceptron training rule. it also can be

6
called the continuous perceptron training rule. The delta learning rule can be
generalized for multilayer networks.

6.3. Widrow-Hoff learning rule

The Windrow-Hoff learning rule (Widrow 1962) is applicable for the


supervised training of neural networks. It is independent of the activation
function of neurons used since it minimizes the squared error between the
desired output value di and the neuron’s activation value neti = wit x . The

learning signal for this rule is defined as follows:

r = d i - wit x .............................. (30)

The weight vector increment under this learning rule is

∆wi = c ( d - wit x ) x ..................... (31)

or, for the single weight the adjustment is

∆wij = c(di - wit x) x j , for j=1,2, .. , n .............. (32)

This rule can be considered a special case of the Delta learning rule.
Indeed, assuming that f(net)=net, we obtain that f’(net)=1. this rule sometimes
called the LMS (Least Mean Square) learning rule. Weights are initialized at
any values in this method.

You might also like