Sf8 - Fai Unit V Notes
Sf8 - Fai Unit V Notes
Learning
What is learning?
Herbert Simon, 1983
Learning denotes changes in a system that enables a system to do the same task more
efficiently the next time.
Marvin Minsky, 1986
Learning is making useful changes in the workings of our minds.
Ryszard Michalski, 1986
Learning is constructing or modifying representations of what is being experienced.
Mitchell, 1997
A computer program is said to learn from experience E with respect to some class of tasks
T and performance measure P, if its performance at tasks in T, as measured by P, improves
with experience E.
TYPES OF LEARNING
Supervised Learning
Unsupervised Learning
Reinforcement Learning
Supervised learning
Supervised learning is the types of machine learning in which machines are trained using
well "labelled" training data, and on basis of that data, machines predict the output. The
labelled data means some input data is already tagged with the correct output.
In supervised learning, the training data provided to the machines work as the supervisor
that teaches the machines to predict the output correctly. It applies the same concept as a
student learns in the supervision of the teacher.
Supervised learning is a process of providing input data as well as correct output data to
the machine learning model. The aim of a supervised learning algorithm is to find a
mapping function to map the input variable(x) with the output variable(y).
In the real-world, supervised learning can be used for Risk Assessment, Image
classification, Fraud Detection, spam filtering, etc.
Attributes necessary for this evaluation may be or say price P, seat capacity S, Engine E.
Our survey deals with whether a given car is family car or not.
We extract important survey information from domain experts.
We represent each car by ordered pair(X,label)
Where x=(price,seat,engine)
Label=label type ‘1’ or ‘0’.
‘1’for positive example.
‘0’ for negative example.
If the training set contains ‘N’ such examples.
It is given by
X= {(Xk , labelk ),1<=K<=N}
Here, Xk =tuple (price,seat,engine ) of the Kth car
A 3-d graph can be plotted with each Kth instance and its label ( ‘0’ or ‘1’) as data point.
Certain ranges can also be imposed on the these features such as
P1<=Price<=P2
S1<=Seat<=S2
E1<=engine<=E2
These conditions fix hypothesis ‘H’ class, from which region ‘C’ containing positive
examples data points in the space can be drawn.
Each tuple of pairs ((p1,p2),(s1,s2),(e1,e2)) defines one hypothesis ‘h’ from ‘H’ and
choose best hypothesis.
The learning algorithm should then find out a particular hypothesis h£H that approx. ‘C’
as close as possible.
All 6 parameters (P1,P2,E1,E2,S1,S2) from training set to include all positive examples
& no –ve examples.
Now for instance ‘x’ defined as h(x)=1
If ‘h’ classified ‘x’ as positive example and ‘o’ otherwise.
UNSUPERVISED LEARNING
In this a typical task of classification or regression using a set of examples prepared by
humans.
In this learning there is no information available about the correct outputs.
Here learners predict future precepts based up on present ones.
This learning includes clustering and discovery.
Clustering classes are formed for data instances and the method of classifying them are
devised.
Discovery specific goal is not given, it is based on deductive and inductive learning's.
Inductive learning involves extracting rules and patterns from given data.
Deductive learning involves deducing new knowledge from already existing knowledge.
So, deductive is related to discovery and inductive is formulation of opinion using
incomplete information.
Unsupervised learning algorithms among neural networks models self organizing map
and ART(Adaptive Resonance theory)
The goal of learning is the self organizing map is to discover some underlying structure
of data and cause different parts of the network to respond. Similarly to certain input
patterns, these systems adapt over a number iterations using feedback process.
ART the number of clusters can be varies according to the problem size, this model
allows users to control the degree of similarity between members of the same clusters
with the help of a user defined constant vigilance process.
REINFORCEMENT LEARNING
In reinforcement learning the decision making system (also known as agent) receives
rewards or feedback (+ve or –ve) for its action at the end of a sequence of steps.
It is required to assign rewards to step while solving the credit assignment problem.
This problem determines which steps should receive credit or blame for the final result.
Reinforcement learning takes place in an environment where the agent cannot directly
compare the results of its action to a desired result.
It may cause system to win or lose a game, inform a system that it has a good move or
poor move.
The primary task of reinforcement learning is to obtain a successful function using these
rewards.
Basic kinds of information to learn by the system:
1) Utility function: the agent learns the utility of being in various states and choose
actions in order to maximize the expected utility of their outcomes. This require the
agent to keep a model of the environment.
2) Action-Value Function: The agent learns an action value function giving the utility of
performing an action in a given state. This is called Q-learning, it is model free
approach.
Now we can draw inference, here the process of learning is happening with a critic.
Here system works based on credit assignment problem. It learns the internal reward
mechanism and uses it to take local actions in order to maximize it.
NEURAL NETWORK BASED LEARNING
A neural network is a computational model similar to human brain.
It consists of several simple units (neurons) that work in parallel, with no central control.
Connection between these working units are weighted and these weights can be modified
by learning method.
Here Data structure consists of a set of neurons connected by weighted links.
Feed-Forward network is a kind of supervised learning, used in this concept.
The training in this network is of 2 parts
An Input vectorà input neuron
An output vectorà Output neuron.
In the process learning, the input vector is provided at the input layer of the system along
with desired responses, one for each neuron, at the output layer.
When a forward pass is done, the errors and discrepancies between the desired and actual
responses for each neuron in the output layer are obtained.