Mod3 - Learning Theory
Mod3 - Learning Theory
The process of acquiring knowledge and expertise through study, experience, or being taught is
called as learning. Generally, humans learn in different ways. To make machines learn, we need to
simulate the strategies of human learning in machines. But will the machines learn? This depends
on the nature of problem that the computers can solve. There are two kinds of problems: well-
posed and ill-posed. Computers can only solve well-posed problems, as these have well-defined
specifications and have the following components inherent to it.
The standard definition of learning proposed by Tom Mitchell is that a program can learn from E
for the Task T, and P improves with experience E.
Let x be the input and X be the input space, which is the set of all inputs, and Y is the output
space, which is the set of all possible outputs. Let the unknown target function be f:X Y, that
maps the input space to output space. The objective of the learning program is to pick a function,
g: X Y to approximate hypothesis f. All the possible formulae from a hypothesis space. In
short, let H be the set of all formulae form which the learning algorithm chooses. The choice is
good when the hypothesis g replicates f for all samples.
It can be observed that training samples and target function are dependent on the given problem.
The learning algorithm and hypothesis set are independent of the given problem. Thus, learning
model is informally the hypothesis set and learning algorithm. Thus, learning model can stated as
follows:
4. Learn by critical thinking, also called as deductive learning, deduces new facts or
conclusions from the related known facts and information.
5. Self learning, also called as reinforcement learning, is a self-directed learning that normally
learns from mistakes punishments and rewards.
6. Learning to solve problems is a type of cognitive learning where learning happens in the
mind and is possible by devising a methodology to achieve a goal. Here, the learner initially is not
aware of the solution or the way to achieve the goal but only knows the goal. The learning
happens either directly from the initial state by following the steps to achieve the goal or indirectly
by inferring the behaviour.
These questions are the basis of a field called 'computational learning theory' or in short (COLT).
It is a specialized field of study of machine learning. COLT deals with the formal method used for
learning systems. It deals with Framework for quantifying learning task and learning algorithms.
It provides of fundamental basis for study of machine learning. Computational learning theory
uses many concepts from divorce areas such as theoretical computer science, artificial Intelligence
and statistics. The core concept of COLT is the concept of learning framework. One such
Step 1) Choosing the Training Experience: The very important and first task is to choose the
training data or training experience which will be fed to the Machine Learning Algorithm. It is
important to note that the data or experience that we fed to the algorithm must have a significant
impact on the Success or Failure of the Model. So Training data or experience should be chosen
wisely.
Below are the attributes which will impact on Success and Failure of Data:
The training experience will be able to provide direct or indirect feedback regarding choices.
For example: While Playing chess the training data will provide feedback to itself like
instead of this move if this is chosen the chances of success increases.
Second important attribute is the degree to which the learner will control the sequences of
training examples. For example: when training data is fed to the machine then at that time
accuracy is very less but when it gains experience while playing again and again with itself
or opponent the machine algorithm will get feedback and control the chess game accordingly.
Third important attribute is how it will represent the distribution of examples over which
performance will be measured. For example, a Machine learning algorithm will get
experience while going through a number of different cases and different examples. Thus,
Machine Learning Algorithm will get more and more experience by passing through more
and more examples and hence its performance will increase.
Step 2- Choosing target function: The next important step is choosing the target function. It
means according to the knowledge fed to the algorithm the machine learning will choose
NextMove function which will describe what type of legal moves should be taken. For example
: While playing chess with the opponent, when opponent will play then the machine learning
algorithm will decide what be the number of possible legal moves taken in order to get succes s.
Step 3- Choosing Representation for Target function: When the machine algorithm will
know all the possible legal moves the next step is to choose the optimized move using any
representation i.e. using linear Equations, Hierarchical Graph Representation, Tabular form etc.
The NextMove function will move the Target move like out of these move which will provide
more success rate. For Example : while playing chess machine have 4 possible moves, so the
machine will choose that optimized move which will provide success to it.
A set of example days, and each is described by six attributes. The task is to learn to predict the
value of EnjoySport for arbitrary day, based on the values of its attribute values.
Concept learning can be viewed as the task of searching through a large space of hypotheses
implicitly defined by the hypothesis representation. The goal of this search is to find the
hypothesis that best fits the training examples. By selecting a hypothesis representation, the
designer of the learning algorithm implicitly defines the space of all hypotheses that the program
can ever represent and therefore can ever learn.
Sky has 3 possible values, and other 5 attributes have 2 possible values. There are 96 (=
3.2.2.2.2.2) distinct instances in X. There are 5120 (=5.4.4.4.4.4) syntactically distinct hypotheses
in H. Two more values for attributes: ? and 0 . Every hypothesis containing one or more 0
symbols represents the empty set of instances; that is, it classifies every instance as negative.
There are 973 (= 1 + 4.3.3.3.3.3) semantically distinct hypotheses in H. Only one more value for
attributes: ?, and one hypothesis representing empty set of instances. Although EnjoySport has
5 Dept of CSE, Geetha N
Module 3 AI & ML
small, finite hypothesis space, most learning tasks have much larger (even infinite) hypothesis
spaces. – We need efficient search algorithms on the hypothesis spaces.
FIND-S Algorithm
Training Examples:
Step 3 - The final maximally specific hypothesis is <Sunny, Warm, ?, Strong, ?, ?>
Example 2 :
3. Apply the FIND-S algorithm by hand on the given training set. Consider the examples in the
specified order and write down your hypothesis each time after observing an example.
Step 1: h0 = (ø, ø, ø, ø, ø)
Step 2:
First iteration X1 = (some, small, no, expensive, many) – No
Negative Example Hence Ignore
h1 = (ø, ø, ø, ø, ø)
h5 = (many, ?, no, ?, ?)
Actually, positive examples are used here as the Find-S algorithm (Basically they are
generalizing from the specification). While the negative example is specified in the
generalizing form.
Algorithm
Training Examples:
Example 2: