Basics of Learning Theory
Basics of Learning Theory
instruction, or study. It involves a change in behavior or mental processes that occurs as a result of new
information or experiences.
• What is the best way to choose a learning algorithm given a particular learning problem?
• How can we ensure that a learning algorithm generalizes well to new data?
Statistical learning theory is a framework for machine learning that draws from statistics and functional
analysis. It deals with finding a predictive function based on the data presented. The main idea in statistical
learning theory is to build a model that can draw conclusions from data and make predictions.
For example, statistical learning theory can be used to analyze the performance of linear regression or support
vector machines on a given dataset and to prove theorems about their convergence rates and error bounds.
Emphasis on Vapnik-Chervonenkis
Learning Models Emphasis on PAC learning
(VC) learning
Mathematical
Model Predict
Training
Data
ML Algorithm Decide
As data increases
Logical and Efficiency increases
Self-Driving Car Mathematical
Model
Direct Indirect
Consider a machine learning algorithm that is being trained to recognize handwritten digits
Complexity
Size of the
of the target
dataset function House Price
Choosing the right function for approximation involves selecting a model that can capture the underlying patterns in the data
and generalize well to new data. This process often involves experimentation and evaluation of different models, until the
best-performing model is identified for use in the machine learning system.
This step involves testing the system on a separate dataset, known as the test set, to measure its performance on
unseen data.
Performance
Data Splitting Deployment
Analysis
Training Data
The problem of inducing general functions from specific training examples is central to learning. Concept learning
can be formulated as a problem of searching through a predefined space of potential hypotheses for the hypothesis
that best fits the training examples.
Each concept or category obtained by learning is a Boolean valued function that takes true or false value.
When a new sample with the values for attributes <Sky, Air Temperature, Humidity, Wind, Water, Forecast> is given, the
value for EnjoySport (ie. 0 or 1) is predicted based on the previous learning.
So we can define, Inductive learning hypothesis is any hypothesis found to approximate the target function well
over a sufficiently large set of training examples will also approximate the target function well over any other
unobserved examples.
All such hypothesis having same semantic is counted as 1. So we can have total number of combinations as below.
1 (hypothesis with one or more 0) + 4×3×3×3×3×3 (add ? to each attribute) = 973 semantically distinct hypotheses
The concept learning can be viewed as the task of searching through a large space of hypotheses. The goal of this search
is to find the hypothesis that best fits the training examples. By selecting a particular hypothesis representation, the
designer of the learning algorithm implicitly defines the space of all hypotheses that the program can ever represent and
therefore can ever learn.
Examples of heuristic search algorithms used in machine learning and AI include hill-climbing, simulated annealing,
genetic algorithms, and particle swarm optimization. Each of these algorithms uses a different heuristic function and
search strategy to efficiently explore the space of possible solutions and find the best fit to the available data.