Chapter 5 Artificial Intelligence Notes
Chapter 5 Artificial Intelligence Notes
Machine Learning
Machine learning is a type of artificial intelligence (AI) that provides computers with the ability to learn
without being explicitly programmed. Machine learning focuses on the development of computer
programs that can teach themselves to grow and change when exposed to new data.
The process of machine learning is similar to that of data mining. Both systems search through data to
look for patterns. However, instead of extracting data for human comprehension, machine learning uses
that data to detect patterns in data and adjust program actions accordingly. Machine learning algorithms
are often categorized as being supervised or unsupervised. Supervised algorithms can apply what has
been learned in the past to new data. Unsupervised algorithms can draw inferences from datasets.
Facebook's News Feed uses machine learning to personalize each member's feed. If a member
frequently stops scrolling in order to read or "like" a particular friend's posts, the News Feed will start to
show more of that friend's activity earlier in the feed. Behind the scenes, the software is simply
using statistical analysis and predictive analytics to identify patterns in the user's data and use to patterns
to populate the News Feed.
Machine learning is particularly attractive in several real life problem because of the following reasons:
Some tasks cannot be defined well except by example
Working environment of machines may not be known at design time
Explicit knowledge encoding may be difficult and not available
Environments change over time
Biological systems learn
Recently, learning is widely used in a number of application areas including,
Data mining and knowledge discovery
Speech/image/video (pattern) recognition
Adaptive control
Autonomous vehicles/robots
Decision support systems
Bioinformatics
WWW
Formally, a computer program is said to learn from experience E with respect to some class of tasks T
and performance measure P, if its performance at tasks in T, as measured by P, improves with experience
E.
Thus a learning system is characterized by:
task T
experience E, and
performance measure P
Examples:
Learning to drive a van
T: Drive on a public highway using vision sensors
P: Average distance traveled before an error (according to human observer)
E: Sequence of images and steering actions recorded during human driving.
Learning
Learning is the act of acquiring new, or modifying and reinforcing,
existing knowledge, behaviors, skills, values, or preferences and may involve synthesizing different types
of information. The ability to learn is possessed by humans, animals, plants and some machines. Progress
over time tends to follow a learning curve. It does not happen all at once, but builds upon and is shaped
by previous knowledge. To that end, learning may be viewed as a process, rather than a collection of
factual and procedural knowledge. Learning produces changes in the organism and the changes produced
are relatively permanent.
The operation of a neural network is determined by the values of the interconnection weights. There is no
algorithm that determines how the weights should be assigned in order to solve specific problems. Hence,
the weights are determined by a learning process
Learning by Analogy
Analogy is a reasoning process of transferring information from a particular subject to
another. In a narrower sense, analogy is an inference or an argument from one particular to
another particular, as opposed to deduction and induction where at least one of the evidences
or the conclusion is general.
Analogy plays a significant role in problem solving such as, decision making, perception,
memory, creativity, emotion, explanation, and communication. It lies behind basic tasks such
as the identification of places, objects and people, for example, in face perception and facial
recognition systems.
Inductive Learning
Inductive Learning is a powerful strategy for helping machine deepen their understanding of
content and develop their inference and evidence-gathering skills. In an Inductive Learning,
Machines examine, group, and label specific "bits" of information to find patterns. For
example, if given 20 specific weather-related terms (e.g., rain, sleet, snow, hygrometer, rain
gauge, thermometer, humid, dry, windy, cold), machine might group the terms into an initial
set of categories labeled precipitation, weather instruments, and how weather feels. Inductive
Learning does not stop at categorization, however; it also instruct machine to use their
labeled groups to develop a set of working hypotheses about the content to come. Then,
during the learning, machine collect evidence to verify or refine each of their hypotheses.
Explanation based Learning
Explanation-based learning (EBL) is a form of machine learning that exploits a very strong,
or even perfect, domain theory to make generalizations or form concepts from training
examples.
An example of EBL using a perfect domain theory is a program that learns to play chess by
being shown examples. A specific chess position that contains an important feature, say,
"Forced loss of black queen in two moves," includes many unrelated features, such as the
specific scattering of pawns on the board. EBL can take a single training example and
determine what the relevant features are in order to form a generalization
Learning may be classified into two categories:
a. Supervised Learning
b. Unsupervised Learning
Consider a machine (or living organism) which receives some sequence of inputs x1, x2, x3, . . ., where
xt is the sensory input at time t. In supervised learning the machine is given a sequence of input & a
sequence of desired outputs y1, y2, . . . , and the goal of the machine is to learn to produce the correct
output given a new input. While, in unsupervised learning the machine simply receives inputs x1, x2, x3,
. ., but obtains neither supervised target outputs, nor rewards from its environment.
Supervised Learning:
In supervised learning, the network is presented with inputs together with the target (teacher signal)
outputs. Then, the neural network tries to produce an
output as close as possible to the target signal by
adjusting the values of internal weights. The most
common supervised learning method is the “error
correction method”.
Error correction method is used for networks which
their neurons have discrete output functions. Neural
networks are trained with this method in order to reduce the error (difference between the network's
output and the desired output) to zero.
Naïve Bayes Classifier
In machine learning, naive Bayes classifiers are a family of simple probabilistic classifiers based
on applying Bayes' theorem with strong (naive) independence assumptions between the features.
Naive Bayes classifiers are highly scalable, requiring a number of parameters linear in the number
of variables (features/predictors) in a learning problem.
Logistics Regression
Logistic regression is used whenever you need to assign an input to one of several classes. It
applies a logistic function to a linear combination of features, yielding a probability for an input to
be in one of the classes. The output is usually binary, but logistic regression can also be applied to
multiclass classification problems.
A function model with smooth switching:
Where:
is the connection strength between unit and unit .