NOTES Unit - II
NOTES Unit - II
The simplest agents were the reflex agents, which base their actions on a direct
mapping from states to actions. Such agents cannot operate well in environments for
which this mapping would be too large to store and would take too long to learn.
Goal-based agents, on the other hand, consider future actions and the desirability
of their outcomes. One kind of goal-based agent called a problem-solving agent. Problem-
solving agents use atomic representations, states of the world are considered as wholes,
with no internal structure visible to the problem-solving algorithms. Goal-based agents
that use more advanced factored or structured representations are usually called planning
agents.
Our discussion of problem solving begins with precise definitions of problems and
their solutions. Several general-purpose search algorithms that can be used to solve these
problems. Several uninformed search algorithms—algorithms that are given no
information about the problem other than its definition. Although some of these
algorithms can solve any solvable problem, none of them can do so efficiently. Informed
search algorithms, on the other hand, can do quite well given some guidance on where
to look for solutions.
PROBLEM-SOLVING AGENTS
In Artificial Intelligence, Search techniques are universal problem-solving methods.
Rational agents or Problem-solving agents in AI mostly used these search strategies or
algorithms to solve a specific problem and provide the best result. Problem-solving agents
are the goal-based agents and use atomic representation. The problem-solving agent
performs precisely by defining problems and its several solutions.
A problem-solving refers to a state where we wish to reach to a definite goal from a
present state or condition or a problem-solving is a part of artificial intelligence which
encompasses a number of techniques such as algorithms, heuristics to solve a problem.
Therefore, a problem-solving agent is a goal-driven agent and focuses on satisfying
the goal.
Steps performed by Problem-solving agent
• Goal formulation: It is the first and simplest step in problem-solving. It organizes
the steps/sequence required to formulate one goal out of multiple goals as well as actions
to achieve that goal. Goal formulation is based on the current situation and the agent's
performance measure.
• Problem formulation: It is one of the core steps of problem-solving which decides
what action should be taken to achieve the formulated goal. In AI this core part is
dependent upon software agent which consisted of the following components to formulate
the associated problem.
• Initial State: It is the starting state or initial step of the agent towards its goal.
• Actions: It is the description of the possible actions available to the agent.
• Transition Model: It describes what each action does.
• Goal Test: It determines if the given state is a goal state.
• Path cost: It assigns a numeric cost to each path that follows the goal. The
problem-solving agent selects a cost function, which reflects its performance measure.
Remember, an optimal solution has the lowest path cost among all the solutions.
The process of looking for a sequence of actions that reaches the goal is called
search. A search algorithm takes a problem as input and returns a solution in the form
of an action sequence. Once a solution is found, the actions it recommends can be carried
out. This is called the execution phase. Thus, we have a simple “formulate, search,
execute” design for the agent.
1. Breadth-first Search:
Breadth-first search is the most common search strategy for traversing a tree or
graph. This algorithm searches breadthwise in a tree or graph, so it is called breadth-first
search.
BFS algorithm starts searching from the root node of the tree and expands all
successor node at the current level before moving to nodes of next level.
The breadth-first search algorithm is an example of a general-graph search algorithm.
Breadth-first search implemented using FIFO queue data structure.
Advantages:
• BFS will provide a solution if any solution exists.
• If there are more than one solution for a given problem, then BFS will provide the
minimal solution which requires the least number of steps.
Disadvantages:
• It requires lots of memory since each level of the tree must be saved into memory
to expand the next level.
• BFS needs lots of time if the solution is far away from the root node.
Example:
In the below tree structure, we have shown the traversing of the tree using BFS
algorithm from the root node S to goal node K. BFS search algorithm traverse in layers, so
it will follow the path which is shown by the dotted arrow, and the traversed path will be:
S---> A--->B---->C--->D---->G--->H--->E---->F---->I---->K
Time Complexity:
Time Complexity of BFS algorithm can be obtained by the number of nodes traversed
in BFS until the shallowest Node. Where the d= depth of shallowest solution and b is a
node at every state.
T (b) = 1+b2+b3+.......+ bd= O (bd)
Space Complexity:
Space complexity of BFS algorithm is given by the Memory size of frontier which is
O(bd).
Completeness:
BFS is complete, which means if the shallowest goal node is at some finite depth,
then BFS will find a solution.
Optimality:
BFS is optimal if path cost is a non-decreasing function of the depth of the node.
2. Depth-first Search
Depth-first search isa recursive algorithm for traversing a tree or graph data
structure. It is called the depth-first search because it starts from the root node and
follows each path to its greatest depth node before moving to the next path.
DFS uses a stack data structure for its implementation. The process of the DFS
algorithm is similar to the BFS algorithm.
Advantage:
• DFS requires very less memory as it only needs to store a stack of the nodes on
the path from root node to the current node.
• It takes less time to reach to the goal node than BFS algorithm (if it traverses in
the right path).
Disadvantage:
• There is the possibility that many states keep re-occurring, and there is no
guarantee of finding the solution.
• DFS algorithm goes for deep down searching and sometime it may go to the
infinite loop.
Example:
In the below search tree, we have shown the flow of depth-first search, and it will
follow the order as:
Root node--->Left node ----> right node.
It will start searching from root node S, and traverse A, then B, then D and E, after
traversing E, it will backtrack the tree as E has no other successor and still goal node is
not found. After backtracking it will traverse node C and then G, and here it will terminate
as it found goal node.
Completeness:
DFS search algorithm is complete within finite state space as it will expand every
node within a limited search tree.
Time Complexity:
Time complexity of DFS will be equivalent to the node traversed by the algorithm. It
is given by:
T(n)= 1+ n2+ n3 +.........+ nm=O(nm)
Where, m= maximum depth of any node and this can be much larger than d
(Shallowest solution depth)
Space Complexity:
DFS algorithm needs to store only single path from the root node, hence space
complexity of DFS is equivalent to the size of the fringe set, which is O(bm).
Optimal:
DFS search algorithm is non-optimal, as it may generate a large number of steps or
high cost to reach to the goal node.
Advantages:
Depth-limited search is Memory efficient.
Disadvantages:
Depth-limited search also has a disadvantage of incompleteness.
It may not be optimal if the problem has more than one solution.
Example:
Completeness:
DLS search algorithm is complete if the solution is above the depth-limit.
Time Complexity:
Time complexity of DLS algorithm is O(bℓ).
Space Complexity:
Space complexity of DLS algorithm is O(b×ℓ).
Optimal:
Depth-limited search can be viewed as a special case of DFS, and it is also not optimal
even if ℓ>d.
Advantages:
Uniform cost search is optimal because at every state the path with the least cost is
chosen.
Disadvantages:
It does not care about the number of steps involve in searching and only concerned
about path cost.
Due to which this algorithm may be stuck in an infinite loop.
Example:
Completeness:
Uniform-cost search is complete, such as if there is a solution, UCS will find it.
Time Complexity:
Let C* is Cost of the optimal solution, and ε is each step to get closer to the goal node.
Then the number of steps is = C*/ε+1. Here we have taken +1, as we start from state
0 and end to C*/ε.
Hence, the worst-case time complexity of Uniform-cost search isO(b1 + [C*/ε])/.
Space Complexity:
The same logic is for space complexity so, the worst-case space complexity of
Uniform-cost search is O(b1 + [C*/ε]).
Optimal:
Uniform-cost search is always optimal as it only selects a path with the lowest path
cost.
Advantages:
It combines the benefits of BFS and DFS search algorithm in terms of fast search and
memory efficiency.
Disadvantages:
The main drawback of IDDFS is that it repeats all the work of the previous phase.
Example:
Following tree structure is showing the iterative deepening depth-first search. IDDFS
algorithm performs various iterations until it does not find the goal node. The iteration
performed by the algorithm is given as:
1'st Iteration-----> A
2'nd Iteration----> A, B, C
3'rd Iteration------>A, B, D, E, C, F, G
4'th Iteration------>A, B, D, H, I, E, C, F, K, G
In the fourth iteration, the algorithm will find the goal node.
Completeness:
This algorithm is complete is if the branching factor is finite.
Time Complexity:
Let's suppose b is the branching factor and depth is d then the worst-case time
complexity is O(bd).
Space Complexity:
The space complexity of IDDFS will be O(bd).
Optimal:
IDDFS algorithm is optimal if path cost is a non- decreasing function of the depth of
the node.
Advantages:
Bidirectional search is fast.
Bidirectional search requires less memory
Disadvantages:
Implementation of the bidirectional search tree is difficult.
In bidirectional search, one should know the goal state in advance.
Example:
In the below search tree, bidirectional search algorithm is applied. This algorithm
divides one graph/tree into two sub-graphs.
It starts traversing from node 1 in the forward direction and starts from goal node 16
in the backward direction.
The algorithm terminates at node 9 where two searches meet.
Completeness:
Bidirectional Search is complete if we use BFS in both searches.
Time Complexity:
Time complexity of bidirectional search using BFS is O(bd).
Space Complexity:
Space complexity of bidirectional search is O(bd).
Optimal:
Bidirectional search is Optimal.
In the informed search we will discuss two main algorithms which are given below:
• Best First Search Algorithm (Greedy search)
• A* Search Algorithm
Disadvantages:
• It can behave as an unguided depth-first search in the worst-case scenario.
• It can get stuck in a loop as DFS.
• This algorithm is not optimal.
Example:
Consider the below search problem, and we will traverse it using greedy best-first
search. At each iteration, each node is expanded using evaluation function f(n)=h(n), which
is given in the below table.
In this search example, we are using two lists which are OPEN and CLOSED Lists.
Following are the iteration for traversing the above example.
Expand the nodes of S and put in the CLOSED list
Initialization: Open [A, B], Closed [S]
Time Complexity:
The worst-case time complexity of Greedy best first search is O(bm).
Space Complexity:
The worst-case space complexity of Greedy best first search is O(bm). Where, m is the
maximum depth of the search space.
Complete:
Greedy best-first search is also incomplete, even if the given state space is finite.
Optimal:
Greedy best first search algorithm is not optimal.
A* Search Algorithm:
A* search is the most commonly known form of best-first search. It uses the heuristic
function h(n), and cost to reach the node n from the start state g(n). It has combined
features of UCS and greedy best-first search, by which it solves the problem efficiently. A*
search algorithm finds the shortest path through the search space using the heuristic
function.
This search algorithm expands fewer search trees and provides optimal results faster.
A* algorithm is similar to UCS except that it uses g(n)+h(n)instead of g(n). In A* search
algorithm, we use the search heuristic as well as the cost to reach the node. Hence, we
can combine both costs as follows, and this sum is called a fitness number.
Algorithm of A* search:
Step 1: Place the starting node in the OPEN list.
Step 2: Check if the OPEN list is empty or not, if the list is empty then return failure and
stops.
Step 3: Select the node from the OPEN list which has the smallest value of evaluation
function (g+h), if node n is the goal node, then return success and stop, otherwise
Step 4: Expand node n and generate all of its successors, and put n into the closed list.
For each successor n', check whether n' is already in the OPEN or CLOSED list, if not then
compute the evaluation function for n' and place it into the Open list.
Step 5: Else if node n' is already in OPEN and CLOSED, then it should be attached to the
back pointer which reflects the lowest g(n') value.
Step 6: Return to Step 2.
Advantages:
• A* search algorithm is the best algorithm than other search algorithms.
• A* search algorithm is optimal and complete.
• This algorithm can solve very complex problems.
Disadvantages:
• It does not always produce the shortest path as it is mostly based on heuristics
and approximation.
• A* search algorithm has some complexity issues.
• The main drawback of A* is memory requirement as it keeps all generated nodes
in the memory, so it is not practical for various large-scale problems.
Example:
In this example, we will traverse the given graph using the A* algorithm. The heuristic
value of all states is given in below table so we will calculate the f(n) of each state using
the formula f(n)= g(n) + h(n), where g(n) is the cost to reach any node from the start state.
Here we will use the OPEN and CLOSED list.
Solution:
Iteration 3: {(S--> A-->C--->G, 6), (S--> A-->C--->D, 11), (S--> A-->B, 7), (S-->G, 10)}
Iteration 4: Will give the final result, as S--->A--->C--->G it provides the optimal path
with cost 6.
Points to remember:
• A* algorithm returns the path which occurred first, and it does not search for all
remaining paths.
• The efficiency of the A* algorithm depends on the quality of the heuristic.
• A* algorithm expands all nodes which satisfy the condition f(n)
Complete:
• A* algorithm is complete as long as:
• The branching factor is finite.
• The cost at every action is fixed.
Optimal:
A* search algorithm is optimal if it follows below two conditions:
Admissible:
The first condition required for optimality is that h(n) should be an admissible
heuristic for A* tree search. An admissible heuristic is optimistic in nature.
Consistency:
Second required condition is consistency for only A* graph-search.
If the heuristic function is admissible, then A* tree search will always find the least cost
path.
Time Complexity:
The time complexity of the A* search algorithm depends on the heuristic function,
and the number of nodes expanded is exponential to the depth of solution d. So, the time
complexity is O(b^d), where b is the branching factor.
Space Complexity:
The space complexity of the A* search algorithm is O(b^d)
Note: Both simple, as well as steepest-ascent hill-climbing search, fails when there
is no closer node.
Local Maxima:
It is that peak of the mountain that is highest than all its neighbouring states but
lower than the global maxima. It is not the goal peak because there is another peak higher
than it.
Plateau:
It is a flat surface area where no uphill exists. It becomes difficult for the climber to
decide in which direction he should move to reach the goal point. Sometimes, the person
gets lost in the flat area.
Ridges:
It is a challenging problem where the person finds two or more local maxima of the
same height commonly. It becomes difficult for the person to navigate the right point and
stuck to that point itself.
Note: A variant of Local Beam Search is Stochastic Beam Search which selects k
successors at random rather than choosing the best k successors.
Genetic Algorithm:
Genetic Algorithm (GA) is a search-based optimization technique based on the
principles of Genetics and Natural Selection. It is frequently used to find optimal or near-
optimal solutions to difficult problems which otherwise would take a lifetime to solve. It is
frequently used to solve optimization problems, in research, and in machine learning.
Introduction to Optimization
Optimization is the process of making something better. In any process, we have a
set of inputs and a set of outputs as shown in the following figure.
Optimization refers to finding the values of inputs in such a way that we get the
“best” output values. The definition of “best” varies from problem to problem, but in
mathematical terms, it refers to maximizing or minimizing one or more objective functions,
by varying the input parameters.
The set of all possible solutions or values which the inputs can take make up the
search space. In this search space, lies a point or a set of points which gives the optimal
solution. The aim of optimization is to find that point or set of points in the search space.
Advantages of GAs
GAs have various advantages which have made them immensely popular. These
include −
• Does not require any derivative information (which may not be available for
many real-world problems).
• Is faster and more efficient as compared to the traditional methods.
• Has very good parallel capabilities.
• Optimizes both continuous and discrete functions and also multi-objective
problems.
• Provides a list of “good” solutions and not just a single solution.
• Always gets an answer to the problem, which gets better over the time.
• Useful when the search space is very large and there are a large number of
parameters involved.
Limitations of GAs
Like any technique, GAs also suffer from a few limitations. These include −
• GAs are not suited for all problems, especially problems which are simple and
for which derivative information is available.
• Fitness value is calculated repeatedly which might be computationally expensive
for some problems.
• Being stochastic, there are no guarantees on the optimality or the quality of the
solution.
• If not implemented properly, the GA may not converge to the optimal solution.
GA – Motivation
Genetic Algorithms have the ability to deliver a “good-enough” solution “fast-
enough”. This makes genetic algorithms attractive for use in solving optimization
problems. The reasons why GAs are needed are as follows −
Basic Structure
The basic structure of a GA is as follows −
We start with an initial population (which may be generated at random or seeded
by other heuristics), select parents from this population for mating. Apply crossover and
mutation operators on the parents to generate new off-springs. And finally these off-
springs replace the existing individuals in the population and the process repeats. In this
way genetic algorithms actually try to mimic the human evolution to some extent.
Each of the following steps are covered as a separate chapter later in this tutorial.
Initial Population
The process begins with a set of individuals which is called a Population. Each
individual is a solution to the problem you want to solve.
An individual is characterized by a set of parameters (variables) known as Genes.
Genes are joined into a string to form a Chromosome (solution).
In a genetic algorithm, the set of genes of an individual is represented using a string,
in terms of an alphabet. Usually, binary values are used (string of 1s and 0s). We say that
we encode the genes in a chromosome.
Population is a subset of solutions in the current generation. It can also be defined
as a set of chromosomes. There are several things to be kept in mind when dealing with
GA population −
The diversity of the population should be maintained otherwise it might lead to
premature convergence.
The population size should not be kept very large as it can cause a GA to slow down,
while a smaller population might not be enough for a good mating pool. Therefore, an
optimal population size needs to be decided by trial and error.
The population is usually defined as a two dimensional array of – size population,
size x, chromosome size.
Population Initialization
There are two primary methods to initialize a population in a GA. They are −
Random Initialization − Populate the initial population with completely random
solutions.
Heuristic initialization − Populate the initial population using a known heuristic
for the problem.
It has been observed that the entire population should not be initialized using a
heuristic, as it can result in the population having similar solutions and very little
diversity. It has been experimentally observed that the random solutions are the ones to
drive the population to optimality. Therefore, with heuristic initialization, we just seed the
population with a couple of good solutions, filling up the rest with random solutions rather
than filling the entire population with heuristic based solutions.
It has also been observed that heuristic initialization in some cases, only effects the
initial fitness of the population, but in the end, it is the diversity of the solutions which
lead to optimality.
Population Models
There are two population models widely in use −
Steady State
In steady state GA, we generate one or two off-springs in each iteration and they
replace one or two individuals from the population. A steady state GA is also known as
Incremental GA.
Generational
In a generational model, we generate ‘n’ off-springs, where n is the population size,
and the entire population is replaced by the new one at the end of the iteration.
Selection
The idea of selection phase is to select the fittest individuals and let them pass their
genes to the next generation.
Two pairs of individuals (parents) are selected based on their fitness scores.
Individuals with high fitness have more chance to be selected for reproduction.
Crossover
Crossover is the most significant phase in a genetic algorithm. For each pair of
parents to be mated, a crossover point is chosen at random from within the genes.
For example, consider the crossover point to be 3 as shown below.
Offspring:
Offspring are created by exchanging the genes of parents among themselves until the
crossover point is reached.
Mutation
In certain new offspring formed, some of their genes can be subjected to
a mutation with a low random probability. This implies that some of the bits in the bit
string can be flipped.
Termination
The algorithm terminates if the population has converged (does not produce offspring
which are significantly different from the previous generation). Then it is said that the
genetic algorithm has provided a set of solutions to our problem.