Genetic Algorithms - Introduction
Genetic Algorithms - Introduction
Introduction to Optimization
Optimization is the process of making something better. In any process, we have a set of inputs and a
set of outputs as shown in the following figure.
Optimization refers to finding the values of inputs in such a way that we get the “best” output values.
The definition of “best” varies from problem to problem, but in mathematical terms, it refers to
maximizing or minimizing one or more objective functions, by varying the input parameters.
The set of all possible solutions or values which the inputs can take make up the search space. In this
search space, lies a point or a set of points which gives the optimal solution. The aim of optimization
is to find that point or set of points in the search space.
GAs were developed by John Holland and his students and colleagues at the University of Michigan,
most notably David E. Goldberg and has since been tried on various optimization problems with a high
degree of success.
In GAs, we have a pool or a population of possible solutions to the given problem. These solutions
then undergo recombination and mutation (like in natural genetics), producing new children, and the
process is repeated over various generations. Each individual (or candidate solution) is assigned a
fitness value (based on its objective function value) and the fitter individuals are given a higher chance
to mate and yield more “fitter” individuals. This is in line with the Darwinian Theory of “Survival of the
Fittest”.
Page 2 of 3
In this way we keep “evolving” better individuals or solutions over generations, till we reach a stopping
criterion.
Genetic Algorithms are sufficiently randomized in nature, but they perform much better than random
local search (in which we just try various random solutions, keeping track of the best so far), as they
exploit historical information as well.
Advantages of GAs
GAs have various advantages which have made them immensely popular. These include −
Does not require any derivative information (which may not be available for many real-world
problems).
Optimizes both continuous and discrete functions and also multi-objective problems.
Always gets an answer to the problem, which gets better over the time.
Useful when the search space is very large and there are a large number of parameters
involved.
Limitations of GAs
Like any technique, GAs also suffer from a few limitations. These include −
GAs are not suited for all problems, especially problems which are simple and for which
derivative information is available.
Fitness value is calculated repeatedly which might be computationally expensive for some
problems.
Being stochastic, there are no guarantees on the optimality or the quality of the solution.
If not implemented properly, the GA may not converge to the optimal solution.
Explore our latest online courses and learn new skills at your own pace. Enroll and become a certified
expert to boost your career.
GA – Motivation
Genetic Algorithms have the ability to deliver a “good-enough” solution “fast-enough”. This makes
genetic algorithms attractive for use in solving optimization problems. The reasons why GAs are
needed are as follows −
Page 3 of 3
In computer science, there is a large set of problems, which are NP-Hard. What this essentially means
is that, even the most powerful computing systems take a very long time (even years!) to solve that
problem. In such a scenario, GAs prove to be an efficient tool to provide usable near-optimal solutions
in a short amount of time.
Traditional calculus based methods work by starting at a random point and by moving in the direction
of the gradient, till we reach the top of the hill. This technique is efficient and works very well for
single-peaked objective functions like the cost function in linear regression. But, in most real-world
situations, we have a very complex problem called as landscapes, which are made of many peaks and
many valleys, which causes such methods to fail, as they suffer from an inherent tendency of getting
stuck at the local optima as shown in the following figure.
Some difficult problems like the Travelling Salesperson Problem (TSP), have real-world applications
like path finding and VLSI Design. Now imagine that you are using your GPS Navigation system, and it
takes a few minutes (or even a few hours) to compute the “optimal” path from the source to
destination. Delay in such real world applications is not acceptable and therefore a “good-enough”
solution, which is delivered “fast” is what is required.