Asymptotic Notations
Asymptotic Notations
ASYMPTOTIC NOTATIONS
1
ALGORITHM DEFINITION
2
Key Characteristics of an Algorithm
1. Input:
An algorithm has zero or more inputs, taken from a specified set.
2. Output:
It produces at least one output, which is the solution to the problem.
3. Definiteness:
Each step is precisely and unambiguously defined.
4. Finiteness:
The algorithm must terminate after a finite number of steps.
5. Effectiveness:
All operations must be basic and feasible (can be done in finite time with basic
tools or machines).
3
Example (Simple Algorithm)
Algorithm (Max-Find):
1.Set max = A[1]
2.For i = 2 to n:
3.Return max
This is a valid algorithm—it has input, output, and follows all five properties. 4
GOOD ALGORITHMS?
5
MEASURING EFFICIENCY
The efficiency of an algorithm is a measure of the amount of resources consumed in
solving a problem of size n.
The resource we are most interested in is time
We can use the same techniques to analyze the consumption of other resources, such as
memory space.
It would seem that the most obvious way to measure the efficiency of an algorithm
is to run it and measure how much processor time is needed
Is it correct ?
6
FACTORS
Hardware
Operating System
Compiler
Size of input
Nature of Input
Algorithm
7
Which should be improved?
RUNNING TIME OF AN ALGORITHM
Depends upon
Input Size
Nature of Input
Generally time grows with size of input, so running time of an algorithm is usually
measured as function of input size.
9
SIMPLE EXAMPLE (1)
N = 10 => 53 steps
N = 100 => 503 steps
N = 1,000 => 5003 steps
N = 1,000,000 => 5,000,003 steps
12
WHAT DOMINATES IN PREVIOUS
EXAMPLE?
13
ASYMPTOTIC COMPLEXITY
14
COMPARING FUNCTIONS:
ASYMPTOTIC NOTATION
f(N) = O(g(N))
16
BIG O NOTATION: SIMPLE DEFINITION
Big O Notation is a mathematical notation used to describe the upper bound of an
algorithm's time or space complexity. It provides a way to express how the runtime
or space requirements of an algorithm grow as the size of the input increases. In
essence, it characterizes the worst-case scenario of an algorithm’s performance.
Key Points
• Focus on Growth Rate: Big O ignores constant factors and lower-order terms, focusing only
on the term that grows the fastest as the input size increases.
• Worst-case Analysis: It provides a way to evaluate the efficiency of an algorithm in the
worst-case scenario.
17
COMMON BIG O NOTATIONS
• O(1): Constant time - The algorithm's runtime does not change with the size of
the input.
• O(n): Linear time - The runtime grows linearly with the input size.
• O(n^2): Quadratic time - The runtime grows quadratically with the input size.
18
EXAMPLE
Example Algorithm: Linear Search
Description: A simple algorithm to find an element in an array by checking
each element one by one.
Analysis:
•In the worst case, the target element is not in the array, and the algorithm checks every element.
•If the array has nn elements, the time complexity is O(n)O(n). 19
BIG O NOTATION DEFINITION
Formal Definition:
20
EXPLANATION OF TERMS
• 𝒇(𝑵): The function representing the time or space complexity of an
algorithm.
• 𝒈(𝑵): A reference function that describes a growth rate to compare against.
• 𝒄: A constant multiplier that scales 𝑔(𝑁).
• 𝑵𝟎: A threshold beyond which the relationship holds true.
• Sufficiently Large 𝑵: This means we only consider values of 𝑁 that are
greater than 𝑁0.
21
BIG OH NOTATION [2]
22
EXAMPLE (2): COMPARING FUNCTIONS
10 n2 Vs n3 3500
3000
2500
10 n^2
2000
n^3
1500
1000
500
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
23
COMPARING FUNCTIONS
0.05 N2 = O(N2)
Time (steps)
3N = O(N)
Input (size)
N = 60
24
BIG-OH NOTATION
Simple Rule:
Drop lower order terms and constant factors
7n-3 is O(n)
8n2log n + 5n2 + n is O(n2log n)
25
BIG OMEGA NOTATION
If we wanted to say “running time is at least…” we use Ω
If f(n) and g(n) are two complexity functions then we can say:
f(n) is Ω(g(n)) if there exist positive numbers c and n0 such that 0<=f(n)>=cΩ (n) for all n>=n0
26
BIG THETA NOTATION
If we wish to express tight bounds we use the theta notation, Θ
27
WHAT DOES THIS ALL MEAN?
If f(n) = Θ(g(n)) we say that f(n) and g(n) grow at the same
rate, asymptotically
28
WHICH NOTATION DO WE USE?
To express the efficiency of our algorithms which of the
three notations should we use?
Why?
log n Logarithmic: when n increases, so does run time, but much slower. Common in programs which
solve large problems by transforming them into smaller problems. Exp : binary Search
n Linear: run time varies directly with n. Typically, a small amount of processing is done on each
element. Exp: Linear Search
n log n When n doubles, run time slightly more than doubles. Common in programs which break a problem
down into smaller sub-problems, solves them independently, then combines solutions. Exp: Merge
n2 Quadratic: when n doubles, runtime increases fourfold. Practical only for small problems; typically
the program processes all pairs of input (e.g. in a double nested loop). Exp: Insertion Search
2n Exponential: when n doubles, run time squares. This is often the result of a natural, “brute force”
solution. Exp: Brute Force.
Note: logn, n, nlogn, n2>> less Input>>Polynomial
n3, 2n>>high input>> non polynomial
30
SIZE DOES MATTER[1]
N log2N 5N N log2N N2 2N
8 3 40 24 64 256
16 4 80 64 256 65536
32 5 160 160 1024 ~109
64 6 320 384 4096 ~1019
128 7 640 896 16384 ~1038
256 8 1280 2048 65536 ~1076
31
COMPLEXITY CLASSES
Time (steps)
32
SIZE DOES MATTER[2]
Suppose a program has run time O(n!) and the run time for
n = 10 is 1 second
Analyzing Loops
Arithmetic operations:
x = 5 * y + 4 - z;
Array referencing:
A[j] = 5;
Array assignment:
j, A[j] = 5;
35
Most conditional tests:
if (x < 12) ...
ANALYZING LOOPS[1]
Any loop has two parts:
How many iterations are performed?
How many steps per iteration?
int j,k;
for (j=0; j<N; j++)
for (k=N; k>0; k--)
sum += k+j;
int j,k;
for (j=0; j < N; j++)
for (k=0; k < j; k++)
sum += k+j;
Gauss figured out that the sum of the first n numbers is always:
41
SEQUENCE OF STATEMENTS
For a sequence of statements, compute their complexity
functions individually and add them up
42
CONDITIONAL STATEMENTS
What about conditional statements such as
if (condition)
statement1;
else
statement2;
However if N ≥ 2, then running time T(N) is the cost of each step taken plus time
required to compute power(x,n-1). (i.e. T(N) = 2+T(N-1) for N ≥ 2)
44
46
REFERENCES
47