Probability Theory
Probability Theory
SUBMITTED BY-
AKASH MODI
DEPARTMENT OF STATISTICS
UNIVERSITY OF LUCKNOW
LUCKNOW, UTTAR PRADESH -226007
CERTIFICATE
PROJECT ACKNOWLEDGEMENT
Akash Modi
CONTENT
ABSTRACT
INTRODUCTION
OBJECTIVES
KEY THEORETICAL CONCEPTS
SAMPLE SPACE
RANDOM EXPERIMENT
EVENT
CONDITIONAL PROBABILITY
BAYES’ THEOREM
CLASSICAL APPROACH
PROBABILITY MASS FUNCTION
PROBABILITY DENSITY FUNCTION
MOMENT GENERATING FUNCTION
CENTRAL LIMIT THEOREM
CONCLUSION
REFERENCE
ABSTRACT
Probability theory forms the foundation of understanding randomness and uncertainty in
various fields such as mathematics, statistics, science, and economics.
This paper explores the fundamental concepts of probability theory, including axioms,
random variables, probability distributions, and key theorems such as the Law of Large
Numbers and the Central Limit Theorem.
Additionally, it examines real-world applications, including risk analysis, machine learning,
and decision-making processes.
Through theoretical explanations and practical examples, the paper aims to highlight the
critical role of probability theory in interpreting data and making informed predictions in
uncertain environment
INTRODUCTION
The theory of probability had its origin in gambling and games of chance. It owes much to
the curiosity of gamblers who pestered their friends in the mathematical world with all sorts
of questions. Unfortunately, this association with gambling contributed to very slow and
sporadic growth of probability theory as a mathematical discipline. The mathematicians of
the day took little or no interest in the development of any theory but looked only at the
combinatorial reasoning involved in each problem.
An extension of the classical definition of Laplace was used to evaluate the probabilities of
sets of events with infinite outcomes. The notion of equal likelihood of certain events played
a key role in this development. According to this extension, if Ω is some region with a well-
defined measure (length, area, volume, etc.), the probability that a point chosen at random
lies in a subregion A of Ω is the ratio measure(A)/measure(Ω). Many problems of geometric
probability were solved using this extension.
The mathematical theory of probability as we know it today is of comparatively recent origin.
It was A. N. Kolmogorov who axiomatized probability in his funda mental work, Foundations
of the Theory of Probability (Berlin), in 1933.
3.EVENT
An event is a subset of the possible outcomes of a random experiment. It is a specific result
or a combination of results that we are interested in.
Examples:
In tossing a coin, getting Head is an event.
In rolling a die, getting a number greater than 4 (i.e., outcomes 5 and 6) is an event.
4.CONDITIONAL PROBABILITY
Conditional probability is the probability of an event occurring given that another event has
already occurred. If A and B are two events, the probability of A given B is denoted as P(A/B)
and
is calculated using the formula:
P(A|B) = P(A∩B) / P(B), where P(B)≠0
Example:
Consider a deck of cards. Let:
A: Event of drawing a king
B: Event of drawing a red card.
If you draw a card and it is known to be red(B), the conditional probability P(A|B) is:
P(A|B) = P(A∩B) / P(B) = 2/26 = 1/13
Here:
P(A∩B) is the probability of drawing a red king (2 cards).
P(B) is the probability of drawing any red card (26) cards.
5.BAYES’ THEOREM
Bayes' Theorem is a fundamental concept in probability theory and statistics that describes
the relationship between conditional probabilities. It allows you to update the probability of
an event based on new evidence or information.
The formula for Bayes' Theorem is:
P(A∣B) = P(B∣A)⋅P(A)/P(B)
Where:
P(A∣B) is the posterior probability — the probability of event AA happening given
that event B has occurred.
P(B∣A) is the likelihood — the probability of event B happening given that
event A has occurred.
P(A) is the prior probability — the initial probability of event A before considering
the new evidence B.
P(B) is the marginal probability — the total probability of event B occurring,
regardless of whether A happens or not.
EXAMPLE:
Suppose a doctor wants to determine the probability of a patient having a disease based on
the results of a diagnostic test. Let's say:
PROBABILITY AXIOMS
Let (Ω, S) be the sample space associated with a statistical experiment. In this section we
define a probability set function and study some of its properties.
Definition 1. Let (Ω, S) be a sample space. A set function P defined on S is called a
probability measure (or simply, probability) if it satisfies the following conditions:
(i) P(A) ≥0 for all A∈S.
(ii) P(Ω) = 1
(iii) If A∩B = φ,
Then P(A∪B) = P(A) + P(B)
1. 1. Expectation (or Expected Value)
The expectation (or expected value) of a random variable is a measure of its central
location—in other words, it provides a weighted average of all possible values that the
random variable can take, weighted by their probabilities. It represents the "long-run average"
of the outcomes if the experiment is repeated many times.
For a Discrete Random Variable:
If XX is a discrete random variable with possible values x1,x2,…,xnx1,x2,…,xn and
associated probabilities P(X=xi)P(X=xi), the expectation is:
E[X]=∑i=1nxi⋅P(X=xi).E[X]=i=1∑nxi⋅P(X=xi).
For a Continuous Random Variable:
If XX is a continuous random variable with probability density function (PDF) fX(x)fX(x), the
expectation is:
E[X]=∫−∞∞x⋅fX(x) dx.E[X]=∫−∞∞x⋅fX(x)dx.
The expectation is a real number and represents the "center" of the distribution of XX.
Properties of Expectation:
1. Linearity: The expectation is linear, which means:
E[aX+b]=aE[X]+b,E[aX+b]=aE[X]+b,
where aa and bb are constants. This property also extends to sums of random variables:
E[X+Y]=E[X]+E[Y]if X and Y are random variables.E[X+Y]=E[X]+E[Y]if X and Y are rando
m variables.
2. Constant: If cc is a constant, then:
E[c]=c.E[c]=c.
2. Variance
The variance of a random variable measures the spread or dispersion of its values around
the expectation. It indicates how much the values of the random variable deviate from the
mean. A high variance means that the values are spread out widely around the mean, while a
low variance indicates that the values are concentrated near the mean.
Formula for Variance:
The variance of a random variable XX, denoted by Var(X)Var(X), is defined
as the expected squared deviation from the
mean:Var(X)=E[(X−E[X])2].Var(X)=E[(X−E[X])2].Expanding this
expression:Var(X)=E[X2]−(E[X])2.Var(X)=E[X2]−(E[X])2.
For a Discrete Random Variable:
If XX is a discrete random variable with values x1,x2,…,xnx1,x2,…,xn and
probabilities P(X=xi)P(X=xi), the variance is:
Var(X)=∑i=1n(xi−E[X])2⋅P(X=xi).Var(X)=i=1∑n(xi−E[X])2⋅P(X=xi).
For a Continuous Random Variable:
If XX is a continuous random variable with PDF fX(x)fX(x), the variance is:
Var(X)=∫−∞∞(x−E[X])2⋅fX(x) dx.Var(X)=∫−∞∞(x−E[X])2⋅fX(x)dx.
Alternatively, the variance can be computed as:
Var(X)=E[X2]−(E[X])2.Var(X)=E[X2]−(E[X])2.
Properties of Variance:
4. Non-negative: Variance is always non-negative:
Var(X)≥0.Var(X)≥0.
If XX is a constant random variable, Var(X)=0Var(X)=0.
5. Linearity: Variance behaves differently under scaling and addition:
If Y=aX+bY=aX+b, where aa and bb are
constants:Var(Y)=a2⋅Var(X).Var(Y)=a2⋅Var(X).The constant bb does
not affect the variance because it does not change the spread of the
distribution, only the location.
For the sum of two independent random
variables XX and YY:Var(X+Y)=Var(X)+Var(Y).Var(X+Y)=Var(X)
+Var(Y).If XX and YY are dependent, you need to add the covariance
term:Var(X+Y)=Var(X)+Var(Y)+2Cov(X,Y).Var(X+Y)=Var(X)+Var(Y)
+2Cov(X,Y).
Example:
Consider a discrete random variable XX with the following probability distribution:
XX 1 2 3 4
Conclusion
Expectation is the mean or average of a random variable, providing a measure
of the "center" of the distribution.
Variance measures the spread or variability of the random variable around the
mean, with higher variance indicating greater spread.