Stat Notes
Stat Notes
1. Probability Mass Function (PMF) – A function that gives the probability of each
possible value of a discrete RV. It satisfies:
2. Cumulative Distribution Function (CDF) – A function that gives the probability that
the random variable is less than or equal to a certain value:
3. Expected Value (Mean) – The average value of the random variable (see formula on
your notes)
4. Variance and Standard Deviation – Measures the spread of the value of random
variables (see formula on your notes)
These concepts are fundamental in probability theory and statistical analysis, providing a
framework for modeling real-world discrete random processes.
Characteristics
1. Symmetric Shape – The distribution is perfectly symmetric around the mean (μ\muμ).
2. Mean, Median, and Mode are Equal – The highest point of the curve is at the mean.
3. Defined by Two Parameters:
o Mean (μ) – Determines the center of the distribution.
o Standard Deviation (σ) – Determines the spread or width of the distribution.
4. Total Probability Equals 1 – The area under the curve sums to 1.
5. Follows the Empirical Rule (68-95-99.7 Rule):
Applications
• Statistics & Hypothesis Testing (e.g., Z-tests, confidence intervals)
• Quality Control (e.g., process variation analysis)
• Natural and Social Sciences (e.g., IQ scores, financial models)
The normal distribution is fundamental in probability and statistics, forming the basis for
many inferential statistical methods.
Key Properties
1. Mean of the Sampling Distribution
The mean of the sampling distribution is equal to the population mean.
The standard deviation of the sampling distribution is called the standard error (SE)
- Regardless of the population's shape, if the sample size n is large (n≥30), the sampling
distribution of the sample mean will be approximately normal.
- If the population itself is normally distributed, then the sampling distribution of the
mean is also normal for any sample size.
Estimation of parameters is the process of using sample data to determine the unknown
values of population parameters, such as the mean (μ\muμ) or standard deviation (σ\sigmaσ).
It is a key concept in inferential statistics.
Types of Estimation
1. Point Estimation – Provides a single value as an estimate of a population parameter.
2. Interval Estimation – Provides a range of values (confidence interval) that is likely to
contain the population parameter.
- Example: A 95% confidence interval for μ means we are 95% confident that the true
population mean lies within the interval.
• Unbiasedness – The expected value of the estimator equals the true parameter value.
• Consistency – The estimator becomes more accurate as sample size increases.
• Efficiency – The estimator has the smallest variance among all unbiased estimators.
• Sufficiency – The estimator uses all available information from the sample.