BPCC 108 Notes
BPCC 108 Notes
Parametric statistics refers to a set of statistical techniques that assume the underlying data
follows a specific distribution, typically a normal distribution. These methods are used to make
inferences about population parameters based on sample statistics. Common parametric tests
include t-tests, ANOVA (Analysis of Variance), and linear regression.
Assumptions of Parametric Statistics:
1. Normality: The data should be approximately normally distributed. This assumption is
particularly important for small sample sizes, as larger samples (typically n > 30) can often
satisfy this assumption due to the Central Limit Theorem.
2. Homogeneity of Variance: The variances among the groups being compared should be
approximately equal. This is crucial for tests like ANOVA, where the assumption is that the
variability within each group is similar.
3. Independence: The observations should be independent of one another. This means that the
data collected from one participant or observation should not influence or be related to the
data collected from another.
4. Linearity: For certain parametric tests, such as linear regression, there should be a linear
relationship between the independent and dependent variables.
5. Interval or Ratio Scale: The data should be measured on an interval or ratio scale, which
means that the data should have meaningful numerical values that allow for the calculation
of means and standard deviations.
When these assumptions are met, parametric statistical methods can provide more powerful and
reliable results. However, if the assumptions are violated, non-parametric methods may be more
appropriate, as they do not rely on these specific distributional assumptions.
Non-parametric statistics refers to a set of statistical methods that do not assume a specific
distribution for the data. These methods are particularly useful when the assumptions required
for parametric tests (such as normality and homogeneity of variance) are not met. Non-
parametric tests are often used for ordinal data or when the sample size is small.
Assumptions of Non-Parametric Statistics:
1. No Assumption of Normality: Non-parametric tests do not require the data to follow a
normal distribution. This makes them suitable for data that is skewed or has outliers.
2. Ordinal or Nominal Data: Non-parametric methods can be used with ordinal data (data that
can be ranked but not measured) or nominal data (categorical data without a specific order).
This flexibility allows for a wider range of data types.
3. Independence: Similar to parametric tests, non-parametric tests assume that the
observations are independent of one another. Each observation should not influence or be
related to another.
4. Homogeneity of Variance (in some cases): While many non-parametric tests do not require
equal variances, some tests may still assume that the groups being compared have similar
variances, although this is less stringent than in parametric tests.
5. Random Sampling: The data should be collected through a process of random sampling to
ensure that the sample is representative of the population.
Non-parametric statistics are valuable tools in research, especially when dealing with non-
normal data or when the sample size is too small to reliably assess the distribution. Common
non-parametric tests include the Mann-Whitney U test, Wilcoxon signed-rank test, Kruskal-
Wallis test, and Chi-square test.
NOTES
Applications and Uses of Parametric Statistics
1. Experimental Research: Parametric statistics are commonly used in experimental
research where the data is expected to follow a normal distribution. For example, in
clinical trials, researchers often use t-tests or ANOVA to compare the effects of
different treatments on health outcomes.
2. Psychology and Social Sciences: In fields like psychology, parametric tests are
frequently employed to analyze data from surveys and experiments. For instance,
researchers might use regression analysis to understand the relationship between
variables such as stress and academic performance.
3. Quality Control: In manufacturing and quality control processes, parametric methods
can be used to analyze measurements and ensure that products meet specified
standards. Techniques like control charts often rely on normality assumptions.
4. Market Research: Businesses use parametric statistics to analyze consumer preferences
and behaviors. For example, they might use t-tests to compare the average
satisfaction ratings of two different products.
5. Finance: In finance, parametric methods are used to model and predict stock prices,
returns, and risks. Techniques such as linear regression are commonly applied to
understand relationships between financial variables.
Applications and Uses of Non-Parametric Statistics
1. Ordinal Data Analysis: Non-parametric statistics are particularly useful for analyzing
ordinal data, such as survey responses on a Likert scale (e.g., strongly agree to
strongly disagree). Tests like the Wilcoxon signed-rank test can be used to compare
medians.
2. Small Sample Sizes: When dealing with small sample sizes where normality cannot be
assumed, non-parametric tests provide a robust alternative. For example, the Mann-
Whitney U test can be used to compare two independent groups without assuming
normality.
3. Non-Normal Distributions: Non-parametric methods are ideal for data that is skewed
or has outliers. For instance, the Kruskal-Wallis test can be used to compare more
than two groups when the data does not meet the assumptions of ANOVA.
4. Categorical Data: Non-parametric statistics are often used for analyzing categorical
data. The Chi-square test is a common non-parametric method used to assess the
association between two categorical variables.
5. Environmental Studies: In environmental research, non-parametric methods are
frequently applied to analyze data that may not meet the assumptions of parametric
tests, such as species abundance data or pollution levels.
In summary, both parametric and non-parametric statistics have their specific
applications and are chosen based on the nature of the data, the research question, and
the underlying assumptions. Parametric statistics are powerful when assumptions are
met, while non-parametric statistics provide flexibility and robustness in various
situations.
NOTES
Differences between parametric and non-parametric statistics:
1. Assumptions
Parametric Statistics: These methods assume that the data follows a specific
distribution, typically a normal distribution. They also assume homogeneity of
variance and that the data is measured on an interval or ratio scale.
Non-Parametric Statistics: These methods do not assume a specific distribution for the
data. They can be used with ordinal or nominal data and are more flexible regarding
the underlying data distribution.
2. Data Type
Parametric Statistics: Suitable for interval or ratio data, where the data can be
meaningfully averaged and has a consistent scale.
Non-Parametric Statistics: Can be used for ordinal data (ranked data) and nominal
data (categorical data), making them applicable in a wider range of situations.
3. Statistical Power
Parametric Statistics: Generally more powerful than non-parametric tests when the
assumptions are met. This means they are more likely to detect a true effect when one
exists.
Non-Parametric Statistics: Typically less powerful than parametric tests, especially
when the sample size is small. However, they are robust in situations where
parametric assumptions are violated.
4. Examples of Tests
Parametric Statistics: Common tests include t-tests (for comparing means), ANOVA
(for comparing means across multiple groups), and linear regression.
Non-Parametric Statistics: Common tests include the Mann-Whitney U test (for
comparing two independent groups), Wilcoxon signed-rank test (for comparing two
related groups), Kruskal-Wallis test (for comparing more than two groups), and Chi-
square test (for categorical data).
5. Interpretation of Results
Parametric Statistics: Results are often interpreted in terms of population parameters
(e.g., means, variances) and can provide estimates of effect sizes.
Non-Parametric Statistics: Results are typically interpreted in terms of ranks or
medians rather than means, and they may not provide estimates of population
parameters.
6. Use Cases
Parametric Statistics: Preferred when the data meets the necessary assumptions, such
as in controlled experiments with normally distributed data.
Non-Parametric Statistics: Preferred when dealing with non-normal data, small
sample sizes, or when the data is ordinal or nominal.
In summary, the choice between parametric and non-parametric statistics depends on the
nature of the data, the research question, and whether the assumptions of parametric
tests are met.
NOTES
Explain the steps in computation of one sample median test with the help of a
suitable example.
The one-sample median test is a non-parametric statistical test used to determine whether
the median of a single sample differs from a specified value (often the population median).
Here are the steps involved in computing the one-sample median test, along with a suitable
example.
Steps in Computation of One-Sample Median Test
Step 1: State the Hypotheses
Null Hypothesis (H0): The median of the population is equal to a specified value (e.g.,
m0m0).
Alternative Hypothesis (H1): The median of the population is not equal to the specified
value.
Example:
Suppose we want to test whether the median height of a group of students differs from 170
cm.
H0: The median height is 170 cm.
H1: The median height is not 170 cm.
Step 2: Collect the Sample Data
Gather the sample data that you will analyze.
Example:
Let's say we have the following heights (in cm) of 10 students:
165,172,168,175,160,180,170,169,173,177165,172,168,175,160,180,170,169,173,177
Step 3: Determine the Sample Size
Count the number of observations in the sample.
Sample Size (n): In this case, n=10n=10.
Step 4: Calculate the Median of the Sample
Sort the data and find the median.
Sorted heights:
160,165,168,169,170,172,173,175,177,180160,165,168,169,170,172,173,175,177,180
Median: Since nn is even, the median is the average of the two middle values:
Median=170+1692=169.5Median=2170+169=169.5
Step 5: Count the Number of Observations Above and Below the Specified Median
Count how many observations are above and below the specified median (170 cm).
Above 170 cm: 172,175,180,173,177172,175,180,173,177 (5 observations)
Below 170 cm: 165,168,169,160165,168,169,160 (4 observations)
Equal to 170 cm: 170170 (1 observation)
Step 6: Calculate the Test Statistic
The test statistic is based on the counts of observations above and below the specified median.
Let n1n1be the number of observations above the median (5).
Let n2n2be the number of observations below the median (4).
The test statistic TT is the smaller of n1n1and n2n2: T=min(n1,n2)=min(5,4)=4T=min(n1,n2
)=min(5,4)=4
Step 7: Determine the Critical Value
Using a binomial distribution, determine the critical value for the test statistic based on the
sample size and significance level (e.gThe one-sample median test is a non-parametric
statistical test used to determine whether the median of a single sample differs from a
specified value (often a hypothesized population median). Here are the steps involved in
computing the one-sample median test, along with a suitable example.