EDU 802 SATI
EDU 802 SATI
EDU 802 SATI
Question 1.
State reasons why statistics is important in education.
The reasons are;
1. Data-Driven Decision Making:
Statistics enables educators to make informed decisions based on data analysis (Lemann, 2013).
By analyzing data, educators can identify trends, patterns, and correlations that inform
instruction, assessment, and policy decisions.
2. Assessment and Evaluation:
Statistics is crucial for assessing and evaluating student learning outcomes (Baker, 2013).
Statistical methods help educators measure student progress, identify areas of improvement, and
evaluate the effectiveness of instructional strategies.
3. Research and Innovation:
Statistics is essential for conducting educational research and driving innovation (Creswell,
2014). Statistical analysis helps researchers identify relationships between variables, test
hypotheses, and develop evidence-based solutions to educational problems.
4. Accountability and Accreditation:
Statistics plays a critical role in ensuring accountability and accreditation in education (U.S.
Department of Education, 2019). Statistical data helps institutions demonstrate compliance with
accreditation standards, measure student outcomes, and evaluate program effectiveness.
5. Informing Policy Decisions:
Statistics informs policy decisions at the local, state, and national levels (National Center for
Education Statistics, 2020). Statistical data helps policymakers understand educational trends,
identify areas of need, and develop evidence-based policies that promote educational equity and
excellence.
6. Enhancing Teaching and Learning:
Statistics helps educators enhance teaching and learning by identifying effective instructional
strategies (Hattie, 2009). By analyzing data, educators can refine their teaching practices,
improve student outcomes, and promote academic achievement.
2
Inferential Statistics:
Inferential statistics involves making conclusions or predictions about a population based on a
sample of data. It helps to answer questions about the population, such as estimating population
parameters or testing hypotheses.
Examples:
1. Conducting a survey of 100 students to estimate the average height of all students in the
school (Cochran, 1977).
2. Testing the hypothesis that there is a significant difference in the average scores of students
who received different teaching methods (Kirk, 2013).
3. Using regression analysis to predict the relationship between study time and exam scores
(Field, 2018).
Key differences:
1. Purpose: Descriptive statistics aims to summarize and describe data, while inferential statistics
aims to make conclusions or predictions about a population (Moore & McCabe, 2017).
2. Scope: Descriptive statistics focuses on the sample data, while inferential statistics focuses on
the population (Cochran, 1977).
3. Methods: Descriptive statistics uses methods such as mean, median, mode, and standard
deviation, while inferential statistics uses methods such as hypothesis testing, confidence
intervals, and regression analysis (Field, 2018).
ii. Parametric and Non-parametric statistics:
Parametric Tests:
Parametric tests are statistical tests that assume a specific distribution (usually normal
distribution) for the data. They require numerical data and are used to test hypotheses about
population parameters.
Examples:
1. t-test: Used to compare the means of two groups (e.g., comparing the average scores of
students who received different teaching methods) (Kirk, 2013).
2. ANOVA: Used to compare the means of three or more groups (e.g., comparing the average
scores of students from different schools) (Field, 2018).
4
3. Regression analysis: Used to model the relationship between a dependent variable and one or
more independent variables (e.g., predicting student scores based on study time and attendance)
(Cohen, 2013).
Non-Parametric Tests:
Non-parametric tests are statistical tests that don't assume a specific distribution for the data.
They are used when the data doesn't meet the assumptions of parametric tests or when the data is
ordinal or categorical.
Examples:
1. Wilcoxon rank-sum test: Used to compare the distributions of two groups (e.g., comparing the
scores of students who received different teaching methods) (Hollander, 2013).
2. Kruskal-Wallis test: Used to compare the distributions of three or more groups (e.g.,
comparing the scores of students from different schools) (Siegel, 1956).
3. Spearman's rank correlation coefficient: Used to measure the correlation between two ordinal
variables (e.g., correlating student scores with their attendance) (Corder, 2014).
Key differences:
1. Assumptions: Parametric tests assume a specific distribution (usually normal distribution) for
the data, while non-parametric tests don't assume any specific distribution (Kirk, 2013).
2. Data type: Parametric tests require numerical data, while non-parametric tests can be used
with ordinal or categorical data (Field, 2018).
3. Purpose: Parametric tests are used to test hypotheses about population parameters, while non-
parametric tests are used to compare distributions or correlations (Siegel, 1956).
Question 1c
Explain the conditions that are required for each of the following statistical techniques to
be applied for research data analysis.
I. Chi-Square test is a statistical technique used to determine whether there is a significant
association between two categorical variables. The following conditions must be met to use the
Chi-Square test:
1. Categorical data: The data must be categorical, meaning it can be divided into distinct
categories or groups.
2. Independent observations: Each observation must be independent of the others.
3. Random sampling: The sample must be randomly selected from the population.
5
1. Expected frequencies: The expected frequency count for each category must be at least 5.
2. No significant outliers: There should be no significant outliers in the data.
3. Two or more categories: There must be two or more categories for each variable.
4. Nominal or ordinal data: The data can be nominal (e.g., gender, color) or ordinal (e.g.,
education level, income level).
When to use Chi-Square test:
1. Testing for independence: To determine whether there is a significant association between two
categorical variables.
2. Testing for homogeneity: To determine whether the proportions of categories are equal across
different groups.
3. Testing for goodness of fit: To determine whether the observed frequencies differ significantly
from the expected frequencies.
ii. t-test is a statistical technique used to determine whether there is a significant difference
between the means of two groups.
The following conditions must be met to use the t-test:
Assumptions
1. Normality: The data should be normally distributed. This can be checked using histograms, Q-
Q plots, or normality tests such as the Shapiro-Wilk test.
2. Independence: Each observation should be independent of the others.
3. Random sampling: The sample should be randomly selected from the population.
4. Equal variances: The variances of the two groups should be equal. This can be checked using
Levene's test or F-test.
Types of t-tests
1. Independent samples t-test: Used to compare the means of two independent groups.
2. Paired samples t-test: Used to compare the means of two related groups, such as before-and-
after observations.
When to use t-test
1. Comparing means: To determine whether there is a significant difference between the means
of two groups.
1. Testing hypotheses: To test hypotheses about the population mean or the difference between
two population means.
6
iii. z-test is a statistical technique used to determine whether a sample mean is significantly
different from a known population mean. The following conditions must be met to use the z-test:
Assumptions;
1. Normality: The data should be normally distributed. This can be checked using histograms, Q-
Q plots, or normality tests such as the Shapiro-Wilk test.
2. Independence: Each observation should be independent of the others.
3. Random sampling: The sample should be randomly selected from the population.
4. Large sample size: The sample size should be large enough to ensure that the sampling
distribution of the mean is approximately normal. A common rule of thumb is to use a sample
size of at least 30.
Conditions for z-test:
1. Known population mean: The population mean must be known.
2. Known population standard deviation: The population standard deviation must be known.
3. Sample mean: The sample mean must be calculated.
4. Sample size: The sample size must be known.
When to use z-test
1. Testing hypotheses: To test hypotheses about the population mean.
2. Comparing means: To compare the sample mean to a known population mean.
3. Quality control: To determine whether a sample mean is within a specified range.
iv. Bivariate correlation is a statistical technique used to measure the strength and direction of
the linear relationship between two continuous variables. The following conditions must be met
to use bivariate correlation:
Assumptions:
1. Linearity: The relationship between the two variables should be linear.
2. Independence: Each observation should be independent of the others.
3. Normality: The data should be normally distributed.
4. No multicollinearity*: The two variables should not be highly correlated with each other.
Conditions for Bivariate Correlation:
1. Two continuous variables: The two variables should be continuous, such as height, weight, or
IQ scores.
2. Same units of measurement: The two variables should have the same units of measurement.
7
2. Predicting outcomes: To predict the value of the dependent variable based on the value of the
independent variable.
3. Identifying relationships: To identify the strength and direction of the relationship between the
independent variable and the dependent variable.
Types of Simple Regression Analysis;
1. Simple linear regression: Used when the independent variable is continuous.
2. Simple logistic regression: Used when the dependent variable is binary.
vi. ANOVA (Analysis of Variance) is a statistical technique used to compare the means of three
or more groups to determine if there is a significant difference between them. The following
conditions must be met to use ANOVA:
Assumptions:
1. Normality: The data should be normally distributed in each group.
2. Independence: Each observation should be independent of the others.
3. Homogeneity of variance: The variance of the data should be equal across all groups.
4. Random sampling: The sample should be randomly selected from the population.
Conditions for ANOVA:
1. Three or more groups: ANOVA requires at least three groups to compare.
2. Continuous data: The dependent variable should be continuous.
3. Categorical independent variable: The independent variable should be categorical, with three
or more levels.
4. Equal sample sizes: While not necessary, equal sample sizes can improve the accuracy of the
results.
When to Use ANOVA:
1. Comparing means: To compare the means of three or more groups.
2. Testing hypotheses: To test hypotheses about the differences between group means.
3. Identifying significant differences: To identify which groups are significantly different from
each other.
Types of ANOVA:
1. One-way ANOVA: Used to compare the means of three or more groups.
2. Two-way ANOVA: Used to compare the means of two or more independent variables.
3. Repeated measures ANOVA: Used to compare the means of three or more related groups.
9
REFERENCES
Baker, E. L. (2013). The role of statistics in education. Journal of Educational and Behavioural
Statistics, 38(1), 5-13.
Bureau of Labor Statistics. (2020). Data-driven careers.
Cochran, W. G. (1977). Sampling techniques. John Wiley & Sons.
Cohen, J. (2013). Statistical power analysis for the behavioural sciences. Routledge.
Conover, W. J. (1999). Practical nonparametric statistics. John Wiley & Sons.
Corder, G. W. (2014). Nonparametric statistics: A step-by-step approach. Wiley.
Creswell, J. W. (2014). Research design: Qualitative, quantitative, and mixed methods
approaches.
Deming, W. E. (1986). Out of the crisis.
Field, A. (2018). Discovering statistics using IBM SPSS statistics. Sage Publications.
Hattie, J. (2009). Visible learning: A synthesis of over 800 meta-analyses relating to
achievement.
Hollander, M. (2013). Nonparametric statistical methods. Wiley.
Kirk, R. E. (2013). Experimental design: Procedures for the behavioural sciences. Sage
Publications.
Lemann, N. (2013). The big test: The secret history of the American meritocracy.
Moore, D. S., & McCabe, G. P. (2017). Introduction to the practice of statistics. W.H. Freeman
and Company.
National Centre for Education Statistics. (2020). Digest of education statistics.
National Science Foundation. (2019). Science and engineering indicators.
Paul, R. (2013). Critical thinking: What every person needs to survive in a rapidly changing
world.
Siegel, S. (1956). Nonparametric statistics for the behavioural sciences. McGraw-Hill.
U.S. Department of Education. (2019). Every Student Succeeds Act (ESSA).