Reseach 04
Reseach 04
Let's break down the key concepts related to Measurement Scales, Tabulation (Coding
Sheet), Data Analysis, Correlation & Regression, and Parametric & Non-Parametric Tests.
1. Measurement Scales
Measurement scales define how variables are measured and categorized. They determine
the types of analyses that can be performed.
Nominal Scale: This is the most basic scale of measurement. It categorizes data
without a specific order or hierarchy.
o Example: Gender (Male, Female), Eye color (Blue, Green, Brown).
Ordinal Scale: This scale ranks data in order but the intervals between the ranks are
not necessarily equal.
o Example: Educational level (High School, College, Postgraduate), Survey
ratings (Strongly Disagree, Disagree, Neutral, Agree, Strongly Agree).
Interval Scale: Data is ordered, and the intervals between values are equal. However,
there is no true zero.
o Example: Temperature in Celsius or Fahrenheit.
Ratio Scale: This is the most informative scale, which has a true zero and equal
intervals.
o Example: Height, Weight, Age, Income.
2. Tabulation (Coding Sheet)
Tabulation involves organizing data into tables for easier analysis. A coding sheet is used to
record responses and systematically organize raw data, particularly for survey or
questionnaire results.
A coding sheet typically includes:
o Variables: The data points to be collected (e.g., age, gender, income).
o Codes: Numeric or categorical codes used to represent responses (e.g., "1"
for male, "2" for female).
o Respondent IDs: Unique identifiers for each participant or observation.
Here’s an example layout of a basic coding sheet:
001 25 1 40
002 30 2 55
Respondent ID Age Gender (1=Male, 2=Female) Income (in thousands)
003 22 1 30
3. Analysis of Data
Data analysis is the process of inspecting, cleaning, and transforming data to extract
meaningful insights. Here's a brief overview:
Descriptive Analysis: Summarizes the data using measures like:
o Central Tendency: Mean, Median, Mode.
o Dispersion: Range, Variance, Standard Deviation.
Inferential Analysis: Makes inferences or predictions about a population based on a
sample.
o Hypothesis Testing: Tests assumptions using various statistical tests.
4. Correlation & Regression
Both are statistical techniques used to understand relationships between variables.
Correlation: Measures the strength and direction of the relationship between two or
more variables.
o Pearson Correlation is used for linear relationships between continuous
variables.
o Spearman’s Rank Correlation is used for ordinal or non-linear data.
Regression: Explores the relationship between a dependent variable and one or
more independent variables.
o Linear Regression: Predicts the value of a dependent variable based on the
linear relationship with one or more independent variables.
o Multiple Regression: Examines the relationship between one dependent
variable and several independent variables.
5. Parametric & Non-Parametric Tests
These tests are used to analyze data based on certain assumptions.
Parametric Tests: Assume that the data follows a specific distribution (usually
normal). These tests are used when data is continuous and normally distributed.
o Examples:
t-Test: Compares means of two groups (independent or paired).
ANOVA (Analysis of Variance): Compares means of more than two
groups.
Pearson Correlation: Measures the relationship between two
continuous variables.
Non-Parametric Tests: These tests do not assume a specific distribution. They are
often used when the data is not normally distributed or when dealing with ordinal
data.
o Examples:
Chi-Square Test: Tests the association between categorical variables.
Mann-Whitney U Test: Compares two independent groups when the
dependent variable is ordinal or not normally distributed.
Kruskal-Wallis H Test: Compares more than two independent groups
for ordinal or non-normally distributed data.
Summary Table
Tabulation (Coding Organizing raw data into a structured Table with codes for responses
Sheet) format for analysis. and variables.
Key Takeaways
Type-I Error: Occurs when we wrongly reject the null hypothesis (false positive).
Type-II Error: Occurs when we wrongly fail to reject the null hypothesis (false
negative).
Significance Level (α): The threshold for Type-I error.
Power (1 - β): The probability of correctly rejecting a false null hypothesis.
Sample Size: Larger sample sizes help reduce both Type-I and Type-II errors by
providing more reliable estimates.
Hypothesis Testing for Means and Proportions for Small and Large Samples
Hypothesis testing allows researchers to make inferences about population parameters
(such as the population mean or proportion) based on sample data. The process for testing
hypotheses varies depending on whether you are dealing with means or proportions and
whether the sample size is small or large.
We will cover:
1. Hypothesis Testing for Means
o Small samples (using t-tests)
o Large samples (using z-tests)
2. Hypothesis Testing for Proportions
o Small and large samples (using z-tests)
1. Hypothesis Testing for Means
a. Testing the Mean for Small Samples (n < 30)
When the sample size is small (less than 30), we use the t-test for hypothesis testing, as the
sample mean's distribution may not be normal. The t-distribution is used because it
accounts for the increased variability in small samples.
Formula for t-test:
t=Xˉ−μ0s/nt = \frac{\bar{X} - \mu_0}{s / \sqrt{n}}
Where:
o Xˉ\bar{X} = sample mean
o μ0\mu_0 = hypothesized population mean
o ss = sample standard deviation
o nn = sample size
Steps for Conducting a t-test:
1. State the hypotheses:
Null Hypothesis (H₀): The population mean is equal to the
hypothesized value (μ=μ0\mu = \mu_0).
Alternative Hypothesis (H₁): The population mean is not equal to the
hypothesized value (μ≠μ0\mu \neq \mu_0) or is greater than or less
than the hypothesized value (μ>μ0\mu > \mu_0 or μ<μ0\mu < \
mu_0).
2. Select the significance level (α\alpha), commonly 0.05.
3. Calculate the t-statistic using the formula above.
4. Find the critical t-value from the t-distribution table based on the degrees of
freedom (df=n−1df = n - 1) and the significance level (α\alpha).
5. Make a decision:
If the calculated t-value exceeds the critical t-value (for two-tailed
tests), reject the null hypothesis.
If the p-value is less than α\alpha, reject the null hypothesis.
b. Testing the Mean for Large Samples (n ≥ 30)
When the sample size is large (greater than or equal to 30), the Central Limit Theorem (CLT)
suggests that the sampling distribution of the sample mean is approximately normal, even if
the population distribution is not normal. For large samples, we typically use the z-test.
Formula for z-test:
z=Xˉ−μ0σ/nz = \frac{\bar{X} - \mu_0}{\sigma / \sqrt{n}}
Where:
o Xˉ\bar{X} = sample mean
o μ0\mu_0 = hypothesized population mean
o σ\sigma = population standard deviation (or sample standard deviation if
population standard deviation is unknown)
o nn = sample size
Steps for Conducting a z-test:
1. State the hypotheses:
Null Hypothesis (H₀): μ=μ0\mu = \mu_0
Alternative Hypothesis (H₁): μ≠μ0\mu \neq \mu_0 (or one-sided
alternatives).
2. Select the significance level (α\alpha), typically 0.05.
3. Calculate the z-statistic using the formula above.
4. Find the critical z-value from the z-distribution table based on the
significance level (α\alpha).
5. Make a decision:
If the calculated z-value exceeds the critical z-value, reject the null
hypothesis.
If the p-value is less than α\alpha, reject the null hypothesis.
Conclusion
For small samples, use a t-test for means and a z-test for proportions if the sample
size is sufficiently large for normality conditions.
For large samples, use a z-test for means and a z-test for proportions.
Let me know if you need further clarifications or additional examples!
Statistical calculations using SPSS: Creating a Data File, Defining
Variables & data, Frequencies, Crosstabs, Hypothesis Testing,
Reliability test (and above tests), factor analysis, Chi-square, test for
goodness of fits and independence. T Test, Z-Test, F-Test, U-Test,
Kruskal-Wallis Test, Mann-Whitney U Test, Wilcoxon, Multivariate
analysis ANOVA one way and two way classification,