0% found this document useful (0 votes)
47 views

Measures of Association: Lesson 1 Data Analysis

This document provides an overview of measures of association used in data analysis. It discusses Pearson's product-moment correlation, which measures the strength and direction of the linear relationship between two continuous variables. The document defines correlation, provides examples of different correlation patterns, and shows how to calculate and test the significance of Pearson's r correlation coefficient. It also briefly introduces the Spearman rank-order correlation, which is appropriate for ordinal or ranked data, and discusses how to compute correlations in SPSS. The key goal is to help readers understand how to select the appropriate statistical technique to analyze relationships between variables in their data.

Uploaded by

lynette cantos
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
47 views

Measures of Association: Lesson 1 Data Analysis

This document provides an overview of measures of association used in data analysis. It discusses Pearson's product-moment correlation, which measures the strength and direction of the linear relationship between two continuous variables. The document defines correlation, provides examples of different correlation patterns, and shows how to calculate and test the significance of Pearson's r correlation coefficient. It also briefly introduces the Spearman rank-order correlation, which is appropriate for ordinal or ranked data, and discusses how to compute correlations in SPSS. The key goal is to help readers understand how to select the appropriate statistical technique to analyze relationships between variables in their data.

Uploaded by

lynette cantos
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 41

MODULE 4 MEASURES OF ASSOCIATION

LESSON 1 DATA ANALYSIS

Learning Objectives:

At the end of the lesson, you should be able to:

1. Select the appropriate statistical technique to be utilized when analyzing data;


2. Evaluate the types of relationships between variables;
3. Distinguish between the uses of the different statistical measures of association or
relationship: Pearson Product Moment correlation. Spearman Rank Order correlation,
and Chi-square;
4. Utilize descriptive and inferential statistics when examining the relationship between
variables; and,
5. Solve problems using the statistical measures for association.

Keywords and Phrases:

 Pearson product-moment correlation


 Spearman rank order correlation  Linear regression
 Chi-sqaure  Hypothesis testing
 Confidence intervals

Introduction

Another major type of data analysis is concerned with determining the extent to which
variables are related or associated with each other. Statistical tests available depend upon the
type of data gathered for the research. Some requires interval data as in Pearson product-Moment
Correlation (r). When data are ordinal or in ranked form, the Spearman Rank-Order Correlation
or rho is the formula which is more appropriate to use when determining the relationship of two
variables.

Measure of Association [Data Analysis]

93
Correlation Analysis

Correlation analysis is a statistical measure used to determine the strength or degree of


relationship between two or more variables. One main purpose in using correlation is to
determine the degree of linear relationship between two variables.

(a) Perfect positive linear association (b) Perfect negative linear


association

(c) Positive linear association (d) Negative linear association

(e) No linear association (f) Nonlinear association

Measure of Association [Data Analysis]

93
Figure 1 Hypothetical scatter diagrams indicating some forms of relationship between X and Y..

The Pearson Product Moment Correlation Coefficient (Pearson r)

If the scatter diagram indicates the presence of linear association between two variables,
the extent of their association can be measured. The most commonly used statistical measure is
the Pearson Product Moment Correlation Coefficient denoted by p and estimated by r.

This coefficient can take any value from -1 to +1, and it is a dimensionless or unit less
quantity. A positive value indicates a negative slope.

If we ignore the sign, the size of the coefficient indicates how close the data is to a
straight line. The closer the size is to one, the more closely the data follow the straight line. A
correlation equal to +1, indicates that the points lie exactly on a line with a positive slope; a -1
correlation indicates that the points lie on a line with a negative slope. A correlation close to 0
indicates that the points are more scattered.

The square of the correlation coefficient (r) is the coefficient of determination (r2).

In the example, the correlation coefficient of 0.95 indicates that (0.95) 2 (100) = 90.25%
of the variation in variables y is due to the linear relationship between x and y while 9.7%, called
an error is due to other unexplained factors.

Calculating the Pearson r:

The formula for calculating the Pearson r is:

n ∑ xy−( ∑ x )( ∑ y )
R=
√ [ n ( ∑ x ) − ( ∑ x ) ] [ n ( ∑ y )− ( ∑ y ) ]
2 2 2 2

X Y X2 Y2 XY
3 9 9 81 27
4 13 16 169 52
5 15 25 225 75
6 16 36 256 96
7 17 49 289 119
∑ x =25 ∑ Y =70 ∑ X 2=135 ∑ Y 2=1020 ∑ XY =369

Measure of Association [Data Analysis]

93
Substituting the Values

5 ( 369 )−25 (70)


R= 2 2
√ [ 5 ( 135 )−( 25 ) ] [ 5 ( 1020 )−( 70 ) ]
1845−1750
¿
√(50)( 200)
95
¿ =.95
100

A correlation coefficient of 0.95 indicates a positive high degree or positive strong


correlation between the values. A positive coefficient indicates a direct relationship; when the
values in one set increases, those of the other set also increase.

In linear correlation analysis, it is important that we know the extent by which the
variation in the dependent variable is related to the variation in the independent variable such
that we may be able to see the need for examining other related variables by such statistical tools
as the multiple or partial correlation.

A process by which we can measure the interdependence of the variables under


consideration is provided for by the square of the correlation coefficient which is known as the
coefficient of determination (r2). This value tells us the proportion of the total variation in the
dependent variable (y) which is explained by the linear relationship with the independent
variable(x).

Testing the Significance of Pearson r

The formula for testing the significance of Pearson r is called the t-test. It is being used to
verity if the value of the computed coefficient of correlation is significant, that is, does it
represent a real correlation, or is the obtained coefficient of correlation merely brought about by
chance?

The formula for t-test is,

n−2
t=r
√ 1−r 2

Where r = correlation coefficient

N = number of samples

Using the example given,


Measure of Association [Data Analysis]

93
n−2
t=r
√ 1−r 2

5−2
¿ .95
√ 1−(.95)2

3
¿ .95
√ 1−.9025

3
¿ .95
√ .0975
¿ .95 ( 5.547 )=5.27

This value with (N-2) degrees of freedom, or 3 degrees of freedom is significant at the 5
percent leve; (t.05 = 3.182, df = 3). Therefore, there is a significant relationship that exists
between x and y at .05 level of significance.

Pearson Product-Moment Correlation Coefficient in SPSS: To run the correlation in SPSS,


select Analyze, Correlate, Bivariate… from the Data Editor window.

Measure of Association [Data Analysis]

93
Select the variables for which we want to calculate the Pearson r correlation and drag
them to the right side of the variable list. Select “Pearson test” from the window. Click on the
“ok” button.

The result window will show the result of the Pearson correlation with the same set of
values used with manual computations.

Correlations

  X Y
X Pearson Correlation 1.000 .950*
Sig. (2-tailed) 0.013
N 5.000 5
Y Pearson Correlation .950* 1.000
Sig. (2-tailed) 0.013
N 5 5.000
*Correlation is significant at the 0.05 level (2-tailed)

CORRELATION FROM RANKS

Spearman Rank-Difference Method

Measure of Association [Data Analysis]

93
Computational procedure:

1. Record the rank order of the first distribution giving the highest score a rank of 1.
2. Do the same with the second distribution.
3. Get the difference between these two ranks for each individual and record in the D
column.
4. Square each of these difference and sum this column.
5. Apply the values in the formula:

6 ∑ D2
R=1−
n ( n2−1 )

Example:

A committee of five men and five women in a government agency evaluated ten
applicants for a vacant position in the office. They were asked to rate each applicant on the basis
of his/her qualification and interview. The ranking of the applicants by men and women are
given in the following table:

Applicant 1 2 3 4 5 6 7 8 9 10
Men(X) 6 2 1 4 9 10 5 3 8 7
Women(Y) 5 8 4 6 3 2 7 10 1 9

Solution:
Rx Ry D D2
1 6 5 1 1
2 2 8 -6 36
3 1 4 -3 9
4 4 6 -2 4
5 9 3 6 36
6 10 2 8 64
7 5 7 -2 4
8 3 10 -7 49

Measure of Association [Data Analysis]

93
9 8 1 7 49
10 7 9 -2 4

∑ D 2=256
Substituting the values:

6 ∑ D2
R=1−
n ( n2−1 )

6(256)
¿ 1−
10(102−1)

1536
¿ 1−
10(99)

1536
¿ 1−
990

¿ 1−1.55

¿ 0.55

A correlation coefficient of -0.55 indicates a negative, indirect or inverse relationship


between the two sets of variables. That is, when one set of variables increases, the other set
decreases. The level of relationship indicated by 0.45 is moderate.

Spearman rank correlation in SPSS: To run the correlation in SPSS, select Analyze,
Correlate, Bivariate…from the Data Editor window.

Measure of Association [Data Analysis]

93
The bivariate correlation window will appear after clicking on the bivariate
option. Select the correlation variable and put them on the right hand side of the variable list.
Then select the “Spearman test” from the window. Click on the “ok” button.

Measure of Association [Data Analysis]

93
The following table will show the Spearman rank correlation from the result window
with the same set of values used with manual computations.

Simple Linear Regression Analysis

Regression analysis is a statistical measure used to determine the functional relationship


between two or more variables. It is concerned with finding an equation that uses the known
values of one or more variables, called the independent or predictor

Measure of Association [Data Analysis]

93
variables, to estimate the unknown value of a quantitative variable called the dependent or
response variable.

The objectives of finding a statistical functional relationship between variables are:

 To establish the possible causation of changes in one variable by changes in other


variables;
 To predict or estimate the value of a variable given the values of other variables;
 To explain some of the variation of one variable by the other variables.

The Data and the Scatter Diagram

Suppose we want to find out whether the values of variable Y depend on the values of
another variable X, where Y is the dependent variable and X is the independent variable. The
first step is to collect a random sample of n pairs of values of X and Y, namely: (X 1, Y1), (X2,
Y2), …(Xn, Yn); then plot these values on the X-Y plane. The plotted points of the basic data on
the X-Y plane is called the scatter diagram. It gives us an idea on the possible relationship
netween X and Y. Figure 2 shows some forms of the scatter diagram. Scatter diagrams (a) and
(b) show a linear relationship between X and Y while; (c) exhibits no relation at all; (d) shows a
curved relation.

(a) Positive linear relationship (b) Negative linear relationship

Measure of Association [Data Analysis]

93
(c) No relationship (d) Nonlinear (quadratic) relationship

Figure 2: Hypothetical scatter diagrams indicating some forms of functional relationships


between X and Y.

The Data and the Regression Equation

The next step in regression analysis is to find a suitable function that expresses the
predicted value of Y given a value of X. The general pattern presented by the points on the
scatter diagram gives us a clue to finding the appropriate regression function. Figure 3 shows that
a linear function seems to fit the pattern of the data points.

The general form of the linear function is

Y =a+bX

Where a is called the Y-intercept of the line (the value of Y when X is equal to zero); and
b is the slope of the line called the regression (the rate of change of Y per unit change in X). An
illustration is made in Figure 3 with b>0.

Measure of Association [Data Analysis]

93
Figure 3: Graph of Y = a + bX when b>0

The line that describes the statistical relationship between X and Y is called the
regression line.

The regression line gives an estimate of the mean value of Y, denoted by Y, given the
value of X. hence the equation of the regression line is

Y =a+bX

Y should be distinguished from the observed data values which we denote simply by Y.
For a specific X, the resulting Y is a predicted value of the dependent variable.

Characteristics of the Regression Line

1. When b> 0, Y increases as X increases. In this case, we say that Y is directly or


positively related to X. Figure (a) is an example of this relationship.
2. When b< 0, Y decreases as X increases, and we say that Y is inversely or negatively
related to X. Figure (b) is an example.
3. When b = 0, Y is a constant and is equal to the y-intercept a. This implies that there is no
change in Y whatever X be. (horizontal line).

Calculation of Simple Regression

Given: Equation of the line Y = a + bX

Computing for a and b:

b yx =
∑ XY −[ ∑ X ∑ Y / N ]
[∑ x 2−(∑ X )2 /N ]
Measure of Association [Data Analysis]

93
a yx =Y −b yx X

Data Scores of 10 students on two entrance tests, Test I in English, Test II in Filipino

Test 1 (X) Test 2 (Y) X2 Y2 XY


35 15 1225 225 525
30 13 900 169 390
27 14 729 196 378
26 10 636 100 260
24 8 576 64 184
23 10 529 100 230
18 12 324 144 216
17 8 289 64 136
15 10 225 100 150
13 8 169 64 104
12 6 144 36 72
11 5 121 25 55
8 2 64 4 16
6 4 36 16 24
5 0 25 0 0
∑ X=270 ∑ Y =125 ∑ X 2=5992 2
∑ Y =1307 ∑ XY =2740

∑ X=270
∑ Y =125
∑ XY =2740
∑ X 2=5992
∑ Y 2=1307
n = 15

X 270
X =∑ = =18
N 15

Y 125
Y =∑ = =8.3
N 15

Substituting to the above equations for a and bxy

b xy =∑ XY −¿ ¿

Measure of Association [Data Analysis]

93
n ∑ xy−∑ x ∑ y
¿ 2
n ∑ x 2−∑ x

¿¿

¿ 0.43286

a yx =Y −b yx X

=8.33 – (0.43286)(18)

=8.33 – 7.7915

=0.5385

The least squares regression line is therefore

Y =a+bX

Y =0.5385+0.43286 X

We may now use this equation to predict Test 2 scores given Test I score.

For instance, when X=30

Y =0.5385+0.4286 ( 30 )=13.496=13.5

The proper interpretation of 13.5 as a predicted value is that on the average a student
obtain a score of 13.5 in Test 2if his score in Test 1 is 30.

Simple Linear Regression in SPSS. To run simple linear regression in SPSS, select Analyze,
Regression, Linear... from the data editor window.

Measure of Association [Data Analysis]

93
Then, specify Test 2 in dependent variable box and test 1 in the independent variable box.

The following table will show the regression result:

Measure of Association [Data Analysis]

93
a.

a. Predictors: (Constant), Test 1

b. Dependent Variable: Test 2

*Difference in manual computation and use of SPSS is due to rounding off errors.

Chi-square (X2)

The chi-square (Greek letter chi, X2 ) is the most commonly used method of comparing
proportions. It is particularly useful in test evaluating a relationship between nominal- or ordinal
data. Typical situations or settings are cases where persons, events, or objects are grouped in two
or more nominal categories such as “Yes – No” responses. “Favor – Against – Undecided” or
class “A, B, C, or D”.

Chi-square analysis compares the observed frequencies of the responses with the
expected frequencies. It is given by the formula:

2 (F ¿ ¿ o−F e )2
X = ¿
Fe

Where: Fo = observed number of cases

Fe = expected number of cases

Measure of Association [Data Analysis]

93
Chi-square and Goodness of Fit: One Sample Case

There are research problems where responses fall in any one of a number of categories.
The data are expressed in frequencies and the observed frequencies (Fo ) are compared to the
frequencies expected (Fe ) on the basis of some hypothesis.

If the differences between the observed and the expected frequencies are small, X2 will be
small. The greater the difference between the observed and expected frequencies under the null
hypothesis, the greater or larger the X2 will be. If the difference between the observed and
expected values are so large collectively as to occur by chance only, say 0.05 or less, when the
null hypothesis is true, then the null hypothesis is rejected.

Illustration:

Consider the nomination of three (3) presidential candidates of a political party, A, B, and
C. The chairman wonders whether or not they will be equally popular among the members of the
party. To this the hypothesis of equal preference, a random sample of 315 men were selected and
interviewed which one of the three candidates they prefer. The following are the results of the
survey:

Candidates Frequency

A 98

B 115

C 102

Are you going to reject the null hypothesis that equal members of men in the party prefer
each of the three-candidates? Or are you going to accept the null hypothesis of equality of
preference?

Calculating the X2 value:

Fo Fe Fo – Fe F – F e )2
( o (Fo – Fe )2 / Fe

A 98 105 -7 49 49/105=0.467

B 115 105 10 100 100/105=0.952

C 102 105 -3 9 9/105=0.085 .

1.505

Measure of Association [Data Analysis]

93
2 [f 0 −f e ]2
X =∑ =1.505
fe

In order to test the significance of the computed X2 value using a specified criterion of
significance, the obtained value is referred to a table with appropriate degrees of freedom which
is equal to k-1, where k is equal to the number of categories of the variable. In this problem,
df=3-1=2. Therefore, for the X2 to be significant at the 0.05 level, the computed value should be
more than (>) the tabular value which is 5.991.

Summarizing,

Level of significance =5%

df = k-1 (number of categories minus one)

=3-1

=2

Critical value: X2 (.05.2) = 5.991

Decision rule: Reject Ho if X computed > 5.991, otherwise do not reject Ho .

Conclusion:

Since 1.505 < 5.991, do not reject Ho . There is no sufficient evidence or reason to
reject the null hypothesis that the frequencies in the population are equal.

Measure of Association [Data Analysis]

93
Chi-square and Goodness of Fit: One Sample Case in SPSS: From the Data Editor window.
Data are entered in the SPSS like this

Click on the “analysis” menu and select “nonparametric test.”

Measure of Association [Data Analysis]

93
The following table will show the Chi-square correlation from the result window:

Correlations
  Observed N Expected N Residual
98 98 105 -7
102 102 105 -3
115 115.000 105 10
Total 315    

Test Statistics
  Votes
Chi- 1.505
square (a)  
Df 2
Asymp. Sig 0.471

a 0 cells (.0%) have expected frequencies less than 5. The minimum expected cell frequency is
105.0.

Measure of Association [Data Analysis]

93
Chi-square as a Test of Independence: Two Variable Problems

Chi-square can also be used to test the significance of the relationship between two
variables when data are expressed in terms of frequencies of joint occurrence. For this kind of
problem, a two-way contingency table with rows and columns are set-up.

In the case of two variable problems, the expected frequencies are those predicated on the
independence of the two variables. The formula is equal to

( rowtotal ) (columntotal)
F e= [ N ]
Example:

Suppose, one wants to know if there is a relationship between gender and school
choice. A sample of 100 female and 100 male freshman students were asked individually for
their school choice. Test the null hypothesis of no significant relationship between the students’
gender and school choice at 5% level of significance.

Gender
School choice
Female Male Total
Public 42 C1 65 C3 107
Private 58 C2 35 C4 93
Total 100 100 200

Computational Procedure:

1. Compute the expected frequency of each cell


Expected Frequency (Fe ) = Raw Total X Column Total
Grand Total

100 x 107
C 1= =53.5
200

100 x 93
C 2= =46.5
200

100 x 107
C 3= =53.5
200

Measure of Association [Data Analysis]

93
100 x 93
C 4= =46.5
200

2. Present in table form


Fo Fe Fo – Fe F – F e )2
( o (Fo – Fe )2 / Fe

42 53.5 -11.5 132.25 2.48

58 46.5 11.5 132.25 2.85

65 53.5 11.5 132.25 2.48

35 46.5 -11.5 132.25 2.85

( f ¿ ¿ o−f e )
∑ fe
=10.68 ¿

Therefore X2 =10.66

3. Prepare the hypothesis

 Null hypothesis: No significant relationship exist between gender and school


choice.
 Level of Significance: 5%
 Degrees of freedom = (c-1)(r-1)

c=no. of columns

r=no. of rows

df=(2-1)(2-1)=(1)(1)=1

CV=3.841

4. Decision Rule

Since the computed value X2 = 10.66 is greater than the critical value (3.841), we
reject the null hypothesis. Thus, the two variables of gender and school are related-
females tend to prefer private schools while males tend to prefer to study in public
schools.

Measure of Association [Data Analysis]

93
Pearson Chi Square in SPSS:Select Analyze, Descriptive Statistics,
Crosstabs...from the Data Editor window

As you click on the cross tab, the following window will appear. From this window, select
the row variable and insert it as a marked row. Select the second variable and put them in to
mark a column

Measure of Association [Data Analysis]

93
The cross tabulation table for the Chi-Square test of Independence will show the expected
value for two nominal variables.

School* Gender Crosstabulation

Measure of Association [Data Analysis]

93
  Gender
  Female Male Total
School Public
42 65 107
School
Private
52 35 93
School
Total 100 100 200

The Chi-square tests table will show the value of Pearson Chi-Square value,
associated with the significance value.

Chi-square test
d
  Value f Asymp. Sig (2 sided) Exact Sig. (2 sided Exact Sig. (1-sided)
10.63
Pearson Chi-square 2 1 0.001  
Continuity Correction 9.728 1 0.002  
Likelihood Ratio 10.73 1 0.001  
Fisher's Exact Test     0.002 0.001
10.57
Linear-by-Linear 9 1 0.001  
Association      
N of Valid Cases 200        

a. 0 cells (.0%) have expected count less than 5. The minimum expected count is
46.50.
b. Computed only for a 2x2 table

MORE SPSS APPLICATIONS

Measure of Association [Data Analysis]

93
CHI-SQUARE TEST

Example

A researcher is interested whether there is a significant relationship between gender and

educational attainment of the respondents. To investigate the issue, the researcher randomly

selected 25 teachers and noted their gender and educational attainment. Gender is coded 1 for

male and 2 for female, and the Educational Attainment is coded 1 for BA/BS Degree Holder,

2 for With Units in MA/MS, 3 for MA/MS Degree Holder. The data are shown in Table 1.

Table 1. The Gender and Educational Attainment of 25 Randomly Selected Teachers

Respondent Gender Educational Attainment


1 1 1
2 1 2
3 1 1
4 1 3
5 1 3
6 2 3
7 2 2
8 2 2
9 2 2
10 1 1
11 2 1
12 1 1
13 2 2
14 1 2
15 2 2
16 1 3
17 2 3
18 2 3

Measure of Association [Data Analysis]

93
19 2 3
20 2 2
21 2 2
22 2 2
23 2 3
24 2 3
25 2 3

Step 1. Enter your data in the SPSS Data Editor.

Figure 1.1 SPSS Data Editor after data are entered

Step 2. Run the chi-square test

 Click on Analyze.
 Click on Descriptive Statistics.
 Click on Crosstabs
 A dialog box will open, with a list of your variables in the white rectangle on the left.
Click on the independent variable, and then click on the small right arrow to the left of
the white box headed Column(s). This will move the independent variable into the
Column(s) box (in this example “EducAttainment”).
 Click on the dependent variable and then click on the small right arrow to the left of the
white box headed Row(s) (see figure 1.2). This will move the dependent variable into
Row(s) box. In this example the dependent variable is “Gender”.

Measure of Association [Data Analysis]

93
Figure 1.2 Entering variables in Crosstabs

 Click on Statistics.
 Click on the chi-square box. A tick will appear there.
 Click on Continue.
 Click on Cells.
 In the percentages area click on Column. A tick will appear
 Click on Continue
 Click on OK.

An output window will open with the results of the analysis as follows.

This table shows the total number of cases broken


down by valid cases and missing values

Measure of Association [Data Analysis]

93
Gender * EducAttainment Crosstabulation

EducAttainment Total

BA/BS Degree With Units in MA/MS Degree


Holder MA/MS Holder

Count 4 2 3 9
1.00
% within EducAttainment 80.0% 20.0% 30.0% 36.0%
Gender
Count 1 8 7 16
2.00
% within EducAttainment 20.0% 80.0% 70.0% 64.0%
Count 5 10 10 25
Total
% within EducAttainment 100.0% 100.0% 100.0% 100.0%

p-value. If this p-value is less than to the assigned


Chi-square level of significance, we say that the result is
statistically significant

Case Processing Summary

Cases

Valid Missing Total

N Percent N Percent N Percent


 

Chi-square was used to determine if the gender of the teachers is significantly related to
educational attainment. Since the computed p-value (0.065_ with chi-square value of 5.469 is
greater than to the assigned level of significance (0.05), therefore, the hypothesis that there is no
significant relationship between gender and educational attainment of the teachers is thus
accepted. A non-significance chi-square indicates that, in the population from which we have
sampled, the data in the each row does not significantly related to data in each column.

Measure of Association [Data Analysis]

93
SPSS APPLICATION – t-TEST FOR INDEPENDENT SAMPLES

Example

A researcher is interested whether there is a significant difference in the teaching


performance of male and female teachers. The researcher randomly selected 24 teachers, noted
their gender, and evaluated their teaching performance. Gender is coded 1 for male and 2 for
female. The data are shown in Table 2.

Case Gender Teaching Performance


1 1 98
2 1 95
3 1 92
4 1 94
5 1 89
6 2 79
7 2 82
8 2 88
9 2 85
10 1 94
11 2 83
12 1 97
13 2 80
14 1 91
15 2 82
16 1 90
17 2 81
18 2 83
19 2 84
20 2 85
21 2 85
22 2 80
23 2 90
24 1 97

Step 1. Enter your data in the SPSS Data Editor.

Measure of Association [Data Analysis]

93
Figure 2.1 SPSS Data Editor after independent samples – test data entered in SPSS Data Editor

Step 2. Run the independent Samples t-test,

 Click on Analyze
 Click on Compare means
 Click on Independent-Samples T Test
 A dialog box will open with a list of your variables in the white rectangle on the left hand
side. Click on the dependent variable, and then click on the small black arrow to the left
of the white box headed Test Variable(s). This will place the dependent variable (in this
case Teaching Performance) in the box.
 Click on the independent variable and then click on the small black arrow to the left of
the white box headed Grouping Variable. This will place the independent variable (in
this case Gender) in the box (see Figure 2.2)

Measure of Association [Data Analysis]

93
Figure 2.2 Entering the test and grouping variables

 Click on the Define Groups button.


 The dialog box shown in Figure 2.3 will open. Enter the two values of the independent
variable in the white boxes. Because in our example males are coded 1 and females 2 (see
Table 2), these are the two values shown in Figure 2.3

Measure of Association [Data Analysis]

93
Figure 2.3 Entering the levels of the independent variable

 Click on Continue
 Click on OK.
The output will be as follows

Group Statistics

Std. Std. Error


Gender N Mean Deviation Mean
TeachingPerformance Male 11 93.5455 3.01210 0.90818
Female 14 83.3571 3.10353 0.82945

The mean of each The standard


Number of cases in each
group deviation of each
group
group

Measure of Association [Data Analysis]

93
Significance level: the
probability of a significant
Significance of Leven’s Test for t-value Degrees of difference in the means as
equality of variances. If this value is freedom great as the one here if the
less than 0.05, use the t-value, df, and
null hypothesis is true. This
significance level given for
value needs to be below
compared to the assigned
level of significance for
statistical

Referring to the Independent Samples Test table again we can see the results of the t-test.
The t-value is 8.253, there 23 degrees of freedom and the (two tailed) significance level is 0.000.
The difference between two groups’ means is significant in this case because 0.000 is less than o
the assigned level of significance (0.05).

The mean teaching performance were 93.55 (SD = 3.01) for males and 83.36 (SD = 3.10)
for females. The teaching performance of males and females were compared using an
independent samples t-test. There was no significant difference between the teaching
performance of males and females t(23) = 8.253, p-value = 0.000

Measure of Association [Data Analysis]

93
PAIRED SAMPLE t-TEST

A researcher is interested whether a training course increases or decreases the teaching


performance of the teachers who attended the training courses. The data are shown in table 3.

Table 3. The Teaching Performance of 25 Teachers

Before and After a Training Program

Cas
e Teaching Performance before Training Teaching Performance After Training
1 85 95
2 84 98
3 86 97
4 87 92
5 89 96
6 82 93
7 80 94
8 84 95
9 86 90
10 82 82
11 89 97
12 87 98
13 82 95
14 81 95
15 86 92
16 89 91
17 89 94
18 84 95
19 85 96
20 88 97
21 81 99
22 80 98
23 86 99
24 87 91
25 89 98

Measure of Association [Data Analysis]

93
Step 1. Enter your data in the SPSS Data Editor.

Figure 3.1 Paired samples t-test data entered in SPSS Data Editor

Step 2. Run Paired Samples t-test.

 Click on Analyze
 Click on Compare Means
 Click on Paired Samples T Test
 A dialog box will appear. Click on both the variables you want to include in the analysis,
and then click on the small black arrow to the left of the white box headed Paired
Variables. The variables will move into the white box as in shown in Figure 3.2.
 Click OK.

Measure of Association [Data Analysis]

93
Figure 3.2 Entering variables in a paired sample t-test
The output will be as follows

Paired Samples Statistics

Std. Std.
Deviatio Error
  Mean N n Mean
Pair 1        
before_TeachingPerformance 85.12 25 3.00444 0.60089
after_TeachingPerformance 94.68 25 3.70495 0.74099

Mean Number of Cases Standard Deviations

Measure of Association [Data Analysis]

93
Paired Samples test

t-value Degrees of freedom

Significance level: The


probability of a difference in the
means as great as the one here if
the null hypothesis is true. This
value needs to be below to the
assigned level of significance
(alpha) for statistical sifnificance

The mean Teaching Performance was 85.12 (SD = 3.00) before the training course
and 94.68 (SD = 3.70) after the training course. The difference between the mean level of
Teaching Performance before and after the training course was examined with a paired samples
t-test. Teaching Performance was significantly higher after the training course than before it t(24)
= 10.347, p-value < 0.05.

Reading Assignment:

 E-Journals/E-books:
 PUP website: infotrac.galegroup.com/itweb/pup
 Password: powersearch

Exercises/Written Assignments:

1. Compute the coefficient of correlation (r) and coefficient of determination (r 2) between


the monthly income (in thousands of pesos) and the amount of expenses (in thousands of
pesos) of ten families of the same size.
Monthly Income 20 14 15 18 30 34 36 47
Monthly expenses 15 12 25 20 25 30 35 32
Measure of Association [Data Analysis]

93
2. The scores on an aptitude test of ten students and the number of hours they reviewed for the are test
are as follows:

Number of Hours 18 27 20 10 30 24 32 27 12 16
Scores (y) 68 82 77 90 78 72 94 88 60 70

2.1 Compute the coefficient of correlation (r) and coefficient of determination ( r 2)

2.2 Test the significance of the correlation coefficient.

3. Table 1 gives the ranks of ten applicants for teaching position on intelligence and emotional
quotients as shown by their psychological test results.

Applicant Intelligent Quotient Emotional Quotient


A 1 1
B 4 3
C 8 9
D 5 2
E 8 6
F 10 8
G 6 7
H 3 5
I 2 4
J 7 10

A. Calculate the Spearman’s rho.

B. Test the significance of the Spearman’s rho.

4. In the following given data, x =number of sessions attended by 15 trainees in a leadership training
seminar, while y = scores obtained by the same trainees in a test given after the seminar.

x 3 2 4 5 5 6 6 7 9 7 8 5 6 3 8
y 65 50 75 70 80 85 79 88 91 87 88 70 71 63 85

a. Determine the regression equation for predicting y from x

b. Predict the mean score of trainees who attend an average of 8 sessions

Measure of Association [Data Analysis]

93
References/Bibliography

Downie, N. M. & Heath, Robert W. Basic Statistical Methods (Fifth Edition). 1984

Ferguson, George A. Statistical Analysis in Psychology and Education. New York:


McGraw-Hill. 1989

Field, M.N., Ryan, J.M., & Hess, R.K. Handbook of statistical procedures and their
computer applications to education and the behavioural sciences. NY: MacMillan
Publishing Co. 1991

Naval, Victoria C. et al. General Statistics. Conana Publishing. 1988

Ostle, Bernard B. Statistics in Research: Basic Concepts and Techniques for Research
Workers. Ames: lowa State University Press. 1988

Kumar, R. Research Methodology, pp. 1-15. Busy Book Distributors. 1996

Measure of Association [Data Analysis]

93

You might also like