Correlation
Correlation
Correlation
Figure 5.1: Data entry for correlation. The spreadsheet tells us that subject 1 was shown 4 adverts and subsequently purchased 10 packets of crisps. We are going to analyse some data regarding undergraduate exam performance. The data for several examples are stored on my web page (http://www.cogs.susx.ac.uk/users/andyf/teaching/) in a single file called ExamAnx.sav. If you open this data file you will see that these data are laid out in the spreadsheet as separate columns and that gender has been coded appropriately. We will discover to what each of the variables refer as we progress through this chapter.
Page 1
3/12/00
Y Axis : Specify the variable that you wish to be plotted on the y axis (ordinate) of the graph. This should be the dependent variable, which in this case is exam performance. Use the mouse to select exam from the list (which will become highlighted) and then click on to transfer it to the space under where it says Y Axis.
X Axis : Specify the variable you wish to be plotted on the x axis (abscissa) of the scatterplot. This should be the independent variable, which in this case is anxiety. You can highlight this variable and transfer it to the space underneath where it says X axis. At this stage, the dialogue box should look like Figure 5.3. Set Markers by : You can use a grouping variable to define different categories on the scatterplot (it will display each category in a different colour). This function is useful, for example, for looking at the relationship between
Page 2
3/12/00
When you have completed these options you can click on type in a title for the scatterplot. You can also click on values.
120
100
80
60
20
Gender Female
Male
Page 3
3/12/00
then another dialogue box appears with two statistics options and two options for missing
Page 4
3/12/00
Figure 5.6: Dialogue box for bivariate correlation options The statistics options are enabled only when Pearsons correlation is selected, if Pearsons correlation is not selected then these options are disabled (they appear in a light grey rather than black and you cant activate them). This is because these two options are meaningful only for parametric data. If you select the tick-box labelled means and standard deviations then SPSS will produce the mean and standard deviation of all of the variables selected for correlation. If you activate the tick-box labelled Cross-product deviations and covariances then SPSS will give you the values of these statistics for each of the variables being correlated (for mo re detail see Field, 2000).
5.3.1.
For those of you unfamiliar with basic statistics (which shouldnt be any of you !), it is not meaningful to talk about means unless we have data measured at an interval or ratio level. As such, Pearsons coefficient requires parametric data because it is based upon the average deviation from the mean. However, in reality it is an extremely robust statistic. This is perhaps why the default option in SPSS is to perform a Pearsons correlation. However, if your data are nonparametric then you should deselect the Pearson tick-box. The data from the exam performance study are parametric and so a Pearsons correlation can be applied. The dialogue box (Figure 5.5) allows you to specify whether the test will be one- or two-tailed. One-tailed tests should be used when there is a specific direction to the hypothesis being tested, and two tailed tests should be used when a relationship is expected, but the direction of the relationship is not predicted. Our researcher predicted that at higher levels of anxiety exam performance would be poor and that less anxious students would do well. Therefore, the test should be one-tailed because she is predicting a relationship in a particular direction. Whats more, a positive correlation between revision time and exam performance is also expected so this too is a one tailed test. To ensure that the output displays the one-tailed significance click on and then click .
SPSS Output 5.1 provides a matrix of correlation coefficients for the three variables. It also displays a matrix of significance values for these coefficients. Each variable is perfectly correlated with itself (obviously) and so r = 1 . Exam performance is negatively related to exam anxiety with a correlation coefficient of r = 0. 441 which is significant at p < 0.001 (as indicated by the ** after the coefficient). This significance value tells us that the probability of this correlation being a fluke is very low (close to zero in fact). Hence, we can have confidence that this relationship is genuine and not a chance result. Usually, social scientists accept any probability value below 0.05 as being statistically meaningful and so any probability value below 0.05 is regarded as indicative of genuine effect. The output also shows that exam performance is positively related to the amount of time spent revising, with a coefficient of r = 0 .3967 , which is also significant at p < 0.001. Finally, exam anxiety appears to be negatively related to the amount of time spent revising (r = -0.7092, p < 0.001). In psychological terms, this all means that as anxiety about an exam increases, the percentage mark obtained in that exam decreases. Conversely, as the amount of time revising increases, the percentage obtained in the exam increases. Finally, as revision time increases the students anxiety about the exam decreases. So there is a complex interrelationship between the three variables.
Page 5
3/12/00
Sig. (1-tailed)
SPSS Output 5.1: Output from SPSS 7.0 for a Pearsons correlation
5.3.1.1.
A considerable amount of caution must be taken when interpreting correlation coefficients because they give no indication of causality. So, in our example, although we can conclude that exam performance goes down as anxiety about that exam goes up, we cannot say that high exam anxiety causes bad exam performance. This is for two reasons: The Third Variable Problem: In any bivariate correlation causality between two variables cannot be assumed because there may be other measured or unmeasured variables effecting the results. This is known as the third variable problem or the tertium quid. In our example you can see that revision time does relate significantly to both exam performance and exam anxiety and there is no way of telling which of the two independent variables, if either, are causing exam performance to change. So, if we had measured only exam anxiety and exa m performance we might have assumed that high exam anxiety caused poor exam performance. However, it is clear that poor exam performance could be explained equally well by a lack of revision. There may be several additional variables that influence the correlated variables, and these variables may not have been measured by the researcher. So, there could be another, unmeasured, variable that affects both revision time and exam anxiety. Direction of Causality: Correlation coefficients say nothing about which variable causes the other to change. Even if we could ignore the third variable problem described above, and we could assume that the two correlated variables were the only important ones, the correlation coefficient doesnt indicate in which direction causality operates. So, although it is intuitively appealing to conclude that exam anxiety causes exam performance to change, there is no statistical reason why exam performance cannot cause exam anxiety to change. Although the latter conclusion makes no human sense (because anxiety was measured before exam performance), the correlation does not tell us that it isnt true. Using r for Interpretation
2
5.3.1.2.
Although we cannot make direct conclusions about causality, we can draw conclusions about variability by squaring the correlation coefficient. By squaring the correlation coefficient, we get a measure of how much of the variability in one
Page 6
3/12/00
5.3.2.
Spearmans Rho
Spearmans correlation coefficient is a nonparametric statistic and so can be used when the data have violated parametric assumptions and/or the distributional assumptions. Spearmans tests works by first ranking the data, and then applying Pearsons equation to those ranks. As an example of nonparametric data, a drugs company was interested in the effects of steroids on cyclists. To test the effect they measured each cyclists position in a race (whether they came first, second or third etc.) and how many steroid tablets each athlete had taken before the race. Both variables are nonparametric, because neither of them was measured at an interval level. The position in the race is ordinal data because the exact difference between the ability of the runners is unclear. It could be that the first athlete won by several metres while the remainder crossed the line simultaneously some time later, or it could be that first and second place was very tightly contested but the remainder were very far behind. The Spearman correlation coefficient is used because one of these variables is ordinal not interval. The data for this study are in the file race.sav. The procedure for doing the Spearman correlation is the same as for the Pearsons correlation e xcept that in the bivariate correlations dialogue box ( Figure 5.5), we need to select and deselect the option for a Pearson correlation. At this stage, you should also specify whether you require a one- or two-tailed test. For the example above, we predict that the more drugs an athlete takes, the better their position in the race. This is a directional hypothesis and so we should perform a one tailed test. SPSS Output 5.2 shows the output for a Spearman correlation on the variables position and tablets . The output is very simple, first a correlation matrix is displayed that tells us that the correlation coefficient between the variables is -0.599. Underneath is a matrix of significance values for the correlation coefficient and this tells us that the coefficient is significant at p < 0.01. Therefore, it can be concluded that there is a significant negative relationship between the number of tablets an athlete took and their position in the race. Therefore, cyclists who took high numbers of tablets had numerically low positions (i.e. 1, 2, 3), which in real terms means that they did better in the race (because 1 is first place, 2 is second place and so on). Finally, the output tells us the number of observations that contributed to each correlation coefficient. In this case there were 18 athletes and so N = 18. It is good to check that the value of N corresponds to the number of observations that were made. If it doesnt then data may have been excluded for some reason.
Page 7
3/12/00
Position In race Spearman's rho Correlation Coefficient Position In race No. of Steroid tablets taken Position In race No. of Steroid tablets taken Position In race No. of Steroid tablets taken 1.000
-.599**
1.000
Sig. (1-tailed)
.004
.004
18
18
18
18
SPSS Output 5.2: Output from SPSS 7.0 for a Spearman Correlation.
5.3.3.
Kendalls tau is another nonparametric correlation and it should be used rather than Spearmans coefficient when you have a small data set with a large number of tied ranks. This means that if you rank all of the scores and many scores have the same rank, the Kendalls tau should be used. Although Spearmans statistic is more popular of the two coefficients, there is much to suggest that Kendalls statistic is actually a better estimate of the correlation in the population (see Howell, 1992, p.279). As such, we can draw more accurate generalis ations from Kendalls statistic than from Spearmans. To carry out Kendalls correlation on the race data simply follow the same steps as for the Pearson and Spearman correlation but select and deselect the Pearson option. The output is much the same as for Spearmans correlation. Youll notice from SPSS Output 5.3 that although the correlation is still highly significant, the actual value of it is less than the Spearman correlation (it has decreased from - 0.599 to - 0.490). We can still interpret this as a highly significant result, because the significance value is still less than 0.05 (in fact, the value is the same as for the Spearman correlation, p = 0.004). However, Kendalls value is a more accurate gauge of what the correlation in the population would be. As with the Pearson correlation we cannot assume that the steroids caused the athletes to perform better.
Page 8
3/12/00
Position In race Kendall's tau_b Correlation Coefficient Position In race No. of Steroid tablets taken Position In race No. of Steroid tablets taken Position In race No. of Steroid tablets taken 1.000
-.490**
1.000
Sig. (1-tailed)
.004
.004
18
18
18
18
This handout contains large excerpts of the following text (so copyright exists!)
Field, A. P. (2000). Discovering statistics using SPSS for Windows: advanced techniques for the beginner. London: Sage.
Go to http://www.sagepub.co.uk to order a copy
Page 9
3/12/00