Correlation and Partial Correlation Flashcards

1
Q

Bivariate linear correlation

A
  • examines the relationship between two variables
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what can bivariate relationships vary in:

A
  • form (linear, curvilinear)
  • direction (positive / negative)
  • magnitude/strength
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

magnitude/strength in bivariate relationships

A
  • r = -1 : perfect negative relationship
  • r = +1 : perfect positive relationship
  • r = 0 : no relationship
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

positive or negative correlation

A

correlation does not mean causation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

strength of correlation : strong negative/positive

A

+/- 0.9, 0.8, 0.7

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

strength of correlation: moderate negative/positive

A

+/- 0.6, 0.5, 0.4

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

strength of correlation: weak negative/positive

A

+/- 0.3, 0.2, 0.1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

correlation hypothesis testing

A
  • linear correlation involves measuring the relationship between two variables measured in a sample
  • we are more interested in if there is a relationship in equivalent population variables
  • use sample statistics to estimate population parameters
  • H0: there is no relationship between the population variables
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

p-value in correlation

A

what is the chance of measuring a relationship of that magnitude when the null hypothesis is true?

  • reject null if p < .05
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

parametric assumptions

A
  • both variables should be CONTINUOUS (if not use non-parametric alternative)
  • related PAIRS: each participant should have a pair of values
  • absence of outliers
  • linearity: point sin scatterplot should be best explained with a STRAIGHT line
  • sensitive to range restrictions: floor and ceiling effects
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

non-parametric alternative

A
  • if assumptions violated
  • Spearman’s rho (or Kendall’s Tau if fewer than 20 cases)

(or fewer than 7 points on a likert scale you use one of these)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

floor effect

A

cluster of scores at bottom of scale
- form of range restriction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

ceiling effect

A

clustering of scores at top of scale
- form of range restriction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

PPMCC

A

pearson product-moment correlation coefficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what does Pearson’s correlation coefficient investigate

A

the relationship between two quantitative, continuous variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

what does Pearson’s produce

A

a correlation coefficient ‘r’ which is a measure of the strength of association between the two variables

17
Q

Covariance

A
  • for each data point, calculate the difference from the mean of x, and the difference from the mean of y
  • multiply the differences
  • sum the multiplied differences
  • divide by N -1
18
Q

what does covariance do

A
  • provides a measure of the variance shared between x and y variables
19
Q

correlation coefficient and covariance

A
  • ‘r’ is a ratio of covariance (shared variance) to separate variances
  • we can obtain a measure of separate variances by multiplying the standard deviation for x and y
  • DON’T NEED TO DO THIS BY HAND
20
Q

correlation coefficient strength

A
  • ‘r’ is a ratio
  • if covariance is large relative to separate variances, r will be further from 0
  • if covariance is small relative to separate variances, r will be closer to 0
21
Q

what can r represent

A

it can tell us how well a straight line fits the data point i.e. the strength of correlation

  • if data points cluster around the line, r is further from 0
  • if data points scattered around line, r is closer to 0
22
Q

SPSS output for correlation

A
  • r tells you strength
  • p tells you if correlation is significant
  • N helps you calculate d.f.
23
Q

degrees of freedom for r

A

N - 2
- report when reporting r
e.g. r(23) = .522, p = .007

23
Q

sampling error

A
  • r value obtained from another sample from the same population would likely be different
  • reflects sampling error
24
Q

Sampling Distribution of Correlation Coefficients

A
  • if we obtained r for all possible samples drawn from the population of interest … the mean resulting distribution would be equivalent to the true population correlation coefficient
  • H0: no relationship between population variables (i.e. r = 0)
  • so under the null the sampling distribution of correlation coefficient will have a mean of 0

(normal distribution)

25
Q

r distribution

A
  • has a mean of 0
  • extent to which an individual sampled r deviates from 0 can be expressed in standard error units
  • distribution depends on r value-of underlying population and number of samples
26
Q

confidence interval around r

A
  • r is a point estimate of underlying population r-value and it is subject to sampling error
  • ’ we have 95% confidence that the population correlation co-efficient falls between __ and __’
27
Q

SPSS output for CIs

A

image shows for CIs around r

(for CIs for independent variables look at main descriptive statistics table at the top of SPSS output)

28
Q

shared variance

A
  • r^2
  • expressed the proportion of the separate variances that is shared
  • e.g. r = .8, r^2 = .64, variables share 64% variance
    … meaning that 18% of variance of each variable is not shared
29
Q

note on r and shared variance - relative strength

A

a weaker r e.g. .4 vs .8 means
.8 is 4x as strong as .8

we use r^2 to talk about relative strength e.g. how strong is .4 compared to .8

30
Q

effect size

A

r is a measure of effect size, and once squared to give shared variance it can be expressed as a proportion of separate variances, telling us how much of variance in y can be ‘explained’ by x (similar to partial eta squared where how much of variance in DV can be explained by manipulation of IV represented by partial eta sqaured)

31
Q

partial correlation

A

allows us to examine the relationship between two variables, while removing the influence of a third variable

e.g. we want to control for a confounding variable: when looking at the relationship between IQ and grade, we want to remove.control the influence of test motivation

32
Q

how can we control for a third variable (confounding)

A
  • recruit p’s who have same levels of variable e.g. same level of motivation
  • control variable through statistical means e.g. ‘partial out’ variable or ‘hold variable constant’
33
Q

‘partialling’ out a variable and SPSS output

A
  • if removing third variable, correlation between x and y is reduced and may no longer be significant
  • if this case you say
    ‘relationship between X and Y may well have been explained by Z (third variable) on both X and Y
  • if relationship had remained significant it would suggest the relationship was partially explained by third variable
  • if correlation had not decreased, relationship is not explained by third variable
34
Q

write up: correlation and partial correlation: design

35
Q

write up correlation and partial correlation: results section: step one: descriptive statistics

A

DS for each variable
- measure of CT: mean
- measure of spread: SD
- interval estimate: 95% CIs (lower and upper)
- - ‘Descriptive statistics and bi-variate correlation between study variables in Table 1’

36
Q

write up correlation and partial correlation: results: step 2: inferential analysis

A
  • state test used
  • always ‘Pearson’s correlation coefficient revealed a (significance)(direction if significant) relationship between X and Y.
  • report: r(df) - _.__, p = .__, 95% CIs [.__, .__] (CIs around mean r) if only 2 variables, if partial correlation report pearson’s in table (PIC) but always include partial in text
  • (if needed) partial correlation revealed that the relationship between X and Y was (significance) when controlling for Z, r(df) = .__, p = .__
  • df as N-2 instead of N
37
Q

Discussion for partial correlation

A

while x increased/decreased with Y, this relationship was eliminated/not affected when Z was controlled. These findings suggest that the relationship measured between X and Y may be explained/ not explained/ partially explained by Z.

38
Q

discussion for correlation

A

X and Y were related, with high/low Xs associated with high/low Ys