Title: Linear%20Correlation
1Linear Correlation
2Perfect Correlation
- 2 variables x and y are perfectly correlated if
they are related by an affine transform - y ax b
- The correlation is positive if agt0 and negative
if alt0. - By corollary, 2 variables are perfectly
positively correlated if and only if each pair of
corresponding values has the same z-score. - If the 2 variables are perfectly negatively
correlated, corresponding z-scores will be equal
in magnitude but opposite in sign.
3Pearsons r
4Scatterplots
5Pearsons r only measures linear dependence
- Two variables can have low correlation and still
be highly dependent.
6Higher-Order Models
7Pearsons r depends on the range of the variables
under study
- r2 measures the proportion of variance in one
variable accounted for by the other. - If the range of variable X is restricted, it will
account for less of the variance in Y.
8Pearsons r is Sensitive to Outliers
9Standard Definition of Correlation (Population)
10Standard Definition of Correlation (Sample)
11Alternative (Equivalent) Formula
12Computational Formula
For a population
For a sample
13Example 6130A 2005-2006 Assignment Marks
14End of Lecture 7
15Correlation and the Power of Matched Tests
16Correlation and the Power of Matched t-tests
- Now that we understand correlation, we can better
understand the power of matched t-tests when
scores in the two conditions are correlated.
17Recall formulae for standard error for
independent and matched tests
18Knowing the expected std error, we can estimate
the expected t-value
19The power of matched t-tests
- Large positive correlations between scores in the
two conditions will mean a greater expected
t-score for the matched design. - But keep in mind that the critical value for the
matched design will be somewhat larger as well,
due to a smaller df. - Which test is more powerful is decided by the
exact tradeoff between these two effects.
20Applying Correlation Analysis
21Adjusted Correlation Coefficient
22Testing Pearsons r for Significance
23Underlying Assumptions (For Inference)
- Independent random sampling
- Bivariate normal distribution
24Applications of Pearsons r
- Measuring reliability and validity
- Examples
- e.g., test-retest reliability
- Split-half reliability
- Inter-rater reliability
- Criterion validity of self-report (correlate
self-report against behavioural measure) - Correlation between tests that are supposed to
measure the same thing. - Correlation between algorithmic model and human
responses in behavioural studies. - Measuring relationships between variables
(correlational studies) - e.g., frequency of cannabis and alcohol use
- Measuring relationships between IVs and DVs
(experimental studies, when IV on interval/ratio
scale - e.g., exam performance as a function of alcohol
consumption on previous night.
25Power Analysis for Pearsons r
26Confidence Intervals for Pearsons r
- Pearsons r is bounded on -1..1.
- Consequently, sampling distribution for r is not
normal. - Sampling distribution for rgt0 is negatively
skewed. - Sampling distribution for rlt0 is positively
skewed. - Thus confidence intervals are generally not
symmetric.
27Fisher Transform
- Fisher transform (Appendix r') Method for
symmetrizing r to facilitate calculation of
confidence interval using standard normal table.
28Confidence Intervals on r
29End of Lecture 8
30Testing Difference of Pearson Correlations from 2
Independent Samples
- Converting the skewed r distribution to an
(approximately) normal distribution allows
straightforward two-sample testing
31Example
N44
N43