reliability and validity Flashcards
1
Q
what is reliability?
A
Very broadly, the dependability of a measurement instrument. Eg weighing scales, should expect to weight roughly the same/ consistently across time of day
2
Q
what is validity?
A
The property that a measurement instrument measures what it claims to measure. Eg clock measures timeHarder to establish in science
3
Q
Internal consistency (Cronbach’s α):
A
- Extent to which each of the tests measures the same thing
- Sampling distribution is unknown s bench mark values have been proposed, where 0-1, higher value = greater reliability and ability of an instrument.
4
Q
Alternate-forms reliability:
A
- Involves the construction of two tests of equal length by randomly sampling from the same domain(‘universe of items’).
- The Pearson correlation between these two tests (forms) is an estimate of the reliability
5
Q
Parallel-forms reliability:
A
- Involves the construction of two tests comprising items of the same difficulty, which leads to similar score distributions.
- The Pearson correlation between these two tests (forms) is an estimate of the reliability.
6
Q
Split-half reliability
A
- Randomly splits the test into two halves, calculates the Pearson correlation between them, and applies the Spearman-Brown prophecy formula to estimate reliability. - Still used, but not recommended as can just use computer methods now.
7
Q
Test-retest reliability:
A
- Also known as ‘temporal stability’. Same test administered on two different occasions spaced about one month or more apart.
- The Pearson correlation between the two scores is an estimate of the reliability. Note this is very different from the foregoing types of reliability.
8
Q
Construct validity:
A
- The degree to which a test measures a specified construct as determined by the interpretation of the psychological meaning of test scores and the implications of this interpretation.
- Term introduced by Cronbach and Meehl (1955).
- Construct validity is important from a scientific perspective and rests on the psychological theory underpinning an instrument.
- Not associated with a number as reliability is
9
Q
Criterion validity:
A
- The degree to which scores on a test correlate with scores on a relevant external criterion.
- It is a broad type of validity encompassing several more specific types (e.g., convergent, predictive, concurrent).
10
Q
Convergent validity
A
- The degree to which scores on a test correlate with variables they are supposed to correlate with, given the nature of the construct.
11
Q
Discriminant validity:
A
- The degree to which scores on a test do NOT correlate with (are ‘independent of’ or ‘orthogonal to’) variables they are NOT supposed to correlate with, given the nature of the construct.
- Eg happiness with IQ, if test of happiness correlates with IQ then it is not good! So if it doesn’t correlate then that is good.
12
Q
Predictive validity:
A
- The degree to which scores on a test predict future behaviour on a criterion variable.
13
Q
Concurrent validity
A
- Is based on the correlation between predictor (test) and criterion scores obtained at approximately the same time (e.g., self-reported and clinically diagnosed depression).
14
Q
Congruent validity:
A
- The degree to which a new test correlates with extant measurement instruments of the construct.
- It is a weak type of validity because extant measures may themselves have low validity.
- Good if establishing a new test if there is already a gold standard that can be used to compare to.
15
Q
Face validity:
A
- ‘experts’ review test contents to determine if they are appropriate ‘on their face’.
- Eg assessing leadership potential –> experts in the field and see if valid on face value.