Midterm 1 - Reliability (Ch. 4) Flashcards

1
Q

(reliability/validity) is a condition for (reliability/validity)

A

reliability is a condition for validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

reliability means ____ of scores

A

consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are 3 assumptions of Classical Test Theory about measurement?

A
  • no measurement instrument is perfectly reliable (observed score = true score + error)
  • measurement errors are random (mean error = 0)
  • measurement errors are normally distributed
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

(random/systematic) error lowers reliability of a test

A

random!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

(T/F) systematic measurement errors lower reliability of a test over time

A

FALSE, does not lower reliability bc test is inaccurate by same amount every time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What impact do random and systematic errors have on the average of a distribution?

A
  • random: no effect on average, just variability around it
  • systematic: changes average (bias)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Classical test theory assumes that the variance of observed scores = ____ + _____

A

variance of true scores + error variance (random error)

σX(squared) = σT(squared) + σE(squared)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

The reliability coefficient (rxx’) measures the proportion of observed test scores accounted for by ____

A

variability in true scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the formula for the reliability coefficient?

A

rxx’ = true score variance/observed score variance

rxx’ = σT(squared)/σX(squared)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

the reliability coefficient is always between __ and __

A

0 and 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Classical test theory assumes that two tests are parallel if:
1.
2.
3.

A
  1. true scores are the same (which means equal observed score means)
  2. equal error variance
  3. same correlations w other tests
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Standard error of measurement (SEM) indicates the amount of _____ expected in an individual’s (true/observed) test score. SEM corresponds to _____

A

uncertainty or error; observed test score
corresponds to SD of distribution of scores one would obtain by repeatedly testing a person

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Identify each of the variables in this equation:
SEM = SDx * √(1-rxx’)

if SDx is high, SEM will be (higher/lower)
if rxx’ is high, SEM will be (higher/lower)

A

SEM = standard error of measurement
SDx = standard deviation of observed scores
rxx’ = reliability of observed scores

if SDx is high, SEM will be HIGHER
if rxx’ is high, SEM will be LOWER

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Confidence interval (CI) is a range of scores that _____

A

we feel confident will include the true score

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Identify each of the variables in the following formula:
CI = X +/- (zCI)(SEM)

A

CI = confidence interval
X = observed score
zCI = z of CI’s boundaries
SEM = standard error of measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

what is the z score (zCI) for a 95% confidence interval?

A

1.96

17
Q

What is the 95% confidence interval for an observed score of 110 with a SEM of 5?

A

CI = X +/- (zCI)(SEM)
CI = 110 +/- 1.96(5)
CI = 110 +/- 9.8
CI = approx 100 and 120

18
Q

Reliability of a test can be increased by ____

A

adding items! (as long as they’re valid)

19
Q

What does the Spearman-Brown formula do?

A

predicts effect of lengthening or shortening a test on reliability

20
Q

Identify the variables in the Spearman-Brown formula:

Predicted rxx’ = [n(rxx’)] / [1+(n-1)rxx’]

A

Predicted rxx’ = reliability of observed scores
n = degree of increase/decrease in test length (ex double = 2; half = .5)

21
Q

What are the 3 main types of reliability?

A
  • test-retest
  • internal consistency
  • interrater agreement
22
Q

What is test-retest reliability?

A
  • coefficient of stability
  • admin same test to same group @ diff occasions and correlate both sets of scores
23
Q

(T/F) when evaluating reliability we always want test-retest reliability to be high

A

FALSE - should match construct (should be lower when measuring a construct we expect to change!)

24
Q

What is another form of test-retest reliability?

A
  • coefficient of equivalence
  • make two similar versions of a test and admin both to same group in very short period, then correlate scores
25
Q

What is internal consistency?

A
  • look at how consistently examinees performed across items or subsets of items
  • one test administration!
26
Q

What are the two methods we discussed for assessing internal consistency?

A
  • Split half methods (older): correlate scores based on first half or second half of items OR based on odd/even items
  • contemporary approach: Cronbach’s alpha (avg of all possible split-half correlations); unaffected by how items are arranged in test
27
Q

(T/F) we always want internal consistency coefficients to be as high as possible

A

FALSE
- higher for narrow constructs (ex sadness)
- lower for broader constructs (ex neuroticism)
**very high scores can indicate insufficient sampling of the domain!

28
Q

What is interrater reliability?

A
  • have diff observers rate a certain behaviour
  • look @ level of agreement btw observers
29
Q

Identify the variables in the Kappa formula for 2 categories
k = (po - pe) / (1 - pe)

This formula assesses ____
k ranges from __ to __ with higher values indicating ___

A

k = kappa (interrater agreement corrected by chance)
po = observed proportions (how often raters agreed out of all observations)
pe = expected proportions

assesses interrater agreement corrected by chance
k ranges from -1 to 1; higher = higher agreement

30
Q

What is one challenge of the interrater reliability method?

A
  • can be hard to precisely define a construct
31
Q

What is the main source of error for each of these forms of reliability coefficients?
- test-retest
- parallel forms
- internal consistency
- interrater

A
  • test-retest: time sampling (length of test-retest interval)
  • parallel forms: item sampling (homogeneity of items)
  • internal consistency: item sampling (homogeneity, length)
  • interrater: observer differences
32
Q
A