Item Analysis and Test Reliability Flashcards

1
Q

classical test theory

A

used as framework to develop and evaluate tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

true score variability

A

result of actual differences among examinees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

measurement error

A

random factors affect test performance. Ex- distractions, examinee fatigue

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

test reliability

A

consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

test-retest reliability

A

consistency of scores over time. administer, readminister later, correlate 2 sets of scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

alternate forms of reliability

A

consistency over different forms of the test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

internal consistency reliability

A

consistency over different test items. Not good for speed tests.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

coefficient alpha (Chronbach’s alpha)

A

administer test to sample and calculate average inter-item consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Kuder-Richardson 20 (KR-20)

A

coefficient used when items are dichotomously scored (correct or incorrect)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

split half reliability

A

administer test to a sample, split test in half and correlate the 2 halves (ex- half do odd, half do even questions)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Spearman Brown prophecy formula

A

determine effects of lengthening or shortening tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Inter-rater reliability

A

consistency of scores over different raters for subjectively scores tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Cohen’s kappa coefficient

A

consistency among 2 raters when rate is nominal scale

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

coefficient of concordonance

A

assess 3 or more raters when ratings are ranked

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

consensual observer drift

A

2 or more raters communicate while rating, which increased consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

content homogeneity

A

tests that are homogenous content have larger reliability coefficients than heterogenous

17
Q

range of scores

A

reliabilty larger when scores are unrestricted range

18
Q

guessing

A

reliability affected by likelihood answered questions correctly by guessing

19
Q

item difficulty

A

dichotomous difficulty level (p) is .50. P ranges from 0 to 1.0.

20
Q

item discrimination

A

(D) -1.0 to +1.0 value of .30 or higher is acceptable

21
Q

confidence interval

A

range of which score is true score

22
Q

standard error of measurement

A

confidence interval

23
Q

Item Response Theory (IRT)

A

alternative to classical test theory, how does performance on each test item relate to latent ability that is measured b the test