Chapter 4: Reliability Flashcards

1
Q

Reliability

A

Consistency or stability of test scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Factors that impact reliability

A
When the test is administered
Items selected to be included
External distractions (ex- noise)
Internal distractions (ex- fatigue)
Person administering test
Person scoring test
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Two components of score

A
True score (representative of true knowledge or ability)
Error score
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Systematic error

A

Error resulting from receiving a different set of instructions for test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Classical test theory equation

A

Xi=T+E
Xi- obtained score
T- true score
E- error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What measurement error reduces

A

Usefulness of measurement
Generalizability of test results
Confidence in test results

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Content sampling error

A

Difference between sample of items on test and total domain of items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How good sampling affects error

A

Reduces it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Largest source of measurement error

A

Content sampling error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Time sampling error

A

Random fluctuations in performance over time

Can be due to examinee (fatigue, illness, anxiety, maturation) or due to environment (distractions, temperature)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Inter-rater differences

A

When scoring is subjective, different scorers may score answers differently

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Clerical errors

A

Adding up points incorrectly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Reliability (mathematic definition)

A

Symbol: rxx

Ratio of true score variance to total score variance (number from 0 to 1, where 0 is total error and 1 is no error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Reliability equation

A

rxx= (sigma^2T)/(sigma^2X)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Reliability’s relation to error

A

Greater the reliability, the less the error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What reliability coefficients mean

A

rxx of 0.9: 90% of score variance is due to true score variance

17
Q

Test-retest reliability

A

Administer the same test on 2 occasions
Correlate the scores from both administrations
Sensitive to sampling error

18
Q

Things to consider surrounding test-retest reliability

A

Length of interval between testing
Activities during interval (distraction or not)
Carry-over effects from one test to next

19
Q

Alternate-form reliability

A

Develop two parallel forms of test
Administer both forms (simultaneously or delayed)
Correlate the scores of the different forms
Sensitive to content sampling error (simultaneous and delayed) and time sampling error (delayed only)

20
Q

Things to consider surrounding alternate-form reliability

A

Few tests have alternate forms

Reduction of carry-over effects

21
Q

Split-half reliability

A

Administer the test
Divide it into 2 equivalent halves
Correlate the scores for the half tests
Sensitive to content sampling error

22
Q

Things to consider surrounding split-half reliability

A

Only 1 administration (no time sampling error)
How to split test up
Short tests have worse reliability

23
Q

Kuder-Richardson and coefficient (Cronbach’s) alpha

A

Administer test
Compare each item to all other items
Use KR-20 for dichotomous answers and Cronbach’s alpha for any type of variable
Sensitive to content sampling error and item heterogeneity
Measures internal consistency

24
Q

Inter-rater reliability

A

Administer test
2 individuals score test
Calculate agreement between scores
Sensitive to differences between raters

25
Q

Composite scores

A

Scores that are combined to form a combined score

Reliability of these is usually better than their individual parts

26
Q

Difference scores

A

Calculated difference between 2 scores
Reliability of these is usually lower than their individual parts (information is lost: only can see change, not initial baseline)

27
Q

Choosing a reliability test to use

A

Multiple administrations: test-retest reliability

One administration: homogeneous content uses coefficient alpha and heterogeneous content uses split-half coefficient

28
Q

Factors to consider when evaluating reliability coefficients

A

Construct being measured
Time available for testing
How the scores will be used
Method of estimating reliability

29
Q

High-stake decision tests: reliability coefficient used

A

Greater than 0.9 or 0.95

30
Q

General clinical use: reliability coefficient used

A

Greater than 0.8

31
Q

Class tests and screening tests: reliability coefficient used

A

Greater than 0.7

32
Q

How to improve reliability

A

Increase number of test items
Use composite scores
Develop better items
Standardize administration

33
Q

Standard error of measurement (SEM)

A

Standard deviation of test administered to the same individual an infinite number of times
Useful when interpreting test scores
When reliability increases, this decreases

34
Q

How to calculate confidence intervals

A

Use SEM and SD

35
Q

Relationship between reliability and confidence interval

A

Reliability increases, confidence interval decreases

36
Q

Test manuals/researchers report: information included

A

Internal consistency
Test-retest
Standard error of measurement (SEM)
Information on confidence intervals

37
Q

Generalizability theory

A

Shows how much variance is associated with different sources of error