Test Construction Flashcards
Classical test theory
- measurement that is used to develop and evaluate tests
- framework
- assumes that obtain test scores (X) are due to the combination of true variability (T) and measurement error (E)
X=T+E
True score variability
- result of actual differences among examinees in regards to whatever the test is measuring
- assumed to be consistent
Measurement error
- due to random factors that affect test performance of examinees in unpredictable ways
Example: distractions, ambiguously worded test items, and examinee fatigue
Test reliability
- The extent to which a test provides consistent information
- there are several methods for evaluating this and each is appropriate for different circumstances
+ Most methods provide a reliability coefficient
Reliability coefficient
- type of correlation coefficient
- ranges from 0 to 1.0
- designated with the letter r with the subscript of two of the same letters or numbers
- always interpreted as the the direct amount of variability obtained in test scores that’s due to true score variability
- .70 or higher is considered the minimally acceptable level, but 90 is usually required for higher stakes tests used to select employees, assign diagnoses, or other important decisions about individuals
The acceptable level of reliability
- depends on the type of test and its purpose
Standardized cognitive ability tests versus personality test
- cognitive ability test have a higher reliability coefficient
Standardized cognitive ability tests versus personality test
- cognitive ability test have a higher reliability coefficient
Test retest reliability
- provides information on the consistency of scores over time
- involves administering the test to a sample of examinees and readministering the test to the same examinees at a later time than correlating the two sets of scores
- useful for tests that are designed to measure a characteristic that is stable over time
Alternative forms reliability
,- provides information about the consistency of scores over different forms of the test and when the second form is administered at a later time, the consistency of scores over time
- involves administering one forum to a sample of examinees administering the other forms to the same sample of examinees and correlating the sets of scores
- important whenever a test has more than one form
Internal consistency reliability
- provides information on the consistency of scores over different test items
- useful for tests that are designed to measure a single content, domain or aspect of behavior
,- not useful for speed tests because it overestimates their reliability
Speed test
- test retest reliability and alternative forms reliability are appropriate
Coefficient alpha
-aka cronbach’s alpha
- involves administering the test to a sample of examinees and calculating the average inter-item consistency
Coefficient alpha
-aka cronbach’s alpha
- involves administering the test to a sample of examinees and calculating the average inter-item consistency
Kuder Richardson’s 20( kr-20)
- alternative to coefficient alpha
- can be used when test items are dichotomously scored
Split half reliability
- involves administering the test to a sample of examinees splitting the test in half, usually in terms of even and odd numbers and correlating the scores on the two halves