Test Construction Flashcards
Reliability
Amt of consistency, repeatability, and dependability in scores on a given test
Classical Test Theory
Obtained score in a combo of truth & error
X=T + E
True score variability vs Error Variability
due to real diffs in ability or knowledge in test takers;
caused by chance or random factors
Range of reliability coefficient for a test
- 00 to 1.00
* Interp directly as the percentage of true score variability
Minimum acceptable Reliability
0.80
Sources of Error in Test Impacting Reliability
- Content Sampling- items that by chance tap into test taker’s knowledge, or don’t
- Time Sampling-impact of giving tests at 2 diff points in time
- Test Heterogeneity
Factors Affecting Reliability
# of items (more items--> higher rel) Homogeneity of items (items tap similar content areas--> higher rel) Range of Scores (unrestricted range maximizes reliability) Ability to guess (easier to guess, lower reliability b/c no longer assessing true knowledge)
4 Estimates of Reliability
Test-Retest
Parallel Forms
Internal consistency
Interrater Reliability
Coefficient of stability
for test-retest reliability
Major source of error is time sampling (time interval btwn administrations)
Coefficient of equivalence
for parallel forms reliability
major sources of error include time sampling & content sampling
Spearman Brown Prophecy Formula
used for split half reliability computation to compensate for impact of reducing # of items on each half of the test
Split half reliability is inappropriate for ___ and ___ tests
Speeded (all items easy, but about how many comp in allotted time)
Power (subjects need to complete all items of varying difficulty, like EPPP)
Ways to measure internal consistency reliability
Split Half
Kuder-Richardson (used when items are scored dichotomously) & Cronbach’s Alpha (used when items are likert scale)
Standard Error of Measurement
standard deviation of a theoretically normal distribution of test scores obtained by one ind on equivalent tests
Range of values for Standard Error of the Measurement
0.0 to standard deviation of the test (when test is totally unreliable, SEM equal the standard deviation of the test)
Due to error in measurement, subjects scores often reported using ___ ___
confidence bands (standard error of measurement corresponds to z scores)
68% btwn -1 and +1 z score
95% -2 and + 2
99% -3 and + 3
Validity of a test
meaningfulness, usefulness, or accuracy of a test
3 types of validity
Content
Criterion related
Construct
Content validity
how adequately does a test sample a particular content area
-quantified by asking a panel of experts
no numerical validity coefficient
Criterion related validity
how adequately a test score can be used to predict criterion outcome
ex- how well do SAT scores predict college GPA
How is criterion related validity calculated?
Pearson r; coefficient range of -1.0 to + 1.0
as low as .20 freq considered acceptable
*Correlation squared tells how much variability in the criterion is accounted for by performance on the predictor
2 subtypes of criterion related validity
Concurrent-predictor & criterion measured at about the same time
Predictive-delay btwn meas of predictor & criterion
Standard Error of the Mean vs
Standard Error of Measurement vs
Standard Error of Estimate
Mean-the means of many many samples should be normally distributed, the amt of variation/deviation in these group means is standard error of the mean (umbrella of inferential stats)
Measurement-under umbrella of reliability; standard deviation of theoretically normal distribution of test scores obtained by ONE Person on equivalent tests
Estimate- under umbrella of validity; standard deviation of a theoretically normal distribution of criterion scores obtained by ONE person measured repeatedly
Standard Error of Estimate
standard deviation of a theoretically normal distribution of criterion scores obtained by ONE person measured repeatedly
min value of 0.0, max value is SD of the criterion
if perfect predictor SEE is 0.0; if no predictive value, SEE=standard deviation of the criterion
3 Applications of Criterion Related Validity Coefficient
Expectancy Tables
Taylor Russell Tables
Decision Making Theory
Expectancy Tables
List the probability that a person’s criterion score will fall in a specified range based on the range in which a person’s predictor score fell
Taylor Russell Tables
Show how much more accurate selection decisions are when using a particular predictor test as opposed to no predictor test
Speaks to INCREMENTAL VALIDITY (amt of improvement in success rate resulting from use of a predictor test)
3 variables affecting incremental validity (and used in Taylor Russell Tables)
Criterion related validity coefficient of the predictor test Company's base rate (rate of selecting succ employees w/o a predictor test) Selection Ratio (proportion of available openings to available applicants)
Item Response Curve
provides one to 3 pieces of info re: test item:
it’s difficulty (rep by position of the curve-left vs. right)
it’s ability to distinguish between high & low scorers (the slope of the curve)
probability of getting item right by guessing (y intercept)
Item Response Theory
Based on complex math assumptions, including invariance of item parameters, which holds that charac of items should be same for all theoretically equivalent groups of subjects chosen from the same population
For example, items should not have diff charac for minority & non minority groups
*used to dev comp programs that tailor tests to the individual’s ability level
Other assumptions of Item Response Theory
assumes test items measure a latent trait
usually has little sig unl;ess working w/very large samples
Incremental Validity is greatest when:
base rate is moderate, validity coefficient is high, & selection ratio is low