Chapter 5 Flashcards
Refers to consistency in measurement.
Not an all-or-none matter.
A test may be reliable
in one context and unreliable in another.
RELIABILITY
Index of reliability, a proportion
that indicates the ratio between the true score variance on a
test and the total variance.
Reliability coefficient
Score on an ability test is presumed to reflect
not only the testtaker’s true score on the ability being measured but also error.
X= T + E
Classical test theory
A statistic useful in describing sources of test score variability
Variance
2 types of measurement error
Systematic
Random
Source of error in measuring a targeted variable
caused by unpredictable fluctuations and inconsistencies of other
variables in the measurement process.
Random error
Source of error in measuring a variable that
is typically constant or proportionate to what is presumed to be the true
value of the variable being measured.
Doesn’t affect score consistency
Systematic error
Source of variance which refers to variation among items
within a test as well as to variation among items between tests.
Differences are sure to be found in the way the items are worded
and in the exact content sampled.
Item sampling
or Content sampling
The testtaker’s reactions to those influences are the source of
one kind of error variance:
Sources of error in test administration
Test environment - physical setting where a test is administered
Testtaker variables - factors related to the individual taking the test
Examiner-related variables - factors related to the person administering the test
2 possible sources of error variance under test scoring and interpretation
Scorers - subjectivity involved
Scoring systems - glitch if computer scoring
3 potential sources of nonsystematic error in assessment
situation
Forgetting
Failing to notice abusive
behavior
Misunderstanding instructions regarding
reporting
Estimate of reliability obtained by
correlating pairs of scores from the same people on two different
administrations of the same test.
Test-retest reliability
The
estimate of test-retest reliability when the interval between testing is greater than six months
Coefficient
of stability
The degree of the relationship between various forms of a test can be
evaluated by means of _______________; often referred to as coefficient of equivalence.
Alternate-forms or parallel-forms
coefficient of reliability
Exist when, for each form of the test, the
means and the variances of observed test scores are equal.
Parallel forms
Simply different versions of a test that have
been constructed so as to be parallel.
Alternate forms
Obtained by correlating two pairs of scores obtained from
equivalent halves of a single test administered once.
Internal consistency estimate of reliability
or
Estimate of interitem consistency
Step 1. Divide the test into equivalent halves.
Step 2. Calculate a Pearson r between scores on the two halves
of the test.
Step 3. Adjust the half-test reliability using the Spearman-Brown formula
3 steps in the computation of coefficient of split-half reliability
3 acceptable ways to split a test
Randomly assign
Odd-even reliability
By content
Assign odd-numbered
items to one half of the test and even-numbered items to the
other half
Odd-even reliability
Each half contains items equivalent with respect to content and
difficulty
By content
Randomly assign items
to one or the other half of the test.
Randomly assign
Formula that allows a test developer or user to
estimate internal consistency reliability from a correlation of two halves of
a test.
Spearman-Brown
Degree of correlation among all
the items on a scale.
Inter-item consistency