Reliability of a test Flashcards
dependability or consistency of the instrument or scores obtained by the same person when re-examined with the same test on different occasions, or with different sets of equivalent items
Reliability
index of reliability, a proportion that indicates the ratio between the true score variance on a test and the total variance
Reliability Coefficient
score on an ability test is presumed to reflect not only the testtaker’s true score on the ability being measured but also the error
Classical Test Theory (True Score Theory)
refers to the component of the observed test score that does not have to do with the testtaker’s ability
Error
Factors that contribute to consistency
stable attributes
Factors that contribute to inconsistency
characteristics of the individual, test, or situation, which have nothing to do with the attribute being measured, but still affect the scores
Goals of Reliability:
EEDT
✓ Estimate errors
✓ Devise techniques to improve testing and reduce errors
useful in describing sources of test score variability
Variance
variance from true differences
True Variance
variance from irrelevant random sources
Error Variance
all of the factors associated with the process of measuring some variable, other than the variable being measured
Measurement Error
- difference between the observed score and the true score
Measurement Error
Sources of Error Variance that refer to variation among items within a test as well as to variation among items between tests
- The extent to which testtaker’s score is affected by the content sampled on a test and by the way the content is sampled is a source of error variance
Item Sampling/Content Sampling
Sources of Error Variance that testtaker’s motivation or attention, environment, etc.
Test Administration
Sources of Error Variance that may employ objective-type items amenable to computer scoring of well-documented reliability
Test Scoring and Interpretation
source of error in measuring a targeted variable caused by unpredictable fluctuations and inconsistencies of other variables in measurement process (e.g., noise, temperature, weather)
Random Error
source of error in a measuring a variable that is typically constant or proportionate to what is presumed to be the true values of the variable being measured
- has consistent effect on the true score -
SD does not change, the mean does
Systematic Error
________ refers to the proportion of total variance attributed to true variance
Reliability
The _____ the proportion of the total variance attributed to true variance, the ________ the test
greater - more reliable
___________ may increase or decrease a test score by varying amounts, consistency of test score, and thus, the reliability can be affected
Error variance
Error: Time Sampling
Test-Retest Reliability
an estimate of reliability obtained by correlating pairs of scores from the same people on two different administrations of the test
Test-Retest Reliability
appropriate when evaluating the reliability of a test that purports to measure an enduring and stable attribute such as personality trait
- established by comparing the scores obtained from two successive measurements of the same individuals and calculating a correlated between the two set of scores
Test-Retest Reliability
the longer the time passes, the greater likelihood that the reliability coefficient would be insignificant
Test-Retest Reliability
happened when the test-retest interval is short, wherein the second test is influenced by the first test because they remember or practiced the previous test = inflated correlation/overestimation of reliability
Carryover Effects
scores on the second session are higher due to their experience of the first session of testing
Practice Effect
test-retest with ______ interval might be affected of other extreme factors, thus, resulting to _____ correlation
longer - low
problems in absences in second session (just remove the first tests of the absents)
Mortality
statistical tool Test-Retest Reliability
Pearson R, Spearman Rho
Error: Item Sampling (Immediate), Item Sampling changes over time (delaued)
Parallel Forms/Alternate Forms Reliability
established when at least two different versions of the test yield almost the same scores
- has the most universal applicability
Parallel Forms/Alternate Forms Reliability
each form of the test, the means, and the variances, are EQUAL; same items, different positionings/numberings
Parallel Forms
simply different version of a test that has been constructed so as to be parallel
Alternate Forms
- technique to avoid carryover effects for parallel forms, by using different sequence for groups
- can be administered on the same day or different time
Counterbalancing
most rigorous and burdensome, since test developers create two forms of the test - main problem: difference between the two test - test scores may be affected by motivation, fatigue, or intervening events - means and the variances of the observed scores must be equal for two forms - Statistical Tool: Pearson R or Spearman Rho
Counterbalancing
used when tests are administered once - consistency among items within the test - measures the internal consistency of the test which is the degree to which each item measures the same construct
Internal Consistency (Inter-Item Reliability)
Error: Item Sampling Homogeneity
Internal Consistency (Inter-Item Reliability)
measurement for unstable traits
- if all items measure the same construct, then it has a good internal consistency
Internal Consistency (Inter-Item Reliability)
if a test contains items that measure a single trait (unifactorial)
Homogeneity