Reliability Versus Validity Lecture Dr Wofford Flashcards
Test-Retest Reliability
- Coefficient: test-retest reliability coefficient
- Common with self-report survey instruments
- ie: a subject takes an identical test on two different occasions under identical testing conditions
- Considerations:
- Test-retest intervals
- Carryover and testing effects
Very important to talk about the Test-retest intervals. Like how long since we last took the test. (like I probably cannot take the cognitive tests Dr. Esmat used in her study for a long time if ever)
Testing effect. Did I do better because I learned the test or because I improved?
Carryover and testing effects.
Internal Consistency
A type of Reliability
- Generally used with questionnaires, written exams, and interviews (more for qualitative research)
- Use correlations among all items in the scale
- Want to see some relationship among the items on an exam, interview….as they should measure the same attribute
- Reliability coefficient: correlations
Criterion-Related Validity
- Most practical and objective approach to validity testing
- Ability of one test to predict results on an external criterion
- High correlation indicates the test is valid based on the external criterion
- External criterion must be valid, reliable, independent and free from bias (Must make sure the gold standard is really a gold standard if you are going to use it for this)
- May also be called reference standard or gold standard or
- Can be tested using concurrent or predictive validity
ICC
- Reliability coefficient: intraclass correlation coefficient (ICC)
- The reliability coefficient for Rater Reliability tests (both Intra- and Inter- rater)
Relationship between validity and reliability
- Validity implies that a measurement is relatively free from error
- Inherently means that a valid measurement is also reliable
- A test can be reliable, but not valid
- A test cannot be valid, but not reliable
What is the validity counterpart to internal reliability?
Content Validity
Random Errors
Random errors: Due to chance and can affect scores in unpredictable ways
- Decrease random errors= increase reliability
- Reliability focuses on amount of random error a measurement has
- Example: fatigue
Three main Types of Reliability
- Test-retest reliability: Stability of the measuring instrument
- Rater reliability: Stability of the human observer
- Inter-rater versus Intra-rater
- Internal consistency: extent to which items measure various aspects of the same characteristic and nothing extraneous
- more for when using questionnaires
Inter-rater reliability
Inter-rater reliability: variation between 2+ raters who measure same subject
- Best if all raters measure a response during one trial
- Ensure blinding of other assessors
Reliability coefficient: intraclass correlation coefficient (ICC)
Three Sources of Measurement Error
- Individual taking the measurements
- Called tester or rater reliability
- Measuring instrument introduces error
- Variability of the measured characteristic
Validity: Convergence and Discrimination
Convergent validity: two measures believed to reflect the same underlying phenomenon will have similar results or correlate highly
- Implies that the theoretical context behind the construct will be supported when the test is administered to different groups in different places at different times
Discriminant validity: different results (low correlations) are expected from measures which are believed to assess different characteristics (Discriminant validity is when two measures that measure different things should not correlate well.)
Construct validity is related to convergent and divergent
Four Types of Measurement Validity
- Face validity
- Content validity
- Criterion-related validity
- Concurrent validity
- Predictive validity
- Construct validity
Systematic errors:
Systematic Errors: predictable errors of measurement
- Consistently overestimates or underestimates the true score
- Constant and biased
- More of a problem with validity than reliability
(Systematic error is a reliable error: For example, an uncalabrated scale is always the same amount off each time you measure it. It causes problems with validity but not reliability.)
Predictive Validity
- Establishes that the outcome of a test can be used to predict future score or outcome
- ie: GPA used to predict success in PT school or BERG balance test used to predict falls
- Criterion and target test are tested at different times
Target test = new test that is untested so far
Responsiveness to Change
- Responsiveness: ability of an instrument to detect minimal change over time
- Used to assess the effectiveness of interventions
- Minimal clinically important difference (MCID): smallest difference in a measured variable that signifies an important difference in a subject’s outcome
- Statistical versus clinically meaningful change