Validity and Reliability Flashcards
What are the three aspects of reliability?
stability, internal consistency and interrater agreement
What is stability?
Test-retest reliability: the extent to which the same results are obtained on repeated applications
What is internal consistency?
the extent to which all of the items of the measure address the same underlying concept. Items that propose to measure the same general construct should produce similar scores.
What is interrater agreement?
The extent to which the results are in agreement when different individuals administer the same instrument to the same individuals/groups
What are the two types of interrater reliability?
- intra-rater: indicates how consistently a rater administers and scores a measure
- inter-rater: how well more than one rater agree in the way they administer and score the measure
What is validity?
The degree to which an instrument measures what it intends to measure
What are the types of validity?
face, content, consensual, criterion, construct and predicitive
What is face validity?
The relevance of the measurement - do the questions yield relevant info to the topic investigated? The perceived relevance to the test taker.
What may indicate poor face validity?
many ‘don’t know’ answers in a questionnaire
What is content validity?
The extent to which a measure represents all facets of a given phenomena or covers the whole concept
What is consensual validity?
a number of experts agree that the measure is valid
What is criterion validity?
concurrent and predictive validity. The extent to which the test agrees with a gold standard test known to represent the phenomena accurately.
What is predictive validity?
The extent to which the measurement can depict what may occur in the future
What is construct validity?
The extent to which an assessment measures a theoretical construct
What are the types of construct validity?
- convergent: scale should be related to variables and other measures of the same construct
- discriminative: demonstrate that is discriminates between groups and individuals
- factorial: items go together to create factors (correlation between test and major factors)
- discriminant: new test should not correlate with dissimilar, unrelated constructs
What is external validity?
The tool’s generalisability to other settings
What research evidence is needed for construct validity?
- hypothesis about relationship between variables
- select test items of behaviors that represent the construct
- collect data to test hypothesis
- determine if data supports hypothesis
What is a Rasch analysis used for?
determines unidimensionality of test items
What are the two types of criterion validity?
- concurrent - with established measure
- predictive - with future outcome
What research evidence is needed for predictive validity?
- identify criterion behavior and population sample
- administer test and keep until criterion data is available
- obtain measure of performance on each criterion
- determine strength of relationship