Validity Flashcards
a judgment or estimate of how well a test measures what it supposed to measure
Validity
Evidence about the appropriateness of inferences drawn from test scores
Degree to which the measurement procedure measures the variables to measure
Validity
logical result or deduction
Inferences
the process of gathering and evaluating evidence about validity
Validation
yield insights regarding a particular population of testtakers as compared to the norming sample described in a test manual
Validation Studies
degree of control among variables in the study (increased through random assignment)
Internal Validity
generalizability of the research results (increased through random selection)
External Validity
focuses on individual with their unique histories and behaviors
Conceptual Validity
Means of evaluating and integrating test data so that the clinician’s conclusions make accurate statements about the examinee
Conceptual Validity
a test appears to measure to the person being tested than to what the test actually measures
Face Validity
describes a judgement of how adequately a test samples behavior representative of the universe of behavior that the test was designed to sample
more logical than statistical
Content Validity
when the proportion of the material covered by the test approximates the proportion of material covered in the course
Content Validity
a plan regarding the types of information to be covered by the items, the no. of items tapping each area of coverage, the organization of the items, and so forth
Test Blueprint
concerned with the extent to which the test is representative of defined body of content consisting the topics and processes
- panel of experts can review the test items and rate them in terms of how closely they match the objective or domain specification
- examine if items are essential, useful and necessary
Content Validity
failure to capture important components of a construct
construct underrepresentation
happens when scores are influenced by factors irrelevant to the construct
construct-irrelevant variance
developed the formula of Content Validity Ratio
Lawshe
exactly half of the experts rate the item as essential
Zero CVR
more statistical than logical
- a judgement of how adequately a test score can be used to infer an individual’s most probable standing on some measure of interestꟷthe measure of interest being criterion
Criterion Validity
standard on which a judgement or decision may be made
Criterion
Characteristics: relevant, valid, uncontaminated
Criterion Validity
occurs when the criterion measure includes aspects of performance that are not part of the job or when the measure is affected by “construct-irrelevant” (Messick, 1989) factors that are not part of the criterion construct
Criterion Contamination
If the test scores obtained at about the same time as the criterion measures are obtained; economically efficient
Concurrent Validity
measures of the relationship between test scores and a criterion measure obtained at a future time
Predictive Validity
the degree to which an additional predictor explains something about the criterion measure that is not explained by predictors already in use; used to improve the domain
Incremental Validity
related to predictive validity wherein it is defined as the degree to which an additional predictor explains something about the ____________ that is not explained by predictors already in use
criterion measure
covers all types of validity
logical and statistical
- judgement about the appropriateness of inferences drawn from test scores regarding individual standing on variable called construct
Construct Validity (Umbrella Validity)
an informed, scientific idea developed or hypothesized to describe or explain behavior; unobservable, presupposed traits that may invoke to describe test behavior or criterion performance
Construct
One way a test developer can improve the homogeneity of a test containing dichotomous items is by ____________________________ that do not show significant correlation coefficients with total test scores
eliminating items
If it is an academic test and high scorers on the entire test for some reason tended to get that particular item wrong while low scorers got it right, then the item is obviously not a ________
good one
demonstrate that scores on the test vary in a predictable way as a function of membership in a group
Method of Contrasted Groups
If a test is a _________ of a particular construct, then the scores from the group of people who does not have that construct would have different test scores than those who really possesses that construct
valid measure
if scores on the test undergoing construct validation tend to highly correlated with another established, validated test that measures the same construct
Convergent Evidence
a validity coefficient showing little relationship between test scores and/or other variables with which scores on the test being construct-validated should not be correlated
Discriminant Evidence
Factor Analysis
designed to identify factors or specific variables that are typically attributes, characteristics, or dimensions on which people may differ
Developed Factor Analysis
Charles Spearman
Used to study the interrelationships among set of variables
Identify the factor or factors in common between test scores on subscales within a particular test
Factor Analysis
estimating or extracting factors; deciding how many factors must be retained
Explanatory FA
researchers test the degree to which a hypothetical model fits the actual data
Confirmatory FA
conveys info about the extent to which the factor determines the test score or scores
can be used to obtain both convergent and discriminant validity
Factor Loading
revalidation of the test to a criterion based on another group different from the original group form which the test was validated
Cross-Validation
decrease in validity after cross-validation
Validity Shrinkage
validation of more than one test from the same group
Co-Validation
norming more than one test from the same group
Co-Norming
factor inherent in a test that systematically prevents accurate, impartial measurement
Bias
Prevention during test dev through a procedure called
Estimated True Score Transformation
numerical or verbal judgement that places a person or an attribute along a continuum identified by a scale of numerical or word descriptors known as
Rating
Rating Scale - intentional or unintentional misuse of the scale
Rating Error
Rating Scale - rater is lenient in scoring (Generosity Error)
Leniency Error
Rating Scale - rater is strict in scoring
Severity Error
Rating Scale - rater’s rating would tend to cluster in the middle of the rating scale
Central Tendency Error
One way to overcome rating errors is to use _____
rankings
Rating Scale - tendency to give high score due to failure to discriminate among conceptually distinct and potentially independent aspects of a ratee’s behavior
Halo Effect
the extent to which a test is used in an impartial, just, and equitable way
Fairness
Attempting to define the validity of the test will be futile if the test is NOT reliable
FACT