Chapter 7 - 8 Flashcards
usefulness or practical value of testing to improve efficiency
Utility
used to refer to the usefulness or practical value of a training program or
intervention
Utility
Factors that affect a test’s utility
- Psychometric Soundness
- Cost
- Benefits
Gives us the practical value of both the scores (reliability
and validity)
Psychometric Soundness
They tell us whether decisions are cost-effective
Psychometric Soundness
A test must be valid to be useful, but a valid test is not always a useful test, especially if testtakers do not follow test directions
True
It refers to disadvantages, losses or expenses in both economic and noneconomic terms
Cost
It refers to profits, gains or advantages
Benefit
It is a family of techniques that entail a cost-benefit analysis designed to yield information relevant to a decision about the usefulness and/or practical value of a tool of assessment
Utility Analysis
provide an indication of likelihood that a testtaker will score within some interval of scores on a criterion measure – an
interval may be categorized as “passing”, “acceptable” or “failing”
Expectancy Table/Chart
estimate of the percentage of employees hired by a particular test who will be successful to their jobs
Taylor-Russell Tables
used for obtaining the difference between the means of the selected and unselected groups to derive an index of what the test is
adding to already established procedure
Naylor-Shine Tables
A formula used to calculate the dollar amount of a utility gain resulting from the
use of a particular selection instrument under specified conditions
Brodgen-Cronbach-Gleser Formula
an estimate of the benefit (monetary/otherwise) of using a particular
test or selection method
Utility gain
a body of methods used to quantitatively evaluate selection procedures,
diagnostic classifications, therapeutic interventions or other assessment or
intervention-related procedures in terms of how optimal they are (most typically
from a cost-benefit perspective)
Decision Theory
a correct classification
hit
a qualified driver is hired; an unqualified driver is not hired
It is a hit
an incorrect classification; a mistake
miss
a qualified driver is not hired; an unqualified driver is hired
It is a miss
the proportion of people that an assessment tool accurately identified
as possessing a particular variable
hit rate
the proportion of qualified drivers with a passing score who actually
gain permanent employee status; the proportion of unqualified drivers with a
failing score who did not gain permanent status
This is a hit rate
the proportion of people that an assessment tool inaccurately identified
as possessing a particular variable
miss rate
the proportion of drivers whom inaccurately predicted to be qualified;
the proportion of drivers whom inaccurately predicted to be unqualified
this is a miss rate
falsely indicates that the testtaker possesses a particular variable; example: a driver who is hired is not qualified
false positive
falsely indicates that the testtaker does not possess a particular variable; the assessment tool says to not hire but driver would have been rated as qualified
false negative
Some practical considerations
The Pool of Job Applicants
The Complexity of the Job
The Cut Score in Use
a (usually numerical) reference point derived as a result of a judgment and used to divide a set of data into two or more classifications, with some action to be taken
or some inference to be made on the basis of these classifications
Cut Score/Cutoff Score
dictate what sort of information will be required as well as the
specific methods to be used
objective of utility analysis
Used to measure costs vs. benefits
Expectancy Data
- Based on norm-related considerations rather
than on the relationship of test scores to a
criterion - Also called norm-referenced cut score
- Ex.) top 10% of test scores get A’s
- normative
Relative cut score
- set with reference to a judgment concerning a minimum level of proficiency required to be included in a particular classification.
- Also called absolute cut score
- criterion
Fixed cut score
using two or more cut scores with reference to one predictor for the purpose of categorizing
testtakers
Multiple cut scores
Ex.) having cut score that marks an A, B, C etc.
all measuring same predictor
Multiple cut scores
the achievement of a particular cut score on one test is necessary in order to
advance to the next stage of evaluation in the selection process
Multiple-stage or Multi Hurdle
written application->group interview->personal interview
Multiple-stage or Multi Hurdle
assumption is made that high scores on one attribute can compensate for low scores on another attribute
Compensatory model of selection
Who devised Angoff method?
William Angoff
Who devised Angoff method?
William Angoff
a way to set fixed cut scores that entails averaging the judgments of experts; must have high inter-rater reliability
Angoff Method
a system of collecting data on a predictor of interest from groups known to
possess (and not to possess) a trait, attribute or ability of interest
Know Groups Method/Method of Contrasting Groups
a system of collecting data on a predictor of interest from groups known to
possess (and not to possess) a trait, attribute or ability of interest
Know Groups Method/Method of Contrasting Groups
a cut score is set on the test that best discriminates the high performance from low
performers
Know Groups Method/Method of Contrasting Groups
-in order to “pass” the test, the testtaker must answer items that are considered that has some minimum level of difficulty, which is determined by the experts and serves as the cut score
Item Response Theory (IRT)-Based Methods
- Based on testtaker’s performance across all items on a test
- Some portion of test items must be correct
IRT Based Method
a technique for identifying cut scores based on the number of positions to be
filled
Method of Predictive Yield
a family of statistical techniques used to shed light on the relationship between certain variables and two or more naturally occurring groups
Discriminant Analysis
determining difficulty level reflected by cut score
Item mapping method
test items are listed, one per page, in ascending level of
difficulty. An expert places a bookmark to mark the divide which separates testtakers who have acquired minimal knowledge, skills, or abilities and those that have not. Problems include training of experts, possible floor and ceiling effects, and the optimal length of item booklets
Bookmark-method
Steps in Test Development
- TEST CONCEPTUALIZATION
- TEST CONSTRUCTION
- TEST TRYOUT
- ITEM ANALYSIS
- TEST REVISION
Conception of idea by the test developer
Test Conceptualization