UNIT 6 Flashcards
The extent to which a score from a selection measure is stable and free
from error
A. Validity
B. Consistency
C. Reliability
C. Reliability
True or False: If a score from a measure is not stable or error-free, it is not useful.
True
Each one of several people take the same test
twice.
A. Test-retest reliability
B. Temporal stability
C. Alternate-Forms Reliability
A. Test-retest reliability
The consistency of test scores across time.
A. Test-retest reliability
B. Temporal stability
C. Alternate-Forms Reliability
B. Temporal stability
True or False: the time interval should be long enough so that the specific test answers have not been memorized, but short enough so that the person has not changed significantly.
True
True or False: Typical time intervals between test administrations range from three weeks to three years
False; three days to three months
The typical test-retest reliability coefficient for tests used by organizations is
A. 0.81
B. 0.92
C. 0.86
D. 0.79
C. 0.86
amount of anxiety that an individual normally has all the time,
A. Trait anxiety
B. State anxiety
A. Trait anxiety
the amount of anxiety an individual has at any given moment.
A. Trait anxiety
B. State anxiety
B. State anxiety
The extent to which two forms of the same test are similar.
A. Alternate-forms reliability
B. Counterbalancing
C. Form stability
A. Alternate-forms reliability
A method of controlling for order effects by giving half of a sample Test A first, followed by Test B, and giving the other half of the sample Test B first, followed by Test A.
A. Alternate-forms reliability
B. Counterbalancing
C. Form stability
B. Counterbalancing
True or False: Applicants retaking the same cognitive ability test will increase their scores about twice as much as applicants taking an alternate form of the cognitive ability test
True
The extent to which the scores on two forms of a test are similar.
A. Alternate-forms reliability
B. Counterbalancing
C. Form stability
C. Form stability
True or False: In alternate-forms reliability, the time interval should be as long as possible.
True
The extent to which responses to the same test items are consistent.
A. Item stability
B. Item homogeneity
C. Kuder-Richardson Formula 20
A. Item stability
The extent to which test items measure the same construct.
A. Item stability
B. Item homogeneity
C. Kuder-Richardson Formula 20
B. Item homogeneity
A statistic used to determine internal reliability of tests that use items with dichotomous answers
A. Item stability
B. Item homogeneity
C. Kuder-Richardson Formula 20
C. Kuder-Richardson Formula 20
A form of internal reliability in which the consistency of item responses is determined by comparing scores on half of the items with scores on the other half of the items.
A. Split-half method
B. Spearman-Brown prophecy formula
C. Coefficient alpha
A. Split-half method
Used to correct reliability coefficients resulting from the split-half method.
A. Split-half method
B. Spearman-Brown prophecy formula
C. Coefficient alpha
B. Spearman-Brown prophecy formula
A statistic used to determine internal reliability of tests that use interval or ratio scales.
A. Split-half method
B. Spearman-Brown prophecy formula
C. Coefficient alpha
C. Coefficient alpha
True or False: K-R 20 is used for dichotomous items, whereas the coefficient alpha can be used not only for dichotomous items but also for tests containing interval and ratio items
True
The extent to which two people scoring a test agree on the test score, or the extent to which a test is scored correctly.
A. Scorer reliability
B. Validity
A. Scorer reliability
When deciding whether a test demonstrates sufficient reliability, two factors must be considered:
A. The contents of the test
B. The magnitude of the reliability coefficient
C. The people who will be taking the test.
D. The validity of the test
B&C
The degree to which inferences from test scores are justified by the evidence.
A. Scorer reliability
B. Validity
B. Validity
The potential validity of a test is limited by its
A. Consistency
B. Accuracy
C. Reliability
D. Utility
C. Reliability
The extent to which tests or test items sample the content that they are supposed to measure.
A. Content Validity
B. Criterion Validity
C. Construct Validity
D. Face Validity
A. Content Validity
The extent to which a test score is related to some measure of job performance
A. Content Validity
B. Criterion Validity
C. Construct Validity
D. Face Validity
B. Criterion Validity
Most theoretical of all the validity types
A. Content Validity
B. Criterion Validity
C. Construct Validity
D. Face Validity
C. Construct Validity
The extent to which a test actually measures the construct that it purports to measure.
A. Content Validity
B. Criterion Validity
C. Construct Validity
D. Face Validity
C. Construct Validity
The extent to which a test appears to be valid
A. Content Validity
B. Criterion Validity
C. Construct Validity
D. Face Validity
D. Face Validity
A measure of job performance, such as attendance, productivity, or a supervisor rating.
A. Criterion
B. Concurrent Validity
C. Predictive Validity
A. Criterion
A form of criterion validity that correlates test scores with measures of job performance for employees currently working for an organization
A. Criterion
B. Concurrent Validity
C. Predictive Validity
B. Concurrent Validity
Given to a group of employees who are already on the job
A. Criterion
B. Concurrent Validity
C. Predictive Validity
B. Concurrent Validity
A form of criterion validity in which test scores of applicants are compared at a later date with a measure of job performance
A. Criterion
B. Concurrent Validity
C. Predictive Validity
C. Predictive Validity
Given to job applicants who are going to be hired
A. Criterion
B. Concurrent Validity
C. Predictive Validity
C. Predictive Validity
A narrow range of performance scores that makes it difficult to obtain a significant validity coefficient
A. Restricted Range
B. Validity Generalization (VG)
C. Synthetic Validity
A. Restricted Range
The extent to which inferences from test scores from one organization can be applied to another organization
A. Restricted Range
B. Validity Generalization (VG)
C. Synthetic Validity
B. Validity Generalization (VG)
A form of validity generalization in which validity is inferred on the basis of a match between job components and tests previously found valid for those job components
A. Restricted Range
B. Validity Generalization (VG)
C. Synthetic Validity
C. Synthetic Validity
A form of validity in which test scores from two contrasting groups “known” to differ on a construct are compared.
A. Known-group Validity
B. Barnum Statements
C. Mental Measurements Yearbook (MMY)
A. Known-group Validity
Statements, such as those used in astrological forecasts, that are so general that they can be true of almost anyone.
A. Known-group Validity
B. Barnum Statements
C. Mental Measurements Yearbook (MMY)
B. Barnum Statements
A book containing information about the reliability and validity of various psychological tests.
A. Known-group Validity
B. Barnum Statements
C. Mental Measurements Yearbook (MMY)
C. Mental Measurements Yearbook (MMY)
Most common source of test information
A. Known-group Validity
B. Barnum Statements
C. Mental Measurements Yearbook (MMY)
C. Mental Measurements Yearbook (MMY)
Should be considered if two or more tests have similar validities
A. Cost
B. Group testing
C. Computer-adaptive testing (CAT)
A. Cost
usually less expensive and more efficient although important information may be lost
A. Cost
B. Group testing
C. Computer-adaptive testing (CAT)
B. Group testing