Chapter 6 EVALUATING S ELECTION TECHNIQUES AND D ECISIONS Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

The extent to
which a score from a test or
from an evaluation is consistent
and free from error.

A

Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The
extent to which repeated administration of the same test will
achieve similar results

A

Test-retest reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The
consistency of test scores
across time

A

Temporal stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The extent to
which two forms of the same
test are similar

A

Alternate-forms
reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

A
method of controlling for order
effects by giving half of a sample
Test A first, followed by Test B,
and giving the other half of the
sample Test B first, followed by
Test A

A

Counterbalancing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

The extent to
which the scores on two forms
of a test are similar

A

Form stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

The extent to
which responses to the same
test items are consistent

A

Item stability or katatagan

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

The
extent to which test items
measure the same construct

A

Item homogeneity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

A
statistic used to determine
internal reliability of tests that
use items with dichotomous
answers (yes/no, true/false

A

Kuder-Richardson
Formula 20 (K-R 20)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

A form of
internal reliability in which the
consistency of item responses is
determined by comparing scores
on half of the items with scores
on the other half of the items

A

Split-half method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Used to
correct reliability coefficients
resulting from the split-half
method

A

Spearman-Brown
prophecy formula

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

A statistic
used to determine internal
reliability of tests that use
interval or ratio scales

A

Coefficient alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The extent
to which two people scoring a
test agree on the test score, or
the extent to which a test is
scored correctly

A

Scorer reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

The degree to which
inferences from test scores are
justified by the evidence

A

Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The extent
to which tests or test items
sample the content that they are
supposed to measure

A

Content validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

The extent
to which a test score is related
to some measure of job
performance

A

Criterion Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

A measure of job
performance, such as attendance,
productivity, or a supervisor rating

A

Criterion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

A form
of criterion validity that correlates
test scores with measures of job
performance for employees currently working for an organization

A

Concurrent validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

A form
of criterion validity in which test
scores of applicants are compared at a later date with a
measure of job performance

A

Predictive validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

A narrow
range of performance scores that
makes it difficult to obtain a
significant validity coefficient

A

Restricted range

21
Q

The extent to which inferences from test scores from
one organization can be applied
to another organization

A

Validity generalization
(VG)

22
Q

A form of
validity generalization in which
validity is inferred on the basis of
a match between job components and tests previously found
valid for those job components

A

Synthetic validity

23
Q

The
extent to which a test actually
measures the construct that it
purports to measure

A

Construct validity

24
Q

A
form of validity in which test
scores from two contrasting
groups “known” to differ on
a construct are compared

A

Known-group validity

25
Q

The extent to
which a test appears to be valid

A

Face Validity

26
Q

Statements, such as those used
in astrological forecasts, that are
so general that they can be true
of almost anyone

A

Barnum statements

27
Q

A book
containing information about the
reliability and validity of various
psychological tests

A

Mental Measurements
Yearbook (MMY

28
Q

A type of test
taken on a computer in which
the computer adapts the difficulty level of questions asked to
the test taker’s success in
answering previous questions

A

Computer-adaptive
testing (CAT

29
Q

A
series of tables based on the
selection ratio, base rate, and
test validity that yield information about the percentage of
future employees who will be
successful if a particular test
is used

A

Taylor-Russell tables

30
Q

The
percentage of applicants an
organization hires

A

Selection ratio

31
Q

Percentage of
current employees who are
considered successful

A

Base rate

32
Q

A utility method
that compares the percentage of
times a selection decision was
accurate with the percentage of
successful employees

A

Proportion of correct
decisions

33
Q

Tables that
use the base rate, test validity,
and applicant percentile on a
test to determine the probability
of future success for that
applicant

A

Lawshe tables

34
Q

Method of
ascertaining the extent to which
an organization will benefit from
the use of a particular selection
system

A

Utility formula

35
Q

The length of time an
employee has been with an
organization

A

Tenure

36
Q

Group
differences in test scores that are
unrelated to the construct being
measured

A

Measurement bias

37
Q

The most common example of adverse impact is changes in work conditions such as new workers being more stressed than ever before. Other examples include increased stress in the workplace due to changes in work schedule, working too hard or too much, or being in a negative work environment.

A

Adverse Impact

38
Q

A situation in
which the predicted level of job
success falsely favors one group
over another

A

Predictive bias

39
Q

The
characteristic of a test that significantly predicts a criterion for
one class of people but not for
another

A

Single-group validity

40
Q

The
characteristic of a test that significantly predicts a criterion for
two groups, such as both minorities and nonminorities, but
predicts significantly better for
one of the two groups

A

Differential validity

41
Q

A
statistical procedure in which the
scores from more than one
criterion-valid test are weighted
according to how well each test
score predicts the criterion

A

Multiple regression

42
Q

Selecting applicants in straight
rank order of their test scores

A

Top-down selection

43
Q

A method of making selection
decisions in which a high score
on one test can compensate for
a low score on another test. For
example, a high GPA might
compensate for a low GRE score

A

Compensatory approach

44
Q

A variation on
top-down selection in which the
names of the top three applicants are given to a hiring authority who can then select any
of the three

A

Rule of three

45
Q

The minimum
test score that an applicant must
achieve to be considered for hire

A

Passing score

46
Q

A selection strategy in which
applicants must meet or exceed
the passing score on more than
one selection test

A

Multiple-cutoff approach

47
Q

Selection practice of administering one test at a time so that
applicants must pass that test
before being allowed to take the
next test

A

Multiple-hurdle approach

48
Q

A statistical technique based on the standard
error of measurement that allows similar test scores to be
grouped

A

Banding

49
Q

The
number of points that a test
score could be off due to test
unreliability

A

Standard error of
measurement (SEM)