Evaluating Selection Techniques and Decisions Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

The extent to which a score from a test or from an evaluation is consistent and free from error

A

Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The extent to which repeated administration of the same test will achieve similar results

A

Test-retest reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The consistency of test scores across time

A

Temporal stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The extent to which two forms of the same test are similar

A

Alternate-forms

reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

A method of controlling for order effects by giving half of a sample Test A first, followed by Test B, and giving the other half of the sample Test B first, followed by Test A

A

Counterbalancing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

The extent to which the scores on two forms of a test are similar

A

Form stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

The extent to which responses to the same test items are consistent

A

Item stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

The extent to which test items measure the same construct

A

Item homogeneity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

A statistic used to determine internal reliability of tests that use items with dichotomous answers (yes/no, true/false)

A
Kuder-Richardson
Formula 20 (K-R 20)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

A form of internal reliability in which the consistency of item responses is determined by comparing scores on half of the items with scores on the other half of the items

A

Split-half method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Used to correct reliability coefficients resulting from the split-half method

A

Spearman-Brown

prophecy formula

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

A statistic used to determine internal reliability of tests that use interval or ratio scales

A

Coefficient alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The extent to which two people scoring a test agree on the test score, or the extent to which a test is scored correctly

A

Scorer reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

The degree to which inferences from test scores are justified by the evidence

A

Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The extent to which tests or test items sample the content that they are supposed to measure

A

Content validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

The extent to which a test score is related to some measure of job performance

A

Criterion validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

A measure of job performance, such as attendance, productivity, or a supervisor rating

A

Criterion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

A form of criterion validity that correlates test scores with measures of job performance for employees currently working for an organization

A

Concurrent validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

A form of criterion validity in which test scores of applicants are compared at a later date with a measure of job performance

A

Predictive validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

A narrow range of performance scores that makes it difficult to obtain a significant validity coefficient

A

Restricted range

21
Q

The extent to which inferences from test scores from one organization can be applied to another organization

A

Validity generalization

VG

22
Q

A form of validity generalization in which validity is inferred on the basis of a match between job components and tests previously found valid for those job components

A

Synthetic validity

23
Q

The extent to which a test actually measures the construct that it purports to measure

A

Construct validity

24
Q

A form of validity in which test scores from two contrasting groups “known” to differ on a construct are compared

A

Known-group validity

25
Q

The extent to which a test appears to be valid

A

Face validity

26
Q

Statements, such as those used in astrological forecasts, that are so general that they can be true of almost anyone

A

Barnum statements

27
Q

A book containing information about the reliability and validity of various psychological tests

A

Mental Measurements

Yearbook (MMY)

28
Q

A type of test taken on a computer in which the computer adapts the difficulty level of questions asked to the test taker’s success in answering previous questions

A

Computer-adaptive

testing (CAT)

29
Q

A series of tables based on the selection ratio, base rate, and test validity that yield information about the percentage of future employees who will be successful if a particular test is used

A

Taylor-Russell tables

30
Q

The percentage of applicants an organization hires

A

Selection ratio

31
Q

Percentage of current employees who are considered successful

A

Base rate

32
Q

A utility method that compares the percentage of times a selection decision was accurate with the percentage of successful employees

A

Proportion of correct

decisions

33
Q

Tables that use the base rate, test validity, and applicant percentile on a test to determine the probability of future success for that applicant

A

Lawshe tables

34
Q

Method of ascertaining the extent to which an organization will benefit from the use of a particular selection system

A

Utility formula

35
Q

The length of time an employee has been with an organization

A

Tenure

36
Q

Group differences in test scores that are unrelated to the construct being measured

A

Measurement bias

37
Q

An employment practice that results in members of a protected class being negatively affected at a higher rate than members of the majority class. Adverse impact is usually determined by the four-fifths rule

A

Adverse impact

38
Q

A situation in which the predicted level of job success falsely favors one group over another

A

Predictive bias

39
Q

The characteristic of a test that significantly predicts a criterion for one class of people but not for another

A

Single-group validity

40
Q

The characteristic of a test that significantly predicts a criterion for two groups, such as both minorities and non-minorities, but predicts significantly better for one of the two groups

A

Differential validity

41
Q

A statistical procedure in which the scores from more than one criterion-valid test are weighted according to how well each test score predicts the criterion

A

Multiple regression

42
Q

Selecting applicants in straight rank order of their test scores

A

Top-down selection

43
Q

A method of making selection decisions in which a high score on one test can compensate for a low score on another test. For example, a high GPA might compensate for a low GRE score

A

Compensatory approach

44
Q

A variation on top-down selection in which the names of the top three applicants are given to a hiring authority who can then select any of the three

A

Rule of three

45
Q

The minimum test score that an applicant must achieve to be considered for hire

A

Passing score

46
Q

A selection strategy in which applicants must meet or exceed the passing score on more than one selection test

A

Multiple-cutoff approach

47
Q

Selection practice of administering one test at a time so that applicants must pass that test before being allowed to take the next test

A

Multiple-hurdle approach

48
Q

A statistical technique based on the standard error of measurement that allows similar test scores to be grouped

A

Banding

49
Q

The number of points that a test score could be off due to test unreliability

A

Standard error of

measurement (SEM)