Utility Flashcards

1
Q

According to article, what’s wrong with traditional indexes like specificity, sensitivity, PPV, and NPV?

A

They vary based on cutoff scores, base rates, and cost/benefits with a particular context or assessment.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What do the article authors encourage to do?

A

Use methods of signal detection theory, like ROC for utility

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What happens when cutoff scores change? Which of the following are affected? (Selection Ratio, Base Rate, Sensitivity, Specificity, Positive Predictive Power, Negative Predictive Power)

A

Only base rate does not change.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What metrics are directly affected by the base rate?

A

PPV and NPV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the ROC-based index is independent of?

A

1) base rate / prevalence
2) cutoff score
3) values of the 4 decision-making outcomes (i.e., TN TP FN FP)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How to use ROC analysis?

A

Look at area under ROC curve. This shows discriminative power of an assessment independent of cutoff value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does it mean when AUC = 1?

What about when AUC = 0.5 or less than 0.5?

A

When AUC = 1, the model is perfect across all cutoffs. Sensitivity and Specificity both are 100%

When AUC = 0.5, it’s no better than guessing.

When AUC < 0.5, the test inverts values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Test Utility

A

Usefulness or practical value of testing to improve efficiency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

3 factors for utility of a test (what makes it have high/low utility?)

A

1) psychometric soundness (validity + reliability)
2) costs
3) benefits

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Does utility require measuring something?

A

No, there can be utility without measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Is utility an empirical or conceptual endeavour?

A

Empirical

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How are costs complicated?

A

It’s complicated to calculate costs due to nonmonetary costs. (eg. what are the costs of not administering a test, resulting in possibly failing to diagnose a disease?)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What general procedure determines costs and benefits of a test?

A

Utility analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Expectancy data

A

One way of doing utility analyses. Uses predictions of outcomes to judge the benefits of a test?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What does a Taylor-Russell table accomplish?

A

A type of expectancy data.

Helps to see how much using a test will improve selection over existing method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Naylor-Shine tables

A

Difference in means between selected and unselected groups to see what a test adds (increase in avg score)

17
Q

What is a Taylor-Russell table based on?

A

Selection Ratio and Validity estimate

18
Q

What formula is used when assessing costs and benefits of a test?

A

Brogden-Cronbach-Gleser formula.

Estimates utility gain based on dollar amount.

19
Q

Higher selection ratio means probably higher or lower false positives?

A

More false positives, less false negatives.

20
Q

Why is the pool of job applicants a practical consideration when determining utility of a test?

A

Utility analyses assume that there exists an unlimited supply of the type of people who is qualified according to the test.

21
Q

Why is the fact that some people won’t accept job offers a practical consideration for utility of a test?

A

It will overestimate utility if not considered. Some people will deny job offers. 80% of utility is a better estimate.

22
Q

Alternative name for relative cut score

A

Norm-referenced cut score

23
Q

Fixed vs. Relative cut scores

A

a specific number vs. a distribution or percentile.

24
Q

Multiple hurdle or multistage selection process

What’s multiple cut scores and how does that differ?

A

Ensures a tester must have a few predictors. Multiple cut scores

Multiple cut scores can refer to different categories of same predictor (eg. A, B, C, D grades have multiple cut scores)

25
Q

Compensatory model of selection

what does this compare with?

A

A high score on one predictor can balance out a low score of another

Compared to a multiple hurdle/multistage process.

26
Q

What statistical tool for compensatory model of selection?

A

Multiple regression

27
Q

Angoff Method

A

A method to determine cut scores.

Average experts judgements on expectations.

28
Q

What can show that Angoff method might not be best?

What are its weaknessess

A

Low interrater reliability (if the experts disagree).

Lack of data-driven techniques, more subjective.

29
Q

Known Groups Method

(alt name)

A

Method of contrasting Groups. A way to determine test scores.

Take scores of those known to possess and not possess some predictor.

30
Q

What’s a problem with known groups method?

A

The choosing of groups are arbitrary

31
Q

IRT-based methods

A

A way to determine cutoff scores.

Each item has some deemed level of difficulty. People must answer items that are above a certain difficulty.

32
Q

Item-mapping method

A

a way to make cut score with IRT-based method.

Arranges items on histogram, judges judge whether the questions will be answered right at least half the time = cut score.

33
Q

Bookmark Method

A

a way to make cut score with IRT-based method.

Experts place a ‘bookmark’ between two pages in a given booklet that separate minimum difficulty (cut off)

34
Q

Drawbacks of IRT-based methods

A

Need expert opinion, floor/ceiling effects, optimal numbers of items in booklet (for bookmark method)

35
Q

Method of predictive yield

A

Thorndike’s method of setting cut scores.

Needs posititions to be filled, projections of likelihood to offer acceptance, and distribution of applicant scores.

36
Q

Discriminant Analysis

A

See relationship between identified variables and two naturally occurring groups. (Discriminant Function Analysis)