Quiz 3 Flashcards

1
Q

three main types of outcomes of interest

A

predicted numerical value
predicted class membership
propensity (probability when categorical)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

predicting class membership

A

classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

new records most likely to be part of class

A

ranking

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

predictive accuracy measures

A

mean absolute error/deviation, average error, mean absolute percentage error, root mean squared error, total sum of squared errors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

compares the model’s predictive performance to a baseline model that has no predictors

A

lift chart

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is a lift chart looking for?

A

subset of records that has the highest cumulative predicted values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

observation that belongs to one class but model put it in another

A

misclassification error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

summarizes the correct and incorrect classifications that a classified produced for a certain dataset (validation)

A

classification/confusion matrix

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

overall error rate

A

incorrect/total

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

error

A

actual - prediction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

ability to correctly detect all important class members

A

sensitivity of classifier

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

ability to rule out negative class members correctly

A

specificity of classifier

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

plots 1 - sensitivity and specificity

A

ROC curve

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

ability to detect only the important class members

A

precision of classifier

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

average cost of miscalculation per classified observation

A

average misclassification cost

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

score model to validation set that is random or score model to an oversampled validation set and reweight the results to remove the effects of oversampling

A

how to adjust for oversampling

17
Q

how many responders from whole data is a sample responder worth

A

oversampling weights

18
Q

accurately classify the most interesting/important cases

A

goal of ranking

19
Q

actual is no, predicted is no (0,0)

A

true negative

20
Q

actual is yes, predicted is yes (1,1)

A

true positive

21
Q

actual is yes, predicted is no (1,0)

A

false negative

22
Q

actual is no, predicted is yes (0,1)

A

false positive

23
Q

accuracy

A

1 - error

24
Q

error with confusion matrix

A

(false negative + false positive) / total

25
Q

if it is important to predict positive values correctly what should you do?

A

lower the cut-off

26
Q

sensitivity

A

true positive / (true positive + false negative)

27
Q

F1

A

(2 * precision * recall) / (precision + recall)

28
Q

what is the ideal part of a ROC curve?

A

top lift, high sensitivity and high specificity