Topic 6: Machine Learning: Performance Evaluation, Backtesting & False Discoveries Flashcards

1
Q

Describe a ranking classifier

A

a classifier that gives scores to instances (classifier + threshold = single confusion matrix)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Describe the ROC graph

A

Two-dimensional plot with false positives rate on the x-axis and true positive rates on the y-axis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Describe the ROC graph.

A

Y-axis shows the true positive rate (sensitivity) and the X-axis shows the false positive rate (1-specificity)

sensitivity = TP / (TP+FN)

specificity = FP / (FP+TN)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Describe the four corners and the diagonal of the ROC graph.

A

Bottom left: Conservative (only make classifications with strong evidance)

Upper right: permissive (make positive classifications with weak evidence)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Define the hit rate and false alarm rate.

A

hit rate = percentage of positives correctly classified (TP/(TP+FP))

false alarm rate = FP/(FP+TN)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Define the AUC measure.

A

Area under the curve is used to assess the performance of the detection of a model independent of the detection threshold.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Describe the cumulative response curve, also known as the lift curve.

A

Lift curve plots the hit rate as a function of the population that is targeted.

(e.g. 20% test instances targets 60% of positives targeted)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Describe why standard statistical tools, such as p-values and t-statistics,
can lead to false discoveries in the presence of multiple tests.

A

the large number of tests will lead to false positives/false negatives so you need a tougher standard.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Calculate the t-statistic based on the reported Sharpe ratio for testing a
single trading strategy.

A

T-statistic = Sharpe Ratio × √Number of years

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Describe and apply Bonferroni tests in the context of the family-wise error rate
(FWER) approach to adjusting p-values for multiple tests.

A

Approaches to the multiple testing problem in statistics:

Bonferroni test, an FWER, accepts no false discoveries. Calculated by 0.05/number of tests.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Recognize and apply the Holm function to calculate adjusted p-values

A

Holm pk = 0.05 / (total number of tests + 1 - k), compare p-value with their hurdles

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Describe the Holm method in the context of the false discovery rate (FDR) approach to adjusting p-values for multiple tests.

A

The holm method is less stringent than the bonferoni method, the false discovery rate (FDR) is less stringent than both of them.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Describe the process of accepting and rejecting tests using the Holm method.

A

P-value should be less than the Holm statistic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Explain the relationship between avoiding false discoveries and missing
profitable opportunities.

A

Adjusting the hurdle when performing multiple tests decreases type I errors (false discoveries) but increases type II errors (missing discoveries).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Define specificity and sensitivity.

A

specificity = TN / (TN + FP)

sensitivity = TP / (TP + FN)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Describe the false discovery rate with the help of a tree diagram

A

top part of tree diagram = sensitivity, bottom part is specificity.

17
Q

Calculate the false discovery rate.

A

FP / (FP + TP)