Week 2: Logistic Regression (only) Flashcards

1
Q

What are two problems with the classes in log reg that can make other criteria than the (overall) misclassification rate the biggest priority?

A

1) An asymmeitric problem - where it is more important to correctly predict some classes than other,
2) An imbalanced problem - where classes occur with very different frequencies.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Give an example of when it can be more important not to falsely predict the negative class than to falsely predict the positive class?

A

In a situation where we predict health status of patiens. Falselsy predicting the negative class could lead to predicting a sick patient as healthy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Give an example of when we can encounter an imbalanced problem?

A

If we are modeling a very rare disease.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the equation that needs to be solved to compute the decision boundary in log reg? Why?

A

g(x) = 1-g(x).

This since the solution are points in input space for which the two classes are predicted to be equally probable and therefore lie on the decision boundary.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is a linear classifier?

A

A model whose decision boundary is linear.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the softmax function?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How do we write the loss function L(y.hat, y.i) more specific (for log reg)?

A

L[p(y.i = 1 | x.i ; theta), y.i]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Write the full cost function where the log lik. function has been simplified, and identify the loss and cost . What do we call this loss function?

A

xx

The loss function is called the binary cross-entropy loss.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Write only the (general) binary cross-entropy loss and simplify. Speficy this general loss for logistic regression.

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the logistic loss?

A

2 ln [1+exp(y.i theta^T x.i)]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the log of (1/n)?

A
  • log (n)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is a classification rule?

A

A rule that maps predicted probabilities into class prediction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is computationally good with the likelihood function for a Bernoulli distribution used in logistic regression?

A

That the pdf will reduce to only one part when the probability of y = 1, and only the other part when the probability of y = (-1).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly