General Flashcards

1
Q

What does the superscript within parentheses mean?

e.g. X(i)

A

The ith row of X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does the symbol ∀ mean?

A

“for all” or “for any”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

In a joint probability distribution table, how many rows are there?

A

One for each possible combination of variable values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Given a joint probability distribution, what can we calculate?

A

Conditional or joint probabilities over any subset of the variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What’s the conditional probability equation?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q
  • What is bias?
  • What is variance?
A
  • The inability for a ML method to capture the true relationship
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does high bias correspond to? (2)

A

High bias—-> underfitting —–> More train set error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does high variance correspond to?

A

High variance —–> overfitting —–> More dev set error and more test set error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the development set (aka dev set)?

A

It’s another term for “validation set”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What’s a useful way to think of bias?

A

how well does my model fit the training data?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What’s a useful way to think of variance?

A

how well does my model generalize to unseen data sets?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Can you have high bias and high variance?

A

Yes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Given training error and validation error, how can you assess the bias and variance?

A
  • Training error can tell you the bias.
  • How much higher your validation error is than the training error can tell you the variance.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Why can’t we use the validation set for testing performance?

A
  • Because we used the validation set to tune our model parameters. If we tested on the validation set, we wouldn’t know whether those performance gains from tuning were beneficial for unseen data, or just improved our performance on the validation set
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q
  • What is the validation set?
  • What is the test set?
A
  • Set you use to pick the best parameters or model to use
  • Only used to get a metric of how well your model is performing on unseen data
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What models can we use higher-order features for?

A

Non-exhaustive list:

  • Logistic regression
  • Linear regression
17
Q

What defines a decision boundary?

A

The hypothesis h and the parameters

18
Q

What do we know about the update rule for logistic regression?

A

Looks very similar to the update rule for linear regression, except the hypothesis is different

19
Q

Does feature scaling work for logistic regression?

A

Yes

20
Q

How does multiclass logistic regression work?

A
  • It’s called one vs all classification
  • Train multiple classifiers (one for each class)
  • For each class, take the binary logistic regression prediction for for each point
    • Then choose the class with the highest probability output