Lecture 7 Flashcards

1
Q

Random Variable

A

Refers to an element/event whose status is unknown

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Domain

A

The set of values a random variable can take

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Conditional probability

A

the chances that some outcome occurs given that another event has also occurred

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Joint Probability

A

The probability that a set of random variables will take a specific value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are three types of classifiers?

A

Instance based classifiers, generative, and discriminative

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Instance based classifiers

A

Use observations directly without models

e.g. k nearest neighbors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Generative classifiers

A

build a generative statistical model

e.g. Bayes classifiers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Discriminative classifiers

A

directly estimate a decision rule/boundary

e.g. decision trees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Gaussian Naive Bayes classifier

A

assumes that features follow a normal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Multinomial Naive Bayes

A

each feature represents an integer count of something, like how often a word appears in a sentence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Bernoulli Naive Bayes

A

Assumes your feature vectors are binary or continuous values which can be precisely split (binarized) with a predefined threshold

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Advantages of Naive Bayes classifiers

A

They are simple, work well with a small amount of training data, and the class with the highest probability is considered as the most likely class

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Disadvantage of Naive Bayes classifiers

A

Estimates parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the complexity of a decision tree model determined by?

A

the depth of the tree

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What causes overfitting in decision trees?

A

Increasing the depth of the tree and thus increasing the number of decision boundries

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the aim of Bayesian Linear Regression?

A

Not to find the single “best’ value of the model parameters, but rather to determine the posterior distribution for the model parameters

17
Q

Advantages of Decision Trees

A

They are suitable for multiclass classification, the model is easily interpretable, it can handle numerical and categorical data, is non-linear, and can tolerate missing values

18
Q

Disadvantages of Decision Trees

A

They’re prone to overfitting without pruning, are weak learners, and singular trees do not make great predictions

19
Q

Which classification algorithms that we know are linear?

A

Logistic regression and

Linear SVMs

20
Q

Which classification algorithms that we know are nonlinear?

A
KNN
Neural Networks (Multi-Layer Perceptron)
Kernel SVMs
Naive Bayes
and Decision Trees
21
Q

What does the vertical line | in probability theory refer to?

A

given

e.g. p(A|B) is the probability of A given B.