Lecture 4: Decision trees & k-means Flashcards

1
Q

What is logical inference?

A

Process of deriving new facts from a set of premises

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the 3 types of logical inference?

A

1.Deduction
2.Abduction
3.Induction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is deduction?

A

-Conclusion follows necessary from the premises

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is abduction?

A

-Conclusion is one hypothetical(most probable) explanation for the premises

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is induction?

A

-Conclusion about all members of a class from the examination of only a few member of the class.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is inductive learning?

A

-Most work in ML
-Examples are given to train a system in a classification task

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are 3 examples of techniques in ML?

A

-Probabilistic methods
-Decision trees
-Neural networks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are decision trees(3)?

A

-Simple, but very successful form of learning algo
-Very well-know algo is ID3 and its successor C4.5

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the Ockham’s razor principle?

A

-Always favor the simplest answer that correctly fits the training data (the smallest tree on average)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What type of assumption is Ockham’s razor principle based on?

A

Inductive bias : making a choice beyond what the training instances contain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is maximum information-gain and what uses it?

A

-Choose the attribute that has the largest information gain.
-ID3 uses it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is essential information theory?

A

-Shannon in 1940s
-Notion of entropy(information content)
-Measure how predictable a random variable(RV) is if:
i. if you already have a good idea about the answer–> low entropy
ii. if you have no idea about the answer–> high entropy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are the 2 things to watch out for in all types of learning?

A

-Noisy input
-Overfitting/underfitting the training data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are 3 cases of noisy input?

A

-Two examples have the same feature-value pairs, but different outputs
-Some values of features are incorrect or missing
-Some relevant attributes are not taken into account in the data set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are 3 cases of overfitting?

A

-Complicated boundaries overfit the data
-Large number of irrelevant features
-They do not generalize well to the new data
*Training error is high
*Testing error is high

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is underfitting(2)?

A

-Model is not expressive enough–> not enough features
-There is no way to fit a linear decision boundary so that the training examples are well separated
*Training error is high
*Testing error is high

17
Q

What is clustering?

A

-The organization of unlabeled data into similarity groups called clusters

18
Q

What are clusters?

A

-Collection of data items which are similar between them, and dissimilar to data items in other clusters.

19
Q

What is the pro and con of k-means?

A

Pro: efficient in time complexity
Con: algo sensitive to outliers