Machine Learning Flashcards

1
Q

Inductive learning

A

generalize from a given set of (training) examples so that accurate predictions can be made about future examples; learn an unknown function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

how to represent a “thing” in machine learning

A

x: example or instance of a specific object; represented by a feature vector; each dimension - feature

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

feature vector representation

A

extract a feature vector x, that describes all attribute relevant for an object; each x is a list of (attribute, value) pairs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

types of features

A

numerical features - discrete or continuous
categorical features - no intrinsic ordering
ordinal features - similar to categorical but clear ordering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

point in feature vector representation

A

each example can be interpreted as a point in a D-dimensional feature space, where D is the number of features/attributes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Training set

A

A training set is a collection of examples (instances), which is the input to the learning process; assume instances are independent and identically distributed. training set = experience given to learning algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

idd

A

independent and identically distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Unsupervised learning

A

training set = x1, …xn; no “teacher” to show how examples should be handled; tasks: clustering, discovery, novelty detection; dimensionality reduction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

goal of clustering

A

group training samples into clusters such that examples in the same cluster are similar, and examples in different clusters are different

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Clustering methods

A

Hierarchical Agglomerative Clustering
K-means Clustering
Mean Shift Clustering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Hierachical Clustering General Idea

A

initially every point is in its own cluster
find the pair of clusters that are the closest
merge the two into a single cluster
repeat
end result: binary tree

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How to measure closeness between 2 clusters (hierarchical clustering)

A

Single linkage, complete-linkage, average linkage

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

single-linkage

A

the shortest distance from any member of 1 cluster to any member of another cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

complete linkage

A

the largest distance from any member of 1 cluster to any member of another cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

average linkage

A

the average distance between all pairs of members, one from each cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How to measure the distance between a pair of examples?

A

Euclidean, manhattan/city block, hamming

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Dendrogram

A

binary tree resulting from hierarchical clustering; the tree can be cut at any level to produce different numbers of clusters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What factors affect the outcome of hierarchical clustering?

A

features used, range of values for each feature, linkage method, distance metric used, weight of each feature

19
Q

K-Means Clustering

A

Specify the desired number of clusters and use an iterative algorithm to find them

20
Q

K-means clustering general idea

A

if have cluster centers, for each point, choose closest center, if have points, choose a cluster center to be the mean/centroid of the points in the cluster; repeat until convergence

21
Q

K-means algorithm

A

input: x…xn
select k cluster centers c1, …ck
fore each point x, determine it’s cluster by finding closest cluster center; update cluster centers as centroid (mean)

22
Q

Distortion

A

Sum of squared distances of each data point to its cluster center; optimal clustering minimizes distortion (over all possible possible cluster locations/assignments)

23
Q

How to pick k in k-means clustering

A

pick number of k to minimize distortion - pick one close to elbow of the distortion curve

24
Q

Does K-means always terminate?

A

Yes (finite number of ways of partioning a finite number of points into k groups)

25
Is K-means guarunteed to find "optimal clustering"?
No, but each iteration will reduce the error/distortion of the clustering
26
Which local optimum k-means goes to is determined by:
solely by the starting cluster centers
27
Ways to pick good starting cluster centers
1. run multiple times with different starting random cluster centers 2. pick starting points based on farthest apart from each other
28
How to pick number of clusters
difficult problem, heuristics depend on number of points/dimensions
29
mean shift clustering
choose search window size and initial location; compute centroid; center search window at mean; repeat
30
mean shift clustering objective:
find the densest region
31
Supervised learning
learns a function H: x->y in some function family H, such that h(x) predictions the true label y on future data (classification if discrete, regression if continuous)
32
what is a label (supervised learning)?
the desired prediction on an instance x
33
what is a class?
discrete label
34
concept learning
determine from a given set of examples if a given example is or is not an instance of the concept/class/category
35
positive/negative example (concept learning)
if an example is or isn't an instance of concept/class/category
36
Supervised Concept Learning by Induction
Given a training set of positive and negative examples of a concept, construct a description that accurately classifies whether future examples are positive or negative
37
Nearest Neighbor Classification
Save each training example in feature space; classify a new example by giving it the same classification as its nearest neighbor
38
How to pick k for nearest neighbor?
split data into training/tuning sets classify turning set with different k values pick k that produces least turning-set error
39
When doesn't k-NN generalize well?
If the examples in each class are not well clustered
40
How is distribution from training set important?
If train on set from different distribution from future data, will not be as accurate
41
Inductive bias
Any knowledge created by generalization from specific facts cannot be proven true, only false; used when one function is chosen over another
42
Inductive biases commonly used in ML?
restricted hypothesis space bias | preference bias
43
Restricted hypothesis space bias
allow only certain types of h's, not aribitrary ones
44
preference bias
define a metric for comparing h's so as to determine whether one is better than another