Course Practice Problems Flashcards

1
Q

Describe the Bayes Classifier

A

The Bayes classifier is a classification procedure that reaches the true minimum misclassification rate. It can be thought of as the underlying model that generates the true categories of the observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Will the Bayes Classifier result in 0 misclassifications?

A

It can, but its highly unlikely

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Under what assumptions is LDA the Bayes Classifier?

A

If each group (or subpopulation) is assumed to be mul-tivariate normally distributed, and all groups have a common covariance matrix

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Under what assumptions is QDA the Bayes Classifier?

A

If each group is assumed to be multivariate normally distributed with uniquecovariance matrices

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Describe the difference between clustering & classification.

A

Clustering is attempting to separate observations into groups according to the predictors (X) — there is noknown response (Y) that we are actively modelling, it is an exploratory procedure.

Classification is the process of fitting a model using predictors (X) to predict a categorical response variable(Y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is a tuning parameter & suggest a common approach to optimize it

A

a set values to help tune your methods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

train test RF

A

Random forests use the nonparametric bootstrap and build trees for each bootstrapped sample. As such,each tree will be fit on approximately 2/3 of the data. So, conveniently, misclassification rates (classification)or MSE (regression) can be calculated using out-of-bag observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

PCA vs NMF

A

Principal component analysis (PCA) is an orthogonal rotation of the original data. There is no dimen-sionality reduction or loss of information (unless we remove components) or assumptions placed on thedistribution of the data (aside from it being numeric). In practice, we often need to scale the data.Non-negative matrix factorization (NMF) assumes that the observed dataXis non-negative, and can bedecomposed into two non-negative matricesWandH. The dimensionality ofWandHhas to be pre-specified and so there will be a loss of information. In contrast to PCA: solutions are non-unique, there isno inherent ordering to the basesH(which are the NMF equivalents to components from PCA), and thenon-negativity assumption results in reconstructions ofXthat canonlybe additive.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

P value definition

A

fill

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

4 steps of K means

A

1.
2.
3.
4.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Whats a hidden layer in a Neural Networks

A

artificial neutron inbetween input n output, takes in a weighted output and provides and output from an activated function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Describe support, confidence, and lift from association rule mining.

A

fill

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Describe support, confidence, and lift from association rule mining.

A

Support : the proportion of time that the item is in the dataset

Confidence: probability that if LHS then RHS

Lift:

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Describe what multidimensional scaling achieves. What is its relationship to principal component analysis?

A

PCA: clusters groups based on correlation
MDS: clusters groups based on distance

they have the same analysis, so you could reverse

How well did you know this?
1
Not at all
2
3
4
5
Perfectly