Easy Flashcards

1
Q

Distance Functions, Euclidean, _, _

A

Manhattan, Minkowski

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

optimal dataset for K neighbours

A

3-10

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

association rules find all sets of itemsets that have _ greater than the minimum

A

support

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

association rules: find desired rules that have _ greater than the min

A

confidence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

association rules are usually needed to satisfy a user-specified _ & _

A

minimum support, confidence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

formula: support for association rules

A

frq(x,y)/n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

formula: confidence for association rules

A

frq(x,y)/frq(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

K-means clustering… place _ at _ locations; repeat until convergence

A

centroids, random

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

K-means clustering… 1. for each point xi:

A

find nearest centroid

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

K-means clustering… 2. assign the point, & for each determine new centroid

A

to cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

K-means clustering… 3. stop when non of the __ change

A

clustering assignments

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

a perceptron is used to classify _ classes

A

linearly separable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

a percepton consists of _, _, _

A

weights, summation processor, activation function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

a percepton takes a weighted sum of input and outputs, 1 if the sum > than _ _ _ _, _

A

some adjusted threshold value, theta

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

the perceptron can have another input known as

A

the bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

perceptron: it is normal practice to treat the bias as

A

just another input

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

the perceptron bias allows us to

A

shift the transfer curve horizontally along the input axis

18
Q

the perceptron weights determine

A

the slope of the curve

19
Q

draw the perceptron

20
Q

perceptron concept: the ouput is set at one of two levels, depending on whether the _, is greater or less than some _ value. This is called:

A

total input, threshold, unit step (threshold)

21
Q

draw the unit step threshold

22
Q

Perceptron function: the _ consists of two functions, _ and _ , ranging from 0 and 1, and -1 to +1

A

sigmoid, logistic, tangential

23
Q

perceptron function: Ouput is proportional to the total weighted output

A

piecewise linear

24
Q

perceptron function: bell shaped curves that are continuos. the node output (high / low) is interpreted in terms of class membership (1/0) depending on how close the net input is to a _

A

Gaussian, chosen value of average

25
what helps us control how much we change the weight and bias in a perceptron, which we do in order to get a smallest error
the learning rate
26
Perceptron: if we have n variables then we need to find \_
n + 1 weight values (n variables + the bias)
27
perceptron: if we have 2 inputs the equation becomes:
w1x1 + w2x2 + b = 0 where wi is the weight of input i and b is the bias (w0 with input value x0 of 1)
28
what is the data mining task of prediction the value of a categorical variable (target or class)
classification
29
transforming attributes from numerical to categorical
binning or discretization
30
transforming attributes from categorical to numerical
encoding or continuization
31
Represents
Linear / Non-linear separability & inseperable
32
list the frequency table methods
* ZeroR * One R * Naive Bayesian * Decision Tree
33
list the covariance matrix methods of classification
* linear discriminate analysis * logistic regression
34
list the similarity functions method of classification
K Nearest Neighbours
35
List the other methods of classication
artificial neural network, support vector machines
36
simplest classification method
ZeroR
37
Zero R classifier relies:
on the targer and ignores all predictors
38
although there is no predictability power in ZeroR it is useful for \_
determining a baseline performance as a benchmark for other classification methods
39
how to implement ZeroR
contruct a frequency table for the target and select it's most frequent value
40
Zero R only predicts \_
the majority class correctly (as shown by the confusion table)