Easy Flashcards

1
Q

Distance Functions, Euclidean, _, _

A

Manhattan, Minkowski

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

optimal dataset for K neighbours

A

3-10

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

association rules find all sets of itemsets that have _ greater than the minimum

A

support

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

association rules: find desired rules that have _ greater than the min

A

confidence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

association rules are usually needed to satisfy a user-specified _ & _

A

minimum support, confidence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

formula: support for association rules

A

frq(x,y)/n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

formula: confidence for association rules

A

frq(x,y)/frq(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

K-means clustering… place _ at _ locations; repeat until convergence

A

centroids, random

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

K-means clustering… 1. for each point xi:

A

find nearest centroid

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

K-means clustering… 2. assign the point, & for each determine new centroid

A

to cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

K-means clustering… 3. stop when non of the __ change

A

clustering assignments

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

a perceptron is used to classify _ classes

A

linearly separable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

a percepton consists of _, _, _

A

weights, summation processor, activation function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

a percepton takes a weighted sum of input and outputs, 1 if the sum > than _ _ _ _, _

A

some adjusted threshold value, theta

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

the perceptron can have another input known as

A

the bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

perceptron: it is normal practice to treat the bias as

A

just another input

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

the perceptron bias allows us to

A

shift the transfer curve horizontally along the input axis

18
Q

the perceptron weights determine

A

the slope of the curve

19
Q

draw the perceptron

A
20
Q

perceptron concept: the ouput is set at one of two levels, depending on whether the _, is greater or less than some _ value. This is called:

A

total input, threshold, unit step (threshold)

21
Q

draw the unit step threshold

A
22
Q

Perceptron function: the _ consists of two functions, _ and _ , ranging from 0 and 1, and -1 to +1

A

sigmoid, logistic, tangential

23
Q

perceptron function: Ouput is proportional to the total weighted output

A

piecewise linear

24
Q

perceptron function: bell shaped curves that are continuos. the node output (high / low) is interpreted in terms of class membership (1/0) depending on how close the net input is to a _

A

Gaussian, chosen value of average

25
Q

what helps us control how much we change the weight and bias in a perceptron, which we do in order to get a smallest error

A

the learning rate

26
Q

Perceptron: if we have n variables then we need to find _

A

n + 1 weight values (n variables + the bias)

27
Q

perceptron: if we have 2 inputs the equation becomes:

A

w1x1 + w2x2 + b = 0

where wi is the weight of input i and b is the bias (w0 with input value x0 of 1)

28
Q

what is the data mining task of prediction the value of a categorical variable (target or class)

A

classification

29
Q

transforming attributes from numerical to categorical

A

binning or discretization

30
Q

transforming attributes from categorical to numerical

A

encoding or continuization

31
Q

Represents

A

Linear / Non-linear separability & inseperable

32
Q

list the frequency table methods

A
  • ZeroR
  • One R
  • Naive Bayesian
  • Decision Tree
33
Q

list the covariance matrix methods of classification

A
  • linear discriminate analysis
  • logistic regression
34
Q

list the similarity functions method of classification

A

K Nearest Neighbours

35
Q

List the other methods of classication

A

artificial neural network, support vector machines

36
Q

simplest classification method

A

ZeroR

37
Q

Zero R classifier relies:

A

on the targer and ignores all predictors

38
Q

although there is no predictability power in ZeroR it is useful for _

A

determining a baseline performance as a benchmark for other classification methods

39
Q

how to implement ZeroR

A

contruct a frequency table for the target and select it’s most frequent value

40
Q

Zero R only predicts _

A

the majority class correctly (as shown by the confusion table)