Exam prep Flashcards
Supervised learning
A subcategory of machine learning which uses inputs, a desired output and labels to train a model.
Semi-supervised learning
Using labelled as well as unlabeled data to perform certain learning tasks.
Dimensionality reduction
Model h in H that represents each instance x with a lower dimension feature vector whilst preserving key features.
Unsupervised learning
A subcategory of machine learning which uses inputs and a desired output to train a model.
Clustering Analysis
A machine learning technique that involves grouping sets of objects in such a way that objects in the same group have similar features.
Anomaly Detection
The identification of rare items, events or observations which deviate significantly from the majority of the data.
Reinforcement learning
A subset of machine learning that allows an agent to learn through trial and error using feedback from its actions.
P(A | B)
P(A n B)/P(B)
Max
Maximum value of a function
Argmax
Sequence of x values to get maximum y values.
For example, for max sin(), values would include 0.5, 1.5 etc, which maximise the y value, but would not include things like 1 or 2, which do not maximise sin().
MAP(A)
argmax_a P(A)
Highest probability in A.
MAP(A, B)
argmax_a, b P(A,B) = argmax_a, b P(B|A)P(A)
Highest probability in P(A, B)
Validation dataset
Part of training dataset.
Regularisation
Techniques that are used to calibrate machine learning models in order to minimize the adjusted loss function and prevent overfitting or underfitting.
Test dataset
Training dataset
Error rate for confusion matrix
1 - accuracy
or
FP + FN / TP + FP + FN + TN
Accuracy for confusion matrix
TP + TN / TP + FP + FN + TN
Loss function
1/m x sum(x - x_)^2
where
x = observed values
x_ = predicted values
Linear regression
Used to minimise the loss function.
Naive Bayes
P(X … Xn | Y) = ΠP(Xi | Y)
Essentially, Xi and Xj are conditionally independent given Y.
Rosenblatt’s Perceptron
- initalise weights randomly
- take one sample x and predict y
- for erroneous predictions update weights
- if output = 0 and y’s true value 1, increase weights
- if output = 1 and y’s true value 0, decrease weights
- repeat until no errors
Perceptron
Machines for binary classifications, with one weight per input.
We multiply the weights with their respective inputs and add bias.
If the result is larger than threshold return 1, else 0.
XOR requires multiple perceptrons.