Supervised Linear Models Flashcards
Notes from Lecture 5, which may help with the exam
What are the main comparison points between Supervised and Unsupervised Learning?
- Both require a training dataset
- Supervised Learning requires labels for each instance/class e.g. two categories, y = 0 and y = 1
- Unsupervised Learning does not use labels
What defines a Linear Algorithm?
Linear models assume that the sample features (X) and the label outputs (Y) are linearly related, described by f(x) = (W, x) + b.
It also often refers to a linear decision boundary in classification models.
What are some examples of Linear Algorithms?
Linear Regression
Logistic Regression (specifically for Classification problems)
Naive Bayes
Support Vector Machines (SVM)
What defines a Non-Linear Algorithm?
Non-linear algorithms assume a non-linear relationship between x and y. Thus, f(x) can be a function of arbitrary complexity
What are some examples of Non-Linear Algorithms?
K-Nearest Neighbour
Kernel SVM
Decision Trees
Neural Networks
What does the term ‘Linearly Separable’ refer to/mean?
- Datasets whose classes can be separated by linear decision surfaces
- Implies no class overlap
- Classes can be divided by lines for 2D data or planes in 3D data
What defines a Parametric Algorithm?
Parametric algorithms are model-driven algorithms that assume the data follows a specific distribution in its feature space or a pre-defined relationship between feature and outcome
What are some examples of Parametric Algorithms?
Linear Regression
Gaussian Naive Bayes
Maximum Likelihood Classifier
What defines a Non-Parametric Algorithm?
Non-Parametric algorithms are data-driven algorithms where approaches are not constrained to prior assumptions on the data distribution.
What are some examples of Non-Parametric Algorithms?
Decision Trees
Neural Networks
What is the primary difference between Classification and Regression?
- Regression estimates values for a given dataset
- Regression predicts numerical values
- Classification assigns class labels to data
- Classification predicts categorical values
What is the definition of Overfitting?
A model learns to map the training data too well, which negatively impacts the performance of the model on new, unknown data. It results in the model having poor generalisability
What is the definition of Underfitting?
If it neither can model the training nor the test data correctly. Underfitting is easier to detect than Overfitting in the training phase using evaluation metrics. It also results in poor performance metrics of the model.
What is the definition of Variance?
Variance is the amount that the estimate of the target function will change, given different training data.
What is the definition of Bias?
Bias indicates the error between the approximated model to the ideal model