MACHINE LEARNING Flashcards

terms

1
Q

Accuracy

A

How often model predicts correctly, e.g: 100 predictions, 80 is correct
80/100 = 0,8 0.8 or 80%

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Activation function

A

A function applied in neural netwroks to introduce non-linear transformations, enabling the network to model complex relationships. Exampes: Sigmoid, Relu

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Autoencoder

A

A type of neural network used for unsupervised learning, mainly for dimensionality reduction and feature learning by encoding inputs to a lower-dimensional space and reconstructing them.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Back-propagation

A

An algorithm for traing neural networks, where errors are propagated backward through the network to adjust weights based on gradients.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Bagging

A

“Bootstrap aggregating” - a technique to reduce variance by training multiple models on random subsets of the data and averaging predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Basis function

A

Functions used in models to transform in put data to make it easier to fit linear models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Bias

A

The error due to overly simplistic models that do not capture patterns well.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Bias-variance tradeoff

A

The balance between a models complexity(variance) and simplicity(bias) to achieve optimal performance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

boosting

A

An ensemble method: building models that correcting errors previos models made which leads to higher accuracy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Bootstrap Sample

A

A sample obtained by randomly sampling with with replacement from dataset, often used in bagging and estimation
example:
- ball from a casket, and identify it with a color
- puts it back, and re-do the process
- point is you might pick same ball twice or more.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Classification

A

A type of supervised learning where the goal is to predict discrete labels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Clustering

A

An unsupervised learning technique for grouping similar data points toghether
- the machine finds patterns by itself
- netflix groups movies after genre

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Complete Linkage

A

A hierarchical clustering method where the distance between clusters is the maximum distance between any pair of points in the clusters
- biggest distance becomes “the distance between the clusters”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Confusion Matrix

A

A table used to evaluate a classification model’s performance by comparing predicted and actual values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Cross-validation

A

A technique for assessing models:
- performance by dividing data into subsets,
- training on some
- testing on others to reduce overfitting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Curse of dimensionality

A
  • problem where data becomes sparse in high-dimensional spaces
  • which makes it harder for models to find patterns
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Deep Learning

A

A subset of machine learning that uses neural networks with many layers to model complex data patterns
- uses many layers
- the more layers, the more complex things can the model learn

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Decision Boundary

A

The boundary that separates different classes in a classifier

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Classifier

A

Algorithm which is trained using labeled data
- employ math and statistics methods to generate predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Decision Tree

A

A model that splits data based on feature values to predict the target label, forming a tree structure.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Ensemble Methods

A

Techniques that combine multiple models to improve predictions, like bagging, boosting, and stacking

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Entropy

A

A measure of uncertainty or disorder in data, used in decision trees to determine the best splits

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Fairness

A

Ensuring that models do not discriminate or create biases against certain groups

24
Q

F1 Score

A

The harmonic mean of precision and recall, providing a balance between the two

25
Feature Extraction
Deriving new features from raw data -often for dimensionality reductin - or improved model performance
26
Feature Selection
choosing the most relevant features for a model to improve accuracy and reduce complexity
27
Generalization
Ability for a model perform well on unseen data - bad generalization: Can 2+2=4, but cant do 3+3=6 - or, only bicycle on one spesific cycle
28
Gradient Descent
An optimization algorithm to minimize a model's loss function by iteratively adjusting parameters - Adjusting a model until it gets perfect
29
Hierarchical Clustering
A clustering method hat creates a tree of clusters by iteratively merging or splitting clusters
30
Imbalanced data
A dataset where some classes are more frequent than others, often requiring techniques like resampling
31
Imputation
Filling in missing data - commonly done using statistical - or model-based techniques
32
Inductive basis
Assumptions to generalize beyond training data
33
Information gain
metric used in decision trees to measure the reduction in entropy after a split
34
Kernel
A function that transforms data into a higher-dimensional space pros: - used in SVM to handle non-linearly separable data
35
K-means
A clustering algorithm that partitions data into k clusters by minimizing the distance of points from the cluster centroid - You want to centralize the clustuers, by doing the same thing multiple times -sensitive for: startposition, outliers, scale on features
36
k-NN
A classification algorithm that assigns labels based on the majority label among the k closest points -
37
Learning rate
Steps a model takes during gradient descent
38
linear regression
A model that finds a linear relationship between a dependent variable and one or more independent variables
39
Lloyds algorithm
A common method for k-means clustering, iterating between assigning points to clusters and updating centroids
40
Logistic regression
A model used for binary classification by modeling the probability of an outcome with a sigmoid function
41
loss function
function that finds error from a models predicts - Used to guide training
42
MAE(mean absolute error)
average of absolute difference between predicted and actual values
43
Missing data(MCAR, MAR, MNAR)
Types of missing data MCAR- completely random MAR- related to observed data MNAR- related to missing data itself
44
Model Evaluation
Process of assessing a model's performance using metrics like accuracy, precision, recall and RMSE
45
Model Selection
The process of choosing the best model among various options based on performance
46
Naive bayes
Probabilistic classifier based on Bayes theorem, assuming features are independent
47
Neural Network
Model inspired by the human brain, consisting of layers of neurons that process data for tasks like classification and regression
48
No free lunch theorem
No model works best for all problems
49
overfitting
When a model learns noise in the training data, performing poorly on new data
50
Oversampling
A technique to balance classes by replicating or generating new samples of the minority class
51
Principal Component Analysis (PCA)
A dimensionality reduction technique that projects data into a lower-dimensional space while retaining variance
52
Pruning
Process of removing parts from a decision tree to prevent overfitting
53
Radial Basis function
A function used in machine learning, often in neural networks and SVMs, to handle non-linear patterns
54
Random Forest
An ensemble of decision trees that improves accuracy and reduces overfitting by aggregating predictions
55
Recall
Ratio true positive predictions to the total actual positives
56
Rectified linear Unit (ReLU)
An activation function that outputs zero for negative inputs and the input itself for positive values
57