L2 Flashcards

1
Q

Formulation of the problem is a :

A

Model : an abstract mathematical description, and not the real thing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Abstraction

A

The process of removing irrelevant detail from a representation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The most common form of ML:

A
  1. Input: (Features) , x
  2. Output: (Label), y
  3. Goal: Learn Mapping, f: x to y.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Supervised (inductive) learning

A

Given: Training data + desired outputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Unsupervised Learning

A

Given: Training data (without desired output).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Why bother transforming to vector?

A

Bias: Ordinality Assumption

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Label Encoding Pros

A

Simplicity and Efficiency.
Memory Efficient
Useful for ordered Categories.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Label Encoding Cons

A

Implies Ordinality
Not suitable for non-ordinal features.
Model Dependency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Linear Model Pros:

A

Many real-world processes are approximated with linear mode.

Can be solve analytically

The simplest yet the fundamental model of the other machine learning models.

Non-Linear problem can be transformed to linear one.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Linear Model Con:

A

May have low effectiveness and capacity.

Model too simple,
Noise in the features
Simple features doesn’t account for all variation,
Error in labels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Data Representation Normalization - Common Techniques:

A

Min-Max Scaling: Transforms features by scaling each feature to a given range, usually [0,1] or [-1,1]

Max Abs Scaling: Scales each features by its max absolute value, resulting in values in the range [-1, 1].

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Data Representation Normalization - Benefits:

A

Faster Convergence
Numerical Stability
Equal Importance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Precision Function

A

True Positive / (True Positive + False Positive)

  • Positive within all the positive predictions (model-wise)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Recall Function

A

True Positive / (True Positive + False negative)

  • Positive within all positive instances (data-wise)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Accurancy Function

A

(True Positive + True Negative) / Total

  • correct prediction over all instance.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Low precision means

A

Many of its positive predictions are incorrect.
Too tolerative
Aka: Type 1

17
Q

Low recall means

A

Many of its positive predictions are incorrect.
Too picky.
Aka: Type 2

18
Q

F1

A

Harmonic mean of precision and recall.

2 * (precision * recall) / (Precision + Recall)

19
Q

Data Splitting

A

The process of dividing a dataset into separate sets of train and test of machine learning model.