LEARNING Flashcards

1
Q

provides a computer with data, rather than explicit instructions. Using these data, the computer learns to recognize patterns and becomes able to execute tasks on its own.

A

machine learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

a task where a computer learns a function that maps inputs to outputs based on a dataset of input-output pairs.

A

supervised learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

This is a task where the function maps an input to a discrete output.

A

classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

algorithm that, given an input, chooses the
class of the nearest data point to that input

A

nearest-neighbor classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

algorithm that, given an input, chooses the
most common class out of the k nearest
data points to that input

A

k-nearest-neighbor classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

drawback of k-nearest

A

using a naive approach, the algorithm will have to measure the distance of every single point to the point in question, which is computationally expensive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

how to optimize k-nearest

A

by using data structures that enable finding neighbors more quickly or by pruning irrelevant observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

drawback of perceptron learning

A

data are messy, and it is rare that one can draw a line and neatly divide the classes into two observations without any mistakes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

perceptron learning

A

goal of the algorithm is to find the best weight vector, when the algorithm encounters new data it updates the current weights.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

perceptron learning rule

A

for each data point, we adjust the weights to make our function more accurate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

sequences of numbers

A

vector

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

The weights and values in Perceptron Learning are represented using?

A

vectors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

unable to express uncertainty, since it can only be equal to 0 or to 1.

A

hard treshold

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

uses a logistic function which is able to yield a real number between 0 and 1, expressing confidence in the estimate

A

soft treshold

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

they are designed to find the maximum margin separator

A

Support Vector Machines

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

A boundary that maximizes the distance between any of the data points

A

maximum margin separator

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Benefit of Support Vector Machines

A

they can represent decision boundaries with more than two dimensions, as well as non-linear decision boundaries

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

this function gains value when the prediction isn’t correct and doesn’t gain value when it is correct

A

loss function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

supervised learning task of learning a function mapping an input point to a continuous value

A

regression

21
Q

functions that can be used when predicting a continuous value

A

L1 and L2 loss functions

22
Q

L2 Loss Function Formula

A

(actual - predicted)^2

22
Q

L1 Loss Function Formula

A

|actual - predicted|

23
Q

L1 vs L2

A

L₂ penalizes outliers more harshly than L₁ because it squares the the difference

24
when a model fits the training data so well that it fails to generalize to other data sets.
overfitting
25
process of penalizing hypotheses that are more complex to favor simpler, more general hypotheses.
regularization
26
used to avoid overfitting
regularization
26
a constant that we can use to modulate how strongly to penalize for complexity in our cost function
Lambda (λ)
27
A technique for testing whether a model is overfitted by splitting data in two (testing and training set)
holdout-cross validation
28
downside of holdout cross validation
we don’t get to train the model on half the data, since it is used for evaluation purposes.
29
we divide the data into k sets. We run the training k times, each time leaving out one dataset and using it as a test set. We end up with k different evaluations of our model, which we can average and get an estimate of how our model generalizes without losing any data.
k-Fold Cross Validation
30
Reinforcement Learning as a Markov Decision Process
- Set of states S - Set of actions Actions(S) - Transition model P(s’ | s, a) - Reward function R(s, a, s’)
30
given a set of rewards or punishments, learn what actions to take in the future
reinforcement learning
31
method for learning a function Q(s,a), outputs an estimate of the value of performing action a in state b
Q Learning
32
Q-Learning Process
- The model starts with all estimated values equal to 0 (Q(s,a) = 0 for all s, a). - When an action is taken and a reward is received, the function does two things: 1. It estimates the value of Q(s, a) based on current reward and expected future rewards, and 2. Updates Q(s, a) to take into account both the old estimate and the new estimate.
33
an algorithm that completely discounts the future estimated rewards, instead always choosing the action a in current state s that has the highest Q(s, a).
Greedy Decision-Making
34
Explore vs. Exploit tradeoff
A greedy algorithm always exploits, taking the actions that are already established to bring to good outcomes. However, it will always follow the same path to the solution, never finding a better path. Exploration, on the other hand, means that the algorithm may use a previously unexplored route on its way to the target, allowing it to discover more efficient solutions along the way.
35
In this type of algorithm, we set ε equal to how often we want to move randomly. With probability 1-ε, the algorithm chooses the best move (exploitation). With probability ε, the algorithm chooses a random move (exploration).
ε (epsilon) greedy algorithm
36
allows us to approximate Q(s, a) using various other features, rather than storing one value for each state-action pair
function approximation
37
given input data without any additional feedback/information, learn patterns.
unsupervised learning
38
organizing a set of objects into groups in such a way that similar objects tend to be in the same group
clustering
39
applications of clustering
- Genetic Research - Image Segmentation - Market Research - Medical Imaging - Social Network Analysis
40
41
algorithm for clustering data based on repeatedly assigning points to clusters and updating those clusters’ center
k-means Clustering
42
Categorize the following: A social network’s AI uses existing tagged photos of people to identify when those people appear in new photos.
This is an example of supervised learning
43
Imagine a regression AI that makes the following predictions for the following 5 data points. What is the total L2 loss across all of these data points (i.e., the sum of all the individual L2 losses for each data point)? The true output is 2 and the AI predicted 4. The true output is 4 and the AI predicted 5. The true output is 4 and the AI predicted 3. The true output is 5 and the AI predicted 2. The true output is 6 and the AI predicted 5.
16
44
If Hypothesis 1 has a lower L1 loss and a lower L2 loss than Hypothesis 2 on a set of training data, why might Hypothesis 2 still be a preferable hypothesis?
Hypothesis 1 might be the result of overfitting.
45
In the ε-greedy approach to action selection in reinforcement learning, which of the following values of ε makes the approach identical to a purely greedy approach?
ε = 0