Session 7.1 Flashcards

1
Q

Deep Learning consists in…

A

training a neural network in which the inputs and outputs are the same

  • Input and output are the same image
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Hidden layers have fewer perceptrons

A
  • Each perceptron in a hidden layer has to represent a more elaborate
    concept
  • Features are automatically created: no need to define them manually
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Deep learning uses relatively old methods:

Neural networks have been around for a long time (1950s).

Why now?

A

Big Data

  • Large amounts of data are now available
  • Possibility of using thousands of machines to solve a single problem
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Underfitting

A

A model that is too simple does not fit the data well (high bias)

e.g., fitting a quadractic function with a linear model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Overfitting

A

A model that is too complex fits the data too well (high variance)

e.g., fitting a quadractic function with a 3rd degree function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Bias

A

a model that underfits is wrong many times (high bias) but is not highly affected by slightly different training data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Variance

A

a model that overfits is right on average, but is highly sensitive to specific training data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

We can reduce variance by

A

by putting many models together and aggregating their outcomes (without necessarily increasing bias)

-> this is the concept of ensemble methods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Ensemble methods use

A

multiple algorithms to obtain better predictive performance than could be obtained from any of the algorithms by itself

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Using multiple algorithms usually increases model performance by:

A

reducing variance: models are less dependent on the specific training data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Bagging (or bootstrap aggregation)

A

Creates multiple data sets from the original training data by bootstrapping – re-sample with repetition.

Runs several models and aggregates output with a voting system

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Other ensemble methods

Random Forest

A

combines bagging with random selection of features (or predictors)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Other ensemble methods

Boosting

A

applies classifiers sequentially, assigning higher weights to observations that have been mis-classified by the previous methods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

We can use predictive models for predicting outcomes based on individual attributes

However, models based only on observational data…

A

do not inform how users would react to a specific intervention

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

In most cases the best way to find out the effect of a specific intervention is to run an experiment:

A

1 Randomly assign customers to different treatment groups

2 Compare differences in behavior among treatment groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly