Chapter 1: Fundamentals of Deep Learning Flashcards

1
Q

Concise AI Definition

A

The effort to automate intellectual tasks normally performed by humans

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Symbolic AI

A

The approach towards creating human-level artificial intelligence by having programmers handcraft a sufficiently large set of explicit rules for manipulating knowledge

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What was the dominant paradigm in AI from the 1950s to the late 1980s?

A

Symbolic AI

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What did Symbolic AI do well? and poorly?

A

did well with well-defined logical problems ex chess not well with fuzzy complex problems ex image classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What replaced Symbolic AI

A

Machine Learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What question gave rise to machine learning?

A

Could a computer go beyond what we know how to order it to perform? and learn on its own to perform a specified task?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the classical programming approach?

A

Humans input rules (a program) and data to be processed according to these rules, yielding answers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Machine Learning steps

A

Humans input data and the answers to the data and the computer outputs the rules that connect the two. These rules can then be applied to new data for original answers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Are machine learning systems programmed?

A

No, they’re trained on data in which it finds statistical structure.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the relationship between Deep Learning, AI and Machine Learning?

A

(AI (ML (DL)))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Predefined set of operations from which ML selects its transformation?

A

Hypothesis Space

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is Deep Learning

A

an approach towards learning representations from data that puts an emphasis on learning successive layers of increasingly meaningful representations. Information-distillation via successive layers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The layered representations in Deep Learning are almost always learned via what models?

A

neural networks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What determines the change a deep learning layer affects on its input?

A

Its weight, the change implemented by the layer is parameterized by its weights

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Loss Function of the network

A

measures how far an output is from what is expected, capturing how well the network does

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

The central algorithm in Deep Learning

A

The Backpropagation algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What happens to a network’s loss score?

A

It is used by the optimizer to implement the backpropagation algorithm to adjust the weights of the layers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the training loop for Deep Learning?

A

The process of starting with random weight values and through repeated feedback from the loss score adjusting layer weights

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is the goal of the training loop?

A

to find the layer weights that minimize the loss score

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Probabilistic Modeling

A

An early form of Machine Learning. The application of the principles of statistics to data analysis

21
Q

What is the best-known algorithm in Probabilistic Modeling?

A

Naive-Bayes

22
Q

Naive Bayes

A

type of ML classifier based on applying Baye’s theorem while assuming that the features of in the input data area all independent (a strong/naive assumption)

23
Q

Logistic Regression

A

a CLASSIFICATION algorithm not regression. Closely related to Naive Bayes, a type of probabilistic modeling

24
Q

What held neural networks back for a long time?

A

The missing piece was an efficient way to train large neural networks

25
Q

What is the Backpropagation Algorithm

A

A way to train chains of parametric operations using gradient-descent optimization

26
Q

Kernel Methods

A

a group of classification algorithms, the best known is SVM

27
Q

Support Vector Machine (SVM)

A

A classification algorithm which solves classification problems by trying to find good decision boundaries between categories.

28
Q

How do SVMs decide on a decision boundary?

A
  1. data mapped onto hyperplane 2. try to compute the decision boundary that will best “maximize the margin” or maximize the distance between the closest points from each class
29
Q

Kernel Trick

A

What gives kernel methods their name since otherwise they would often be computationally intractable. You dont have to compute coordinates of your points in new space just the distance between pairs of points; done via Kernel Function

30
Q

Decision Trees

A

Flowchart-like, successive questions, allow for input classification or predict output values given inputs

31
Q

Random Forest

A

A specific Decision tree algorithm involves building a large number of specialized decision trees and ensembling their outputs.

32
Q

Gradient Boosting Machines

A

a ML technique based on ensembling weak prediction models, generally decision trees. “gradient boosting’ = Iteratively training new models that specialize in addressing the weak points of the previous models

33
Q

Go-to algorithm for computer-vision tasks and generally all perceptual tasks

A

Deep Convolutional Neural networks

34
Q

What were the key factors in Deep learning’s success?

A
  1. That it offered better performance on a wide array of problems. 2. Was much easier because it completely automates what used to be the most crucial step in ML workflow: Feature Engineering. in DL you learn all the features in one pass rather than having to reengineer them yourself
35
Q

Why can’t multiple shallow-layer ML methods be applied in sequence to replicate Deep learning?

A

Because the optimal first representation layer in a three-layer model isn’t the optimal first layer in a one-layer or two-layer model.

36
Q

What is transformative about deep learning?

A

it allows a model to learn al layers of representation JOINTLY, at the same time, rather than in succession (greedily)

37
Q

Deep Learning could also be called?

A

Joint Learning, layers adjust together as any layer changes

38
Q

2 essential characteristics of how Deep Learning learns from Data?

A
  1. the incremental, layer by layer way in which increasingly complex representations are developed 2. that the intermediate incremental representations are learned jointly
39
Q

Two most important ML techniques to be familiar with today and their most effective uses

A

gradient-boosting machines for shallow-learning problems where structured data is available. Deep Learning for perceptual problems

40
Q

Dominant Deep Learning Library?

A

Keras

41
Q

Dominant gradient-boosting library?

A

XGBoost

42
Q

The fundamental algorithm for Deep Learning in time series

A

The Long Short-term Memory Algorithm (LSTM)

43
Q

Why are deep neural networks parallelizable?

A

Because they perform mostly many small matrix multiplications

44
Q

TPU

A

Tensor Processing Unit. A chip designed and built by google from the ground up specifically for deep neural networks.

45
Q

What algorithmic improvements allowed for better gradient-propagation circa 2010

A
  1. better activation functions for neural layers
  2. better weight-initialization schemes
  3. Better optimization schemes (RMSProp and Adam)
46
Q

Theano and Tensorflow

A

two symbolic tensor-manipulation frameworks for python that support autodifferentiation simplifying the implementation of new models

47
Q

What makes Deep Learning Neural nets so scalable?

A

Highly amenable to parallelization, DL models trained by iterating over small batches of data, so they can be trained on data sets of arbitrary size

48
Q

What makes Deep Learning so much simpler than other Machine Learning techniques?

A

No need for feature engineering