week 6 Flashcards

chp 6

1
Q

What is the primary focus of the document titled ‘Deep Learning for Computer Vision’?
(Think about how deep learning relates to visual data processing.)
cp vision overview (key concept)

A

The primary focus is on the application of deep learning techniques in the field of computer vision.
explanation: Deep learning is a subset of machine learning that uses neural networks with many layers to analyze various forms of data, including images, which is central to computer vision.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the purpose of Data Augmentation in deep learning? (Think about how we can create more data from existing data.)
deep learning intro (key concept)

A

To improve model performance and generalization by artificially increasing the size of the training dataset.
Explanation: Data Augmentation involves techniques like rotating, flipping, or cropping images to create variations, which helps the model learn better and become more robust to changes in input.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does Transfer Learning allow you to do in deep learning?(Consider how knowledge from one task can help with another task.)

A

It allows you to apply pre-trained models to new tasks with limited data, speeding up the learning process.
explanation: Transfer Learning leverages the knowledge gained from training on a large dataset to improve performance on a smaller, related dataset, making it easier to train models when data is scarce.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is Invariant Feature Learning in the context of deep learning?(Think about features that do not change even if the input data changes.)

A

It refers to the ability of models to identify and utilize latent patterns in data that remain consistent across variations.
explanation: Invariant Feature Learning helps models generalize better by focusing on essential characteristics of the data, regardless of transformations like scaling or rotation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the manifold hypothesis in the context of hidden structures in data?(Think about how complex data can often be simplified into lower dimensions.) hidden struc (key concept)

A

The manifold hypothesis states that natural data resides in a low-dimensional (non-linear) manifold due to the mutual dependence of variables in the data.
Explanation: The manifold hypothesis suggests that while data may appear high-dimensional (like images with many pixels), it can often be represented in a lower-dimensional space where the underlying structure is more apparent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How does the ideal feature extractor relate to hidden structures in high-dimensional data?
(Consider how features can be separated into meaningful components.)

A

The ideal feature extractor captures the coordinates of data points on the manifold and their positions away from it, effectively disentangling the factors of variation.
Explanation: An ideal feature extractor identifies the essential characteristics of data, allowing for better understanding and classification by focusing on relevant dimensions while ignoring irrelevant variations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are invariant features and why are they important in learning hidden structures?(Think about how we recognize objects regardless of their position or size in an image.)

A

Invariant features are stable characteristics that remain recognizable despite transformations like translation or scaling, crucial for effective representation learning.
(Explanation: Invariant features help in creating robust models that can generalize well across different variations of input data, making them essential for understanding hidden structures in complex datasets.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is a Deep Neural Network (DNN)?
(Think about how many layers are involved in processing data.)

A

A Deep Neural Network is a type of artificial neural network with multiple layers between the input and output layers.
Explanation: DNNs consist of an input layer, several hidden layers, and an output layer, allowing them to model complex relationships in data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the purpose of training a Deep Neural Network?(Consider what happens during the learning process of a model.)

A

The purpose of training a DNN is to adjust its weights to minimize the difference between predicted and actual outputs.
Explanation: Training involves using a dataset to optimize the DNN’s parameters through techniques like gradient descent, which iteratively reduces the error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is Gradient Descent in the context of DNN training?(Think about how you would find the lowest point on a hill.)

A

Gradient Descent is an optimization algorithm used to minimize the loss function by updating the model’s weights in the opposite direction of the gradient.
Explanation: By calculating the gradient of the loss function, Gradient Descent helps in finding the optimal weights for the DNN, leading to better predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the purpose of backpropagation in neural networks?(Think about how errors are adjusted in a network.)

A

To compute the contribution of each node to the overall error using the chain rule. Explanation: Backpropagation uses the chain rule to propagate error derivatives backwards through the network, allowing for the calculation of how much each node contributed to the error, which is essential for updating weights during training.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a potential issue with backpropagation in deep neural networks?(Consider what happens to gradients as they are propagated back through many layers.)

A

The vanishing gradient problem, which can hinder learning in lower layers. (Explanation: The vanishing gradient problem occurs when gradients become very small as they are propagated back through the layers of a neural network, making it difficult for lower layers to learn effectively.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How does gradient descent relate to backpropagation?(Think about the process of minimizing error in training.)

A

Gradient descent is used to update the weights based on the error computed through backpropagation.
explanation: Gradient descent is an optimization algorithm that adjusts the weights of the neural network to minimize the error calculated during backpropagation, effectively improving the model’s performance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the role of the learning rate (𝜖) in backpropagation?(Think about how quickly or slowly the model learns from the data.)

A

The learning rate controls how much to update the weights during training.
Explanation: The learning rate (𝜖) is a crucial hyperparameter in backpropagation that determines the size of the steps taken towards the minimum of the loss function. A fixed learning rate remains constant throughout training, while an adaptive learning rate adjusts based on the gradient history, allowing for faster convergence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the purpose of momentum (𝜈) in the context of backpropagation?(Consider how it relates to physical motion and inertia.)

A

Momentum helps the gradient descent algorithm to maintain its direction and speed up convergence.
Explanation: Momentum in backpropagation is a technique that helps accelerate gradient descent by adding a fraction of the previous update to the current update. This allows the optimization process to continue moving in the same direction, reducing oscillations and speeding up convergence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are some common regularization techniques used in backpropagation to prevent overfitting?(Think about methods that help a model generalize better to unseen data.)

A

Common techniques include weight decay, dropout, and early stopping.
explanation: Regularization techniques are essential in backpropagation to avoid overfitting, which occurs when a model learns noise in the training data rather than the underlying pattern. Techniques like weight decay, dropout, and early stopping help ensure that the model performs well on new, unseen data.

16
Q

What is weight decay in the context of regularization?(Think about how large weights can affect model complexity.)

A

Weight decay adds a penalty to the loss function based on the size of the model’s weights. Explanation: Weight decay discourages overly large weights, which can lead to complex models that overfit the training data by adding a penalty to the loss function.

17
Q

How does dropout help in regularization?(Consider how this technique affects neuron participation in training.)

A

Dropout randomly ignores selected neurons during training to prevent over-reliance on any single neuron. Explanation: By randomly dropping out neurons, dropout prevents the model from becoming overly dependent on specific neurons, promoting a more robust learning process.

18
Q

What is the purpose of data augmentation in regularization? (Think about how changing the training data can help the model learn better.)

A

Data augmentation artificially increases the size and variability of the training set by applying transformations. Explanation: By introducing noise or random transformations to the training data, data augmentation helps the model generalize better by providing a more diverse set of examples.

19
Q

What is data augmentation in the context of deep learning for computer vision?(Think about how we can modify existing images to create new training examples.)

A

Data augmentation refers to techniques used to increase the diversity of training data without actually collecting new data. Explanation: Data augmentation helps improve the robustness of models by artificially expanding the training dataset through various transformations such as flipping, rotating, and adjusting brightness.

20
Q

What are some common techniques used in data augmentation for images?(Consider how you can change an image’s orientation or appearance.)

A

Common techniques include image flipping, rotation, zooming, shifting, and adjusting brightness. Explanation: These techniques help create variations of the original images, allowing models to learn from a wider range of examples and improving their performance.

21
Q

How does rotating an image contribute to data augmentation? (Think about how objects can appear differently when viewed from various angles.)

A

Rotating an image allows the model to learn features from different angles, making it more invariant to orientation changes. Explanation: By including rotated versions of images in the training set, the model can generalize better to new images that may not be aligned in the same way as the training data.

22
Q

What is the purpose of model deployment in deep learning?(Think about how models are used after training.)

A

To make a trained model available for use in real-world applications. Explanation: Model deployment refers to the process of integrating a trained machine learning model into an existing production environment, allowing it to make predictions on new data.

23
Q

What are pre-trained models and why are they useful in model deployment? (Consider how using existing knowledge can save time and resources.)

A

Pre-trained models are models that have been previously trained on a large dataset and can be fine-tuned for specific tasks. Explanation: Pre-trained models allow developers to leverage existing knowledge from large datasets, reducing the time and computational resources needed to train a model from scratch.

24
Q

What is transfer learning in the context of model deployment?(Think about how skills learned in one area can apply to another.)

A

Transfer learning is a technique where a model developed for a specific task is reused as the starting point for a model on a second task. Explanation: In model deployment, transfer learning enables the adaptation of pre-trained models to new but related tasks, improving efficiency and performance.

25
Q

What is Transfer Learning in the context of deep learning?(Think about how knowledge from one task can help with another similar task.)

A

Transfer Learning is a technique where a model developed for a particular task is reused as the starting point for a model on a second task. Explanation: In Transfer Learning, pre-trained models are adapted to new tasks, leveraging previously learned features to improve performance and reduce training time.

26
Q

What are some key components of a Convolutional Neural Network (CNN) architecture used in Transfer Learning? (Consider the layers that process and reduce the dimensionality of input data.)

A

Key components include Convolution layers, Max Pooling layers, Dense layers, and Dropout layers. Explanation: CNN architectures typically consist of convolutional layers for feature extraction, pooling layers for down-sampling, dense layers for classification, and dropout layers to prevent overfitting.

27
Q

What is the purpose of ‘freezing’ layers in a Transfer Learning model? (Think about how you might want to keep certain knowledge intact while learning something new.)

A

Freezing layers prevents their weights from being updated during training, allowing the model to retain learned features from the pre-trained model. Explanation: By freezing certain layers, you can maintain the general features learned from a large dataset while fine-tuning the model on a smaller, task-specific dataset.

28
Q

What is a neural network? (Think about how the brain learns and makes decisions.)

A

A neural network is a computational model inspired by the way biological neural networks in the human brain process information. Explanation: Neural networks consist of interconnected nodes (neurons) that work together to recognize patterns and make predictions based on input data.

29
Q

What is gradient descent? (Consider how you would find the lowest point on a hill.)

A

Gradient descent is an optimization algorithm used to minimize the loss function in machine learning models by iteratively adjusting the model parameters. Explanation: By calculating the gradient (slope) of the loss function, gradient descent updates the parameters in the opposite direction of the gradient to reduce the error

30
Q

What is backpropagation? (Think of it as a way to ‘propagate’ errors backward through the network.)

A

Backpropagation is a method used in neural networks to calculate the gradient of the loss function with respect to each weight by the chain rule. Explanation: This process allows the network to learn by adjusting weights based on the error of the output compared to the expected result.