week 6 Flashcards
chp 6
What is the primary focus of the document titled ‘Deep Learning for Computer Vision’?
(Think about how deep learning relates to visual data processing.)
cp vision overview (key concept)
The primary focus is on the application of deep learning techniques in the field of computer vision.
explanation: Deep learning is a subset of machine learning that uses neural networks with many layers to analyze various forms of data, including images, which is central to computer vision.
What is the purpose of Data Augmentation in deep learning? (Think about how we can create more data from existing data.)
deep learning intro (key concept)
To improve model performance and generalization by artificially increasing the size of the training dataset.
Explanation: Data Augmentation involves techniques like rotating, flipping, or cropping images to create variations, which helps the model learn better and become more robust to changes in input.
What does Transfer Learning allow you to do in deep learning?(Consider how knowledge from one task can help with another task.)
It allows you to apply pre-trained models to new tasks with limited data, speeding up the learning process.
explanation: Transfer Learning leverages the knowledge gained from training on a large dataset to improve performance on a smaller, related dataset, making it easier to train models when data is scarce.
What is Invariant Feature Learning in the context of deep learning?(Think about features that do not change even if the input data changes.)
It refers to the ability of models to identify and utilize latent patterns in data that remain consistent across variations.
explanation: Invariant Feature Learning helps models generalize better by focusing on essential characteristics of the data, regardless of transformations like scaling or rotation.
What is the manifold hypothesis in the context of hidden structures in data?(Think about how complex data can often be simplified into lower dimensions.) hidden struc (key concept)
The manifold hypothesis states that natural data resides in a low-dimensional (non-linear) manifold due to the mutual dependence of variables in the data.
Explanation: The manifold hypothesis suggests that while data may appear high-dimensional (like images with many pixels), it can often be represented in a lower-dimensional space where the underlying structure is more apparent.
How does the ideal feature extractor relate to hidden structures in high-dimensional data?
(Consider how features can be separated into meaningful components.)
The ideal feature extractor captures the coordinates of data points on the manifold and their positions away from it, effectively disentangling the factors of variation.
Explanation: An ideal feature extractor identifies the essential characteristics of data, allowing for better understanding and classification by focusing on relevant dimensions while ignoring irrelevant variations.
What are invariant features and why are they important in learning hidden structures?(Think about how we recognize objects regardless of their position or size in an image.)
Invariant features are stable characteristics that remain recognizable despite transformations like translation or scaling, crucial for effective representation learning.
(Explanation: Invariant features help in creating robust models that can generalize well across different variations of input data, making them essential for understanding hidden structures in complex datasets.)
What is a Deep Neural Network (DNN)?
(Think about how many layers are involved in processing data.)
A Deep Neural Network is a type of artificial neural network with multiple layers between the input and output layers.
Explanation: DNNs consist of an input layer, several hidden layers, and an output layer, allowing them to model complex relationships in data.
What is the purpose of training a Deep Neural Network?(Consider what happens during the learning process of a model.)
The purpose of training a DNN is to adjust its weights to minimize the difference between predicted and actual outputs.
Explanation: Training involves using a dataset to optimize the DNN’s parameters through techniques like gradient descent, which iteratively reduces the error.
What is Gradient Descent in the context of DNN training?(Think about how you would find the lowest point on a hill.)
Gradient Descent is an optimization algorithm used to minimize the loss function by updating the model’s weights in the opposite direction of the gradient.
Explanation: By calculating the gradient of the loss function, Gradient Descent helps in finding the optimal weights for the DNN, leading to better predictions.
What is the purpose of backpropagation in neural networks?(Think about how errors are adjusted in a network.)
To compute the contribution of each node to the overall error using the chain rule. Explanation: Backpropagation uses the chain rule to propagate error derivatives backwards through the network, allowing for the calculation of how much each node contributed to the error, which is essential for updating weights during training.
What is a potential issue with backpropagation in deep neural networks?(Consider what happens to gradients as they are propagated back through many layers.)
The vanishing gradient problem, which can hinder learning in lower layers. (Explanation: The vanishing gradient problem occurs when gradients become very small as they are propagated back through the layers of a neural network, making it difficult for lower layers to learn effectively.)
How does gradient descent relate to backpropagation?(Think about the process of minimizing error in training.)
Gradient descent is used to update the weights based on the error computed through backpropagation.
explanation: Gradient descent is an optimization algorithm that adjusts the weights of the neural network to minimize the error calculated during backpropagation, effectively improving the model’s performance.
What is the role of the learning rate (𝜖) in backpropagation?(Think about how quickly or slowly the model learns from the data.)
The learning rate controls how much to update the weights during training.
Explanation: The learning rate (𝜖) is a crucial hyperparameter in backpropagation that determines the size of the steps taken towards the minimum of the loss function. A fixed learning rate remains constant throughout training, while an adaptive learning rate adjusts based on the gradient history, allowing for faster convergence.
What is the purpose of momentum (𝜈) in the context of backpropagation?(Consider how it relates to physical motion and inertia.)
Momentum helps the gradient descent algorithm to maintain its direction and speed up convergence.
Explanation: Momentum in backpropagation is a technique that helps accelerate gradient descent by adding a fraction of the previous update to the current update. This allows the optimization process to continue moving in the same direction, reducing oscillations and speeding up convergence.