Lecture 10 - Recognition Flashcards

1
Q

What is image classification?

A

Image classification is the task of assigning a label from a predefined set of categories to an entire image.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Describe the architecture of LeNet.

A

LeNet is one of the first convolutional networks designed for handwritten digit recognition, featuring layers of convolutional filters followed by subsampling (pooling) and fully connected layers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is AlexNet and its significance?

A

AlexNet is a convolutional network that won the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) in 2012, significantly outperforming previous models with its deep architecture and use of ReLU activation and dropout.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Explain dropout in neural networks.

A

Dropout is a regularization technique where a random subset of neurons is ignored during training, preventing overfitting by ensuring the model does not rely too heavily on any single neuron.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is data augmentation and its purpose?

A

Data augmentation artificially enlarges the training dataset by applying label-preserving transformations such as cropping, flipping, and color jittering to improve model generalization.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Describe the purpose of batch normalization.

A

Batch normalization normalizes the inputs of each layer to have a mean of zero and variance of one, accelerating training and improving model performance by reducing internal covariate shift.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the key features of VGGNet?

A

VGGNet uses small 3x3 convolutional filters stacked together, leading to deep networks with more non-linearities, which improve performance while maintaining computational efficiency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Explain the Inception module in GoogLeNet.

A

The Inception module in GoogLeNet combines multiple convolutional filters of different sizes and pooling operations within the same layer, allowing the network to capture features at various scales.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is a Residual Network (ResNet)?

A

ResNet introduces residual blocks with shortcut connections that bypass one or more layers, allowing very deep networks to be trained by mitigating the vanishing gradient problem.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Describe the concept of feature learning in CNNs.

A

Feature learning in CNNs refers to the automatic extraction and learning of hierarchical features from raw input images, replacing the need for manual feature engineering.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the role of the softmax function in CNNs?

A

The softmax function is used in the output layer of CNNs for multi-class classification, converting raw output scores into probabilities by normalizing them.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Explain the concept of transfer learning in CNNs.

A

Transfer learning involves using a pre-trained network on a large dataset and fine-tuning it on a smaller, task-specific dataset, leveraging the learned features to improve performance and reduce training time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Write the formula for the softmax function.

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Provide the formula for the ReLU activation function.

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the formula for updating weights using gradient descent in neural networks?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How did AlexNet overcome overfitting?

A

AlexNet used dropout on the fully connected layers and data augmentation techniques to overcome overfitting.

17
Q

What are the benefits of using small convolutional filters in VGGNet?

A

Small convolutional filters (3x3) in VGGNet allow for deeper networks with more non-linearities, improving feature learning while maintaining computational efficiency.

18
Q

Explain how the Inception module improves performance in GoogLeNet.

A

The Inception module improves performance by combining multiple convolutional filters and pooling operations, capturing features at various scales and reducing computational complexity.

19
Q

Why are residual connections important in ResNet?

A

Residual connections in ResNet allow the network to bypass certain layers, helping to prevent the vanishing gradient problem and enabling the training of very deep networks.

20
Q

How does batch normalization accelerate training?

A

Batch normalization accelerates training by normalizing the inputs of each layer, reducing internal covariate shift, and allowing for higher learning rates.

21
Q

Describe the role of data augmentation in improving model generalization.

A

Data augmentation improves model generalization by increasing the diversity of the training data through label-preserving transformations, helping the model learn to be invariant to various image distortions.

22
Q

What is the significance of using multiple GPUs in training AlexNet?

A

Using multiple GPUs allowed AlexNet to train faster and handle larger models by dividing the workload, making it feasible to train on large datasets like ImageNet.

23
Q

How does dropout prevent overfitting in neural networks?

A

Dropout prevents overfitting by randomly dropping neurons during training, ensuring the network does not rely too heavily on any single neuron and promoting redundancy.

24
Q

What are the main contributions of the LeNet architecture?

A

LeNet was one of the first successful applications of convolutional networks, demonstrating their effectiveness in recognizing handwritten digits with high accuracy and efficiency.

25
Q

Describe the process of transfer learning in the context of CNNs.

A

Transfer learning involves taking a pre-trained CNN on a large dataset, such as ImageNet, and fine-tuning it on a smaller, task-specific dataset to leverage learned features and improve performance.

26
Q

Explain the concept of feature learning in CNNs.

A

Feature learning in CNNs refers to the automatic extraction and learning of hierarchical features from raw input images through convolutional layers, enabling the network to learn relevant patterns for classification tasks.

27
Q

What challenges do deeper networks face and how are they addressed?

A

Deeper networks face challenges like vanishing/exploding gradients, slower training, and overfitting. These are addressed with techniques such as residual connections, batch normalization, dropout, and data augmentation.