chatGPT questions exam2 Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

What does Stochastic Gradient Descent (SGD) optimize in machine learning models?

A

SGD is used to minimize the loss function of a model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How does SGD differ from traditional gradient descent?

A

Unlike traditional gradient descent that uses the entire dataset for updates, SGD updates model parameters using only a single sample or a small batch of samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the key benefits and drawbacks of using SGD?

A

SGD is more computationally efficient for large datasets and can help escape local minima, but it may lead to slower and less stable convergence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does the FROC curve evaluate in medical image analysis?

A

the FROC curve evaluates detection and localization performance, plotting sensitivity against the average number of false positives per image.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why can accuracy be a misleading performance metric?

A

Accuracy can be misleading in cases of class imbalance, where it doesn’t reflect the model’s performance on the minority class

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What was the major limitation of the original R-CNN in object detection?

A

The original R-CNN was slow due to its reliance on selective search for generating region proposals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How did Fast R-CNN improve over R-CNN?

A

Fast R-CNN improved efficiency by sharing computations across region proposals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What innovation did Faster R-CNN introduce to object detection?

A

Faster R-CNN introduced a Region Proposal Network (RPN), allowing for end-to-end training and faster proposal generation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How does setting a high threshold for predictions affect sensitivity and specificity?

A

A high threshold generally increases specificity but may decrease sensitivity by rejecting true positives with lower confidence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What approach do convolutional networks use for segmentation tasks?

A

They utilize architectures with downsampling and upsampling layers, and sometimes strided or dilated convolutions, to produce a pixel-wise segmentation map.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How is the number of trainable parameters in a convolutional layer calculated?

A

The formula is (filter height × filter width × input channels + 1) × number of filters, where “+1” accounts for the bias term.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Why is the learning rate considered a critical hyperparameter in neural network training?

A

t influences the training dynamics, where too high a rate may cause divergence and too low a rate results in slow convergence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does dropout do in neural networks?

A

Dropout prevents overfitting by randomly dropping units and their connections during training, simulating training multiple networks in parallel.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the advantages of ReLU over sigmoid functions?

A

ReLU mitigates the vanishing gradient problem, accelerates convergence, and maintains gradient flow for positive inputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Why is feature normalization important before training a machine learning model?

A

it scales features to have zero mean and unit variance, ensuring consistent scaling and preventing information leakage from the test set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How do residual networks (ResNets) facilitate the training of deep models?

A

resNets use skip connections to learn residual mappings, alleviating the vanishing gradient problem and enabling deeper architectures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What does the receptive field refer to in convolutional networks?

A

It refers to the size of the input area that influences the network’s output, determined by the cumulative effect of convolution and pooling operations.

18
Q

How does the softmax function work in neural networks?

A

It converts raw output scores into probabilities by taking the exponential of each output and normalizing these values by the sum of all exponentials.

19
Q

What does a false negative (FN) indicate in detection tasks?

A

A FN occurs when an object is present but not detected by the classifier, which is critical in applications where missing detections can have serious consequences.

20
Q

What is a dilated (atrous) convolution and its purpose?

A

It increases the receptive field without increasing the number of weights by inserting spaces between kernel elements, allowing for broader spatial aggregation.

21
Q

How does U-Net architecture specialize for medical image segmentation?

A

U-Net uses a symmetric architecture with downsampling and upsampling paths, enabling precise localization and context integration without needing a pre-defined weight map.

22
Q

How is specificity calculated in a model’s performance evaluation?

A

Specificity is calculated as the number of true negatives divided by the sum of true negatives and false positives.

23
Q

What determines the output size of a valid convolution operation?

A

The output size is determined by (W − F + 1) × (H − F + 1), where W and H are the width and height of the image, and F is the filter size.

24
Q

What makes the YOLO object detection framework unique?

A

YOLO processes the entire image in a single evaluation and makes predictions for each grid cell, combining bounding box predictions and class probabilities.

25
Q

How does backpropagation work in training neural networks?

A

It computes gradients of the loss with respect to network parameters by applying the chain rule, allowing for efficient parameter updates.

26
Q

What are histogram-based features in image processing?

A

These features, like mean, variance, skewness, and kurtosis, describe the distribution of pixel intensities and are invariant to/do not consider spatial relationship and correlation between pixels (identical histograms can belong to different textures)

27
Q

how does the shift-and-stitch algorithm benefit fully convolutional networks?

A

It enables the generation of fine-grained outputs by reconstructing the original resolution through multiple shifted input passes.

28
Q

What improvement does Fast R-CNN offer over the original R-CNN?

A

Fast R-CNN processes the entire image at once and uses RoI pooling to extract features, improving efficiency and speed

29
Q

How does histogram matching adjust an image’s brightness and contrast?

A

It modifies an image so that its histogram matches that of a reference image, using cumulative distribution functions for mapping pixel values.

30
Q

What are the implications of not using pooling layers in CNNs?

A

Omitting pooling layers requires convolutional layers with strides for downsampling, potentially leading to higher memory and computational costs.

31
Q

How does backpropagation facilitate learning in CNNs?

A

It calculates gradients (using chain rule) for all network parameters by propagating errors backward through the network, enabling parameter updates for learning.

32
Q

What are the steps in mini-batch gradient descent training?

A

The steps are selecting a mini-batch, performing a forward pass, computing loss, backpropagating to compute gradients, and updating weights.

33
Q

How does the choice of pooling size and stride affect max-pooling operations?

A

Incompatible pooling sizes and strides may lead to suboptimal coverage and downsampling uniformity in the feature map.

34
Q

What strategy is used in transfer learning with small datasets?

A

Freeze most of the pre-trained network’s parameters and fine-tune the last few layers to adapt the features to the new task.

35
Q

How does dilated convolution affect the receptive field in CNNs?

A

dilated convolution expands the receptive field without increasing the number of weights, allowing for broader spatial information aggregation.

36
Q

What is the purpose of the softmax function in neural networks?

A

it converts logits into probabilities, facilitating multi-class classification by ensuring output values sum to 1.

37
Q

What does the term “dying ReLU” problem refer to?

A

it refers to the issue where neurons stop learning due to always outputting zero, which can stem from improper initialization or gradient descent issues. - this can be solved with leakyrelu

38
Q

How is the number of trainable parameters in a dense layer calculated?

A

It is calculated as (number of inputs + 1) × number of outputs, accounting for each input connection and bias term.

39
Q

What does zero padding do in convolutional layers?

A

Zero padding allows control over output dimensions, maintaining spatial sizes or adjusting them for specific convolutional effects.

40
Q

How is the “best model” identified during neural network training?

A

The best model is often the one with the lowest validation loss, indicating effective learning and generalization capabilities without overfitting.