LESSON 9 - Supervised deep learning - discriminative Flashcards
What is the primary use of deep learning, especially in the context of neural networks?
Deep learning is mainly used for classification problems, such as object recognition, and involves hierarchical processing in neural organization.
How does the structure of hierarchical processing in the brain differ from traditional neural network architectures?
Unlike the traditional neural network architectures with limited hierarchy (input – hidden neuron – output), the brain’s circuits are more sophisticated, involving many layers of processing.
What neuroscientific model reflects the processing specialized in object recognition, and how was it implemented?
The neuroscientific model reflecting object recognition processing involves simple cells extracting features from an image, which are then picked up by complex cells. It was implemented as a computer simulation without learning.
What hindered the application of deep learning in the past, and what changed to make it successful?
Limited computational power in the past hindered deep learning. Success was facilitated by the availability of parallel computing architectures, particularly Graphic Processing Units (GPUs), which significantly enhanced computational capabilities.
How does deep learning differ in its approach to supervised learning compared to other machine learning methods?
In deep learning, learning is typically run end-to-end, meaning the model learns directly from raw data (e.g., pixels) to object classes without the need for pre-processing or feature extraction.
What is the role of Rectified Linear Units (RELU) in addressing the problem of weak gradients in deep learning?
RELU is used to replace the Sigmoid activation function, providing a linear response and preventing the weakening of gradients during backpropagation, which accelerates learning.
What challenges arise when dealing with many hidden layers in deep learning, and how can overfitting be addressed?
With many hidden layers, backpropagating errors becomes challenging. Overfitting, due to the model’s complexity, is tackled by employing stronger regularization methods, such as weight decay and dropout.
What is the significance of sparse connectivity in convolutional neural networks (CNNs)?
Sparse connectivity, as seen in CNNs, means not every neuron in one layer is connected to a hidden neuron in the next layer. This approach is similar to the receptive field idea and is crucial for pattern detection.
How are convolutional layers implemented in CNNs, and what do they capture?
Convolutional layers use filters to convolve over input images, capturing features through weighted sums. These features, such as oriented edges, are crucial for recognizing patterns.
What is the purpose of pooling layers in CNNs?
Pooling layers reduce the dimensionality of images and emphasize salient features. The pooling operation, like max pooling, helps compress information while highlighting important aspects of the image.
How does the soft-max activation function in the output layer of a neural network work, and what does it represent?
The soft-max activation function ensures that all activations sum up to 1, allowing the output to be interpreted as probabilities. It represents the confidence of the network in various output classes.
What is the role of weight sharing in CNNs, and how does it contribute to the network’s architecture?
Weight sharing, where neurons share the same connections, is a key aspect of CNNs. It helps capture specific features from different portions of the input image, enhancing the network’s ability to recognize patterns.
What was the significance of LeCun’s convolutional neural network, developed before the deep learning era?
LeCun’s convolutional neural network was a pioneering model, particularly working on the MNIST problem. It laid the foundation for convolutional architectures and is considered a precursor to deep learning.
What marked the start of deep learning in image processing and vision in 2012, and which neural network was instrumental?
AlexNet, developed in 2012, marked the beginning of deep learning in image processing and vision, representing a game-changer for the field.
How did GoogLeNet contribute to the progress of deep learning, and what characterized its architecture?
GoogLeNet, known for its sophistication, contributed significantly to deep learning progress. Training required weeks on supercomputers, highlighting the model’s complexity.