Neural Networks Flashcards
What is a neural network?
A neural network is a computational model inspired by the organization and functioning of the human brain. It is a core component of artificial intelligence (AI) and machine learning, designed to process and analyze complex data, recognize patterns, and make decisions or predictions. Neural networks consist of interconnected nodes, called neurons, organized into layers. These networks can be used for a wide range of tasks, including image recognition, natural language processing, and autonomous control.
What are neural networks related to AI and machine learning?
Neural networks play a pivotal role in the fields of AI and machine learning. They are the foundation for many AI applications, providing the ability to learn from data and adapt to new information. Neural networks excel at tasks such as image classification, speech recognition, and recommendation systems, making them a fundamental technology in the advancement of AI and machine learning.
What is a perceptron?
A perceptron is one of the simplest forms of a neural network. It serves as a binary linear classifier, taking multiple inputs, applying weights to those inputs, summing them up, and then passing the result through an activation function to produce an output. Perceptrons are often used as building blocks to understand the basic principles of neural network operation.
How is a perceptron related to neural networks?
Perceptrons are building blocks for more complex neural networks. While they are limited to linearly separable problems, they provide a foundational understanding of how neurons process information through weighted inputs and activation functions. Neural networks extend the concept of perceptrons by adding multiple layers and non-linear activation functions, enabling them to solve more intricate problems.
How do neurons in a neural network work? What are the characteristics?
Neurons in a neural network receive inputs, each multiplied by a weight, which determines their contribution to the neuron’s output. These weighted inputs are summed, a bias term is added, and the result is then passed through an activation function. The activation function introduces non-linearity, allowing the neuron to capture complex relationships in the data. This process is fundamental to the operation of neurons in a neural network.
What is a bias in a neural network?
A bias is an additional parameter associated with each neuron in a neural network. It allows the neuron to shift the activation function’s output. The bias term is crucial for fine-tuning a neuron’s responsiveness and ensuring that the network can model various patterns and data distributions effectively.
What are weights in neural networks?
Weights are parameters assigned to the connections between neurons in a neural network. They determine the strength of influence that each input has on a neuron’s output. Adjusting these weights during training is how the network learns to make accurate predictions or classifications.
What is an activation in a neural network?
Activation in a neural network refers to the output value produced by a neuron after processing its inputs. It represents the neuron’s response to the weighted sum of inputs and the activation function applied to that sum.
What is an activation function in a neural network?
An activation function is a mathematical function applied to the weighted sum of inputs and bias in a neuron. It introduces non-linearity into the neuron’s output, enabling the network to model complex, non-linear relationships in the data. Common activation functions include sigmoid, ReLU (Rectified Linear Unit), and tanh (hyperbolic tangent).
What are some different activation functions? What are the pros and cons of each?
There are several activation functions used in neural networks. Sigmoid functions produce outputs in the range (0, 1), which can represent probabilities. ReLU functions are computationally efficient and help mitigate the vanishing gradient problem but can suffer from the dying ReLU problem. Tanh functions produce outputs in the range (-1, 1) and are zero-centered. Each activation function has its advantages and disadvantages, making their choice task-dependent.
Given a bunch of neurons, how is a neural network constructed?
A neural network is constructed by organizing neurons into layers. Typically, there is an input layer to receive data, one or more hidden layers to process and learn from the data, and an output layer to produce the final results or predictions. The connections between neurons are defined by weights, and each neuron has its activation function and bias.
What are feed-forward neural networks?
Feed-forward neural networks, also known as multilayer perceptrons (MLPs), are a type of neural network architecture where information flows in one direction, from the input layer to the output layer. These networks do not have feedback loops or connections that create cycles, making them suitable for a wide range of supervised learning tasks.
How should weights be initialized?
Weights in a neural network are typically initialized randomly, but careful consideration must be given to the initialization method. Techniques like Xavier/Glorot and He initialization are often used to promote efficient training by ensuring gradients do not vanish or explode during backpropagation.
What are deep networks?
Deep networks refer to neural networks with multiple hidden layers. The depth of a network allows it to capture increasingly abstract and hierarchical features in the data, making it highly effective for complex tasks such as image recognition and natural language processing.
What is the width and depth of a network?
Width refers to the number of neurons in a layer, while depth indicates the number of layers in a neural network. The combination of width and depth determines the network’s capacity to represent and learn from data.