GAMLE SPM Flashcards
Artificial neural networks are?
Parallel distributed computing systems
In the delta rule, the weight changes are given by?
By the difference between target and output, multiplied by the input
In simulated annealing
At high temperatures the network randomly explores the state space
Difference between batch and online (or mini batch learning?
Gradient descent against the global error function in batch learning (computed first across all patterns and then weights are changed.
Stochastic gradient descent against the partial error function (i.e. weights are changed according to the gradient computed for a single example (or for a small number of examples in the minibatch)
What does AUC mean, range in between and what does 1 in AUC mean?
Area under the curve. Ranges between 0 and 1. Represents how good the classifier is, where AUC means the predictions are 100 pct. correct (no false positives, 100 pct. true positives).
What are variational autoencoders?
Variational Autoencoders (VAEs) are a type of generative model in the field of machine learning and artificial intelligence. They belong to the broader family of autoencoders but have distinctive features, particularly in generating new data samples.
what is associative learning (hopfield networks)?
Associative learning, particularly in the context of Hopfield Networks, refers to a type of artificial neural network that is capable of learning and recalling associations or patterns. Hopfield Networks are recurrent neural networks with the ability to store and retrieve specific patterns or memories.
What is a discount factor / gamma?
The discount factor, often denoted by the symbol “γ” (gamma), is a parameter used in reinforcement learning algorithms. It determines the importance of future rewards in the decision-making process of an intelligent agent. The discount factor is a value between 0 and 1.
When an agent makes decisions in a sequential environment, it often receives rewards at each time step. The discount factor allows the agent to weigh immediate rewards against future rewards. A discount factor of 1 means that the agent considers future rewards with equal importance to immediate rewards, while a discount factor less than 1 gives less weight to future rewards.
How can a gamma / discount factor of 1 be useful?
(Reinforcement learning)
In practical terms, a discount factor less than 1 encourages the agent to prioritize short-term rewards, making its planning horizon finite. This is often useful in scenarios where the long-term consequences of actions are uncertain or less relevant.
Handling Infinite Horizons:
When dealing with infinite horizons, a discount factor less than 1 ensures that the sum of future rewards converges to a finite value. This is particularly important in mathematical formulations of reinforcement learning algorithms.
Question 1: What does it mean when a problem is not linearly separable, and which models are used to address this issue?
a) The problem is easy to solve with linear models.
b) The problem is complex and requires non-linear models.
c) The problem has no solution; no models can address it.
d) The problem is linear, and any model can be applied.
Answer 1: The correct answer is b) The problem is complex and requires non-linear models.
Question 2: What is the purpose of the Hebb rule?
a) To suppress neural activation
b) To reinforce neurons with correlated connections
c) To randomize weight changes
d) To activate hidden neurons randomly
Answer 2: The correct answer is b) To reinforce neurons with correlated connections.
Question 3: In the delta rule, what does the weight change depend on?
a) The learning rate
b) The input
c) The difference between target and output multiplied by bias
d) The activation function
Answer 3: The correct answer is c) The difference between target and output multiplied by bias.
Question 4: In contrastive divergence, what method is effective for generalization?
a) Backpropagation
b) Pruning
c) Simulated annealing
d) Reinforcement learning
Answer 4: The correct answer is b) Pruning.
How do Boltzmann Machines utilize hidden units to learn higher-order correlations in the data?
Boltzmann Machines use hidden units to learn higher-order correlations in the data by employing stochastic activations, Gibbs sampling, and simulated annealing during the iterative update process.
What is the positive phase in the training of Restricted Boltzmann Machines using contrastive divergence?
The positive phase involves presenting the pattern to the network, clamping it to the visible neurons, and computing the correlations between all visible and hidden neurons.