Essentials Flashcards
What is the formula for the entropy of a discrete probability distribution?
What is the formula for KL-divergence for two probability distributions?
What is the formula for the entropy of a continuous probability distribution?
What is the formula for the KL-divergence of a continuous probability distribution?
What is the entropy of a Gaussian Distribution?
What is the entropy of a d-dimensional Gaussian distribution?
What is the KL-divergence between two d-dimensional multivariate Gaussian Distributions?
What is the Wasserstein difference for two multivariate Gaussian Distributions?
What is the cross entropy error for a binary classification task?
What is the Gaussian Distribution equation?
What is the multivariate Gaussian Distribution equation?
For softmax, what is Prob(i)?
For softmax what is log Prob(i)
What is the equation for the gradient using softmax?
What is the true value (V*) of the current state?
What is the formula for Q*(s,a)?
What is the formula for number of weights per filter?
(filter width) x (filter height) x (input depth) + 1 (for bias)
Number of neurons in this layer?
(output width) x (output height) x (depth), width= ((input width - filter width)/stride)+1
Number of connections into the neuron’s in a layer?
(num neurons) x (connections per neurons / filter wights - bias)
Number of independent parameters?
(num filters) x (num weights)