Entropy Flashcards
What is the entropy of a random variable X?
H(X) = -ΣxP(x)log2(P(x))
Entropy of a uniform distribution of N outcomes
log2(N)
What has more entropy? A non-unfiorm distribution over N outcomes or a uniform distribution over N outcpmes?
Any non-uniform distribution over N outcomes has lower entropy than the corresponding uniform distribution
What is cross-entropy and what does cross-entropy model/measure?
(-1/n)log(P_{M}(W)
How well a model M predicts the data
What is perplexity and what does it measure?
2^{cross-entropy}, it models the same thing as cross-entropy, how well a probability distribution or probability model predicts a sample.
Better models q of the unknown distribution p will tend to assign higher probabilities q(xi) to the test events. Thus, they have lower perplexity: they are less surprised by the test sample.