T1: 4. Information Theory Flashcards
Define Shannon entropy (words)
A measure of how unsure we are about the value of X, or on average how much information we gain when we learn the value of X.
Define Shannon entropy (formula)
H(X) = -SUM_x[p(x)log(p(x)]
where x = {finite set}
What is 0log(0) in this case?
0
Define Shannon’s Noiseless Coding Theorem
The Shannon entropy gives a lower bound for the number of bits per letter required to encode a message.
What probability would specify N yes or no questions to Shannon entropy?
p(x) = (1/2)^n
What distribution of probabilities maximises H(X)
p(x_n)=1/n
Equal probability for each outcome.
Give the formula for joint entropy
H(X,Y) = -SUM_x,y[p(x,y)log(p(x,y)]
where x = {finite set}
How does joint entropy relate to individual entropies of two events X and Y?
Under what condition is this equality?
H(X,Y) <= H(X) + H(Y)
If X and Y are independent, with p(x,y)=p(x)p(y).
Where is relative entropy useful?
Where we have the two random variables which take the same values but with different probability distributions.
State the relative entropy formula H(p(x)||q(x))
- H(X) - SUM_x p(x) log q(x)
Under what case is relative entropy =0?
p(x) = q(x)
Define the Von-Neumann entropy
S(ρ) = Tr (ρ logρ) = SUM_i ei log ei
where ei are eigenvalues of ρ
Define the relative Von-Neumann entropy S(ρ_1 ||ρ_2)
Tr (ρ_1 log ρ_1) - Tr (ρ_1 log ρ_2) >= 0
Define conditional uncertainty H(X|Y)
This is the remaining uncertainty in X given we know the value of Y
State the conditional uncertainty formula H(X|Y)
H(X|Y) = H(X,Y) - H(Y)