2) Entropy and Fisher Information Flashcards
What is the surprise / shannon information of an event
โlog(๐)
What is the log-odds ratio
What is Entropy
A quantatitive measure of the distribution of probability mass. Low entropy means probability mass is concentrated
What is Shannon Entropy of the Distribution P
What are the Bounds for Shannon Entropy
What is the difference between Shannon Entropy and Differential Entropy
- Shannon entropy is only defined for discrete random variables
- Differential Entropy results from applying the definition of Shannon entropy to a continuous random variable
- Differential entropy is not bounded below by zero and can be negative
What is Differential Entropy
What does the size of the entropy imply
Large entropy implies that the distribution is
spread out whereas small entropy means the distribution is concentrated
What are maximum entropy distributions and what do they signify about a random variable
Maximum entropy distributions are considered minimally informative, indicating the distribution is highly spread out and thus less informative about the random variableโs outcomes
Key Examples Include -
* The discrete uniform distribution, representing maximum entropy among all discrete distributions.
* The normal distribution, as the maximum entropy distribution for continuous variables over [โโ, โ] with specified mean and variance.
* The exponential distribution, for continuous distributions on [0, โ] with a given mean.
What is Cross-Entropy
What is the Cross Entropy of Discrete and Continous Variables
What is the properties of cross-entropy between two distributions ๐น and ๐บ
- Cross-entropy is not symmetric with regard to ๐น and ๐บ, because the expectation is taken with reference to ๐น
- If both distributions are identical cross-entropy reduces to Shannon and differential entropy, respectively: ๐ป(๐น, ๐น) = ๐ป(๐น).
What is Gibbโs Inequality
What is KL Divergence / Relative Entropy
What does KL Divergence measure
- ๐ทKL(๐น, ๐บ) measures the amount of information lost if ๐บ is used to approximate ๐น.
- If ๐น and ๐บ are identical (and no information is lost) then ๐ทKL(๐น, ๐บ) = 0