Today I tried to write about the risk of passing the coronavirus when going to visit someone. It looked nice enough and I'm copying an edited version here, using KaTeX instead of the original LaTeX version I wrote. Also, I was experimenting with unicode symbol support, for which I used XeTeX.
I didn't make up the numbers that appear, but I haven't included the references either. It is just a simple note with no big pretension.
Information content (of a random variable): $I(X) = – \log(P(X))$
Entropy (of a random variable): $H(X) = E[I(X)] = E[– \log(P(X))]$
Conditional Entropy (of a random variable on another): $H(X|Y) = E{PY}[H(X|Y=y)]$
Cross Entropy (between two probability distributions): $H(p, q) = Ep[– \log q] = H(p) + D{KL}(p | q)$
Kullback-Leibler divergence (from probability distribution $q$ to $p$): $D_{KL}(p | q) = H(p, q) – H(p)$
The Kullback-Leibler divergence is surprisingly useful in many places. For example, it can be something to maximize (its expected value), as a criterion for experimental design. I wish I had known more about it during the time of writing my thesis.
I wish I'd write about the interpretation of $D_{KL}$ as a measure of the information gained when one changes from using the probability distribution $q$ to $p$. But if I do, I'll never click the “Publish” button :)