Definition Measures how much knowing one variable reduces uncertainty about another: $$ I(X;Y) = \sum_{x,y} p(x,y) \log \frac{p(x,y)}{p(x)p(y)} $$ Properties $I(X;Y) = I(Y;X)$ (symmetric) $I(X;Y) \geq 0$ (non-negative) $I(X;X) = H(X)$ (self-information is entropy) $I(X;Y) = 0$ if $X$ and $Y$ are independent Python …
Read More