Mutual Information
Definition
Measures how much knowing one variable reduces uncertainty about another:
$$ I(X;Y) = \sum_{x,y} p(x,y) \log \frac{p(x,y)}{p(x)p(y)} $$
Properties
- $I(X;Y) = I(Y;X)$ (symmetric)
- $I(X;Y) \geq 0$ (non-negative)
- $I(X;X) = H(X)$ (self-information is entropy)
- $I(X;Y) = 0$ if $X$ and $Y$ are independent
Python
1from sklearn.metrics import mutual_info_score
2
3# Discrete variables
4mi = mutual_info_score(x, y)
Further Reading
Related Snippets
- Channel Capacity
Shannon's theorem and noisy channels - Data Compression
Lossy vs lossless compression, Huffman coding - Entropy & Information Measures
Shannon entropy, cross-entropy, and KL divergence - Information Theory Basics
Fundamental concepts of information theory