Mutual Information and Dependence
Assumed audience
- Reading level: technical.
- Background: probability and basic entropy.
- Goal: understand mutual information as shared uncertainty.
Definition
Mutual information measures how much knowing one variable reduces uncertainty about another:
[ I(X;Y) = H(X) + H(Y) - H(X,Y). ]
Interpretation
- If and are independent, .
- If determines , mutual information is high.
Why this matters
Mutual information is a core tool for dependency detection and feature selection.