Chapter 21: Problem 11
Mutual information by another name In the chapter, we introduced the concept of mutual information as the average decrease in the missing information associated with one variable when the value of another variable in known. In terms of probability distributions, this can be written mathematically as \\[I=\sum_{y} p(y)\left[-\sum_{x} p(x) \log _{2} p(x)+\sum_{x} p(x | y) \log _{2} p(x | y)\right]\\] where the expression in square brackets is the difference in missing information, \(S_{x}-S_{x} y,\) associated with probability of \(x, p(x),\) and with probabilify of \(x\) conditioned on \(y, p(x | y)\) Using the relation between the conditional probability \(p(x | y)\) and the joint probability \(p(x, y)\) \\[p(x | y)=\frac{p(x, y)}{p(y)}\\] show that the formula for mutual information given in Equation 21.77 can be used to derive the formula used in the chapter (Equation 21.17 ), namely \\[I=\sum_{x, y} p(x, y) \log _{2}\left[\frac{p(x, y)}{p(x) p(y)}\right]\\].
Short Answer
Step by step solution
Key Concepts
These are the key concepts you need to understand to accurately answer the question.