Data processing inequality
The data processing inequality is an information theoretic concept that states that the information content of a signal cannot be increased via a local physical operation. This can be expressed concisely as 'post-processing cannot increase information'.[1]
Statement
[edit ]Let three random variables form the Markov chain {\displaystyle X\rightarrow Y\rightarrow Z}, implying that the conditional distribution of {\displaystyle Z} depends only on {\displaystyle Y} and is conditionally independent of {\displaystyle X}. Specifically, we have such a Markov chain if the joint probability mass function can be written as
- {\displaystyle p(x,y,z)=p(x)p(y|x)p(z|y)=p(y)p(x|y)p(z|y)}
In this setting, no processing of {\displaystyle Y}, deterministic or random, can increase the information that {\displaystyle Y} contains about {\displaystyle X}. Using the mutual information, this can be written as :
- {\displaystyle I(X;Y)\geqslant I(X;Z),}
with the equality {\displaystyle I(X;Y)=I(X;Z)} if and only if {\displaystyle I(X;Y\mid Z)=0}. That is, {\displaystyle Z} and {\displaystyle Y} contain the same information about {\displaystyle X}, and {\displaystyle X\rightarrow Z\rightarrow Y} also forms a Markov chain.[2]
Proof
[edit ]One can apply the chain rule for mutual information to obtain two different decompositions of {\displaystyle I(X;Y,Z)}:
- {\displaystyle I(X;Z)+I(X;Y\mid Z)=I(X;Y,Z)=I(X;Y)+I(X;Z\mid Y)}
By the relationship {\displaystyle X\rightarrow Y\rightarrow Z}, we know that {\displaystyle X} and {\displaystyle Z} are conditionally independent, given {\displaystyle Y}, which means the conditional mutual information, {\displaystyle I(X;Z\mid Y)=0}. The data processing inequality then follows from the non-negativity of {\displaystyle I(X;Y\mid Z)\geq 0}.
See also
[edit ]References
[edit ]- ^ Beaudry, Normand (2012), "An intuitive proof of the data processing inequality", Quantum Information & Computation, 12 (5–6): 432–441, arXiv:1107.0740 , Bibcode:2011arXiv1107.0740B, doi:10.26421/QIC12.5-6-4, S2CID 9531510
- ^ Cover; Thomas (2012). Elements of information theory. John Wiley & Sons.
External links
[edit ]
This computer science article is a stub. You can help Wikipedia by expanding it.