Information Security and Cryptography Research Group

About the Mutual (Conditional) Information

Renato Renner and Ueli Maurer

2002, Proceedings version (ISIT 2002): [RenMau02b].

In general, the mutual information between two random variables X and Y, I(X;Y), might be larger or smaller than their mutual information conditioned on some additional information Z, I(X;Y|Z). Such additional information Z can be seen as output of a channel C taking as input X and Y. It is thus a natural question, with applications in fields such as information theoretic cryptography, whether conditioning on the output Z of a fixed channel C can potentially increase the mutual information between the inputs X and Y.

In this paper, we give a necessary, sufficient, and easily verifiable criterion for the channel C, i.e., the conditional probability distribution PZ|XY, such that I(X;Y)I(X;Y|Z) for every distribution of the random variables X and Y. Furthermore, the result is generalized to channels with n inputs (for nN), that is, to conditional probability distributions of the form PZ|X1Xn.

BibTeX Citation

@unpublished{RenMau02a,
    author       = {Renato Renner and Ueli Maurer},
    title        = {About the Mutual (Conditional) Information},
    year         = {2002},
    note         = {Proceedings version (ISIT 2002): \cite{RenMau02b}},
}

Files and Links