12. Set Disjointness II

At the beginning of the last lecture, we introduced the claim that the randomized communication complexity measure of the Set Disjointness function is R(Disjn)=Θ(n). By the end of the lecture, we had only established the weaker lower bound of Ω(n).

The optimal lower bound was first established by Kalyanasundaram and Schnitger (1992) and shortly afterwards an alternative proof was given by Razborov (1992). But we will cover a different proof of the same theorem that provides us with a great excuse to introduce a fundamental tool in communication complexity (and other areas of theoretical computer science): information theory.

Basics of Information Theory

The starting point for information theory is the notion of entropy. When X is a random variable with distribution μ over some finite set X, the entropy of X is

H(X)=xXμ(x)log1μ(x).

The notion of entropy satisfies a number of useful properties. Perhaps the most fundamental is a direct consequence of Jensen’s inequality which gives a universal upper bound on the entropy of any random variable.

Proposition 1. For any random variable X with distribution μ over a finite set X, the entropy of X is abounded above by

H(X)log|X|.

When μ is some joint distribution on X×Y and (X,Y)μ, the conditional entropy of X relative to Y is

H(XY)=EyμY[H(XY=y)]

where we write μY to represent the marginal distribution of Y under μ. The most basic property of conditional entropy is that conditioning never increases entropy.

Proposition 2. For any jointly distributed random variables X and Y,

0H(XY)H(X).

The mutual information between jointly distributed random variables X and Y is

I(X;Y)=H(X)H(XY).

(There are many other equivalent formulations of mutual information as well.) Informally, mutual information measures “how much information” about X that we learn by observing Y.

We can also define conditional mutual information similarly.

Mutual information satisfies a number of useful properties. One of these particularly useful ones is the chaining rule.

Proposition 3. When X and Y1,,Yn are jointly distributed random variables,

I(X;Y[n])=i=1nI(X;YiY[i1]).

And another useful property is that independent random variables do not affect conditional mutual information.

Proposition 4. When X,Y,Z are jointly distributed and Y and Z are independent under this joint distribution, then

I(X;YZ)I(X;Y).

Intuition for a Set Disjointness Lower Bound

We can use the information theory results introduced above to obtain a linear lower bound on the randomized communication complexity of Set Disjointness.

The starting point of this lower bound requires us to go back to a very basic question: why do we believe that the randomized communication complexity of the set disjointness function should be linear in n in the first place? Well, recalling that our function is defined as

Disjn(x,y)=¬i[n](xiyi),

there is one very natural answer to this question: the communication complexity of disjointness should be Ω(n) because it appears to implicitly make us compute the AND function on a linear number of instances. Letting And:{0,1}{0,1}{0,1} denote the simple AND function on a pair of bits, this intuition suggests that we should maybe try to prove an inequality of the form

R(Disjn)?Ω(nR(And)).

If such an inequality can be established, then we are done because R(And)1 follows from the easy observation that any protocol for the AND function must involve some communication between Alice and Bob.

The problem with the intuition described abovbe is that inequalities like the conjectured one above do not hold in general. Indeed, this intuition implicitly assumes that the best way to compute a function which can be represented as multiple instances of a smaller function must be to compute each of those instances separately to obtain the final result. This is certainly not always the best course of action! In fact, we can see how wrong this implicit assumption can be by considering a slight variant of the disjointness function where we replace the with a . The exact same intuition we described could also be applied to the resulting function i[n](xiyi), and yet this function, in strong contrast to disjointness, can always be computed with constant communication cost.

At this point, it would be natural to assume we need to go back to the very beginning and find a different line of attack to try to prove a lower bound on disjointness. But remarkably, using information theory we don’t need to do that: we can show that our original intuition on the disjointness function was in fact generally correct if we simply relax our target inequality slightly to replace R(And) with a different information-theoretic measure of complexity of the AND function.

Let us see how we obtain such a result.

From Set Disjointness to AND

Fix a randomized protocol π for the Disjn function. We let X and Y denote the inputs to Alice and Bob drawn from some distribution that we will define shortly. And we let Π denote the transcript of π on input (X,Y). The transcript is simply a binary string that describes the bits that are communicated by Alice and Bob during the execution of π.

The distribution μ on (X,Y,Π) is obtained by running the following procedure:

Let ν be a distribution (X,Y,Σ) defined analogously except that in this case n=1 and Σ is the transcript of a protocol σ for the AND function.

We can use the information theory tools developed earlier to establish the following result.

Lemma 4. With μ and ν defined as above and letting CICν(And)=infσIν(Σ;XYD), we have

R(Disjn)nCICν(And).

Proof. Let π represent a randomized communication protocol that computes Disjn with minimal worst-case cost. Then

R(Disjn)=max|Π|H(Π)(Entropy upper bound)H(ΠD)(Conditioning does not increase entropy)H(ΠD)H(ΠXYD)(Non-negativity of entropy)=I(Π;XYD)(Mutual information definition)=i=1nI(Π;XiYiDX1Xi1Y1Yi1)(Chain Rule)=i=1nI(Π;XiYiDi).(Independence)

For any i[n], consider now the protocol σi for the AND function that proceeds as follows. On input x,y, it generates D,Z,X,Y as in the procedure for μ and then overwrites Xi=x and Yi=y. Then Alice and Bob simulate the protocol π on the resulting input (X,Y). Note that for any ji, we will have XjYj=0 by construction, and so the outcome of π determines the value of xy. Furthermore, if (x,y,Σi) is drawn from the ν distribution with the protocol σi, then we have

I(Σi;XiYiDi)=I(Π;XiYiDi)

and by definition in the lemma statement this value is bounded below by CICν(And).

Note that we are not quite done at this point. Earlier, we obtained the trivial lower bound of R(And)1 from the observation that any protocol for the AND function must involve some amount of communication. But with information complexity, this basic observation does not suffice to give a constant lower bound on the conditional information cost of the AND function. After all, it’s possible that there are infinitely many different communication protocols that each have positive information cost but that this cost tends to 0 as we consider more and more sophisticated protocols. Such a situation would of course invalidate our entire proof approach. But, as our intuition suggests, this is not the case: it’s possible to give a (positive constant) lower bound on CICν(And) and this will complete the proof of the claim that Set Disjointness has linear communication complexity.

We will see this lower bound in the next lecture.