7. Query Hierarchy Theorems
The last two lectures showed that even the most basic questions “Does randomness help reduce query complexity?” and “If so, by how much?” lead to surprisingly rich lines of research. Let us now continue along this line by asking another deceptively simple question:
What values can
take?
Every non-constant function
But before we dive into these questions, let’s first take a detour and examine the randomized query complexity of two special functions.
Parity and Gap Majority
The parity (or XOR) function on
With an adversary argument, we can show that
And even in the small bias regime where we consider
Proposition. For every bias function
, In particular,
As we will see, the parity function is useful in establishing a number of complexity theoretic results.
Another useful function for query complexity in particular is the gap majority function on
In other words, the gap majority function is the usual Majority function restricted to the domain
The restriction on the domain makes the function a little bit easier to compute than the standard majority function because it guarantees a gap of size at least
In the small bias regime, however, the gap does make a meaningful difference.
Proposition. For every bias function
, In particular,
Note that this bound is must stronger in terms of
A Hierarchy Gap Theorem
Using the parity function on a subset of the
But the observation above is not entirely satisfactory because the functions used in the construction only depend on a subset of the
Definition. A total function
is truly -variate if for every input , there exists an input such that .
Can truly
Proposition. Every truly
-variate Boolean function has deterministic query complexity . This bound is asymptotically tight: there exist such functions with .
Proof. The lower bound on
follows immediately from the observation that a binary tree of height contains at most nodes and that if is truly -variate, any decision tree that computes it must query all of the indices in at least one of the nodes. Many different functions can be constructed to show that the
bound is asymptotically tight. One particularly useful construction because of its versatility is the function. This function on bits interprets an input as a tuple , and and is defined by In other words, it returns the value of the coordinate of indexed by the address . As we can verify directly, the function is truly variate and it can be computed by a deterministic decision tree of depth .
What about for randomized query complexity?
The simple argument that we used for deterministic decision trees no longer holds when we consider randomized decision tree, because the trees in the support of the distribution that corresponds to a randomized decision tree do not have to all have nodes that query all of the indices of the input.
Yet, the same lower bound on the query complexity of truly
Theorem. Every truly
-variate Boolean function has randomized query complexity .
Proof. The lower bound follows from a matching lower bound on the maximum sensitivity of truly
-variate functions. This result that for every non-constant truly -variate function is due to Simon (1982).
The bound on the minimum randomized query complexity of truly
And the set of all truly
Theorem. Every non-constant symmetric function
has randomized query complexity .
And in other cases, the additional restrictions do not affect the minimum query complexity of functions:
Theorem. Every monotone truly
-variate function has randomized query complexity . This bound is tight for some monotone truly -variate functions.
Proof. The lower bound is the one that holds for all truly
-variate functions. For the matching upper bound, consider the monotone variant of the index function. This function is defined on
bits for some even . We again consider the first bits to represent an address, but in this case only the strings with Hamming weight exactly represent valid addresses. The function is defined as the standard function on valid addresses. On all other addresses, the function evaluates to 1 if the address has more 1s than 0s and to 0 otherwise. We can again verify directly that the function is monotone and truly -variate, and that it can be computed by a deterministic decision tree of depth .
What about other classes of functions?
One particularly interesting special case is the set of all transitive-symmetric Boolean functions,
where
Randomized Evasiveness Conjecture. Every non-constant monotone transitive-symmetric function
has randomized query complexity .
A Hierarchy Theorem for Partial Functions
If we turn our attention to partial Boolean functions, we can again easily obtain a hierarchy theorem for randomized query complexity.
This time, we can start with the Gap Majority function and vary the size of the gap in Hamming weights in the definition of the domain
But again this answer is not completely satisfying. Or, perhaps a better way to put it is that this observation raises an immediate follow-up question: is it just the size of the gap (appropriately defined/generalized) that determines the randomized query complexity of all partial functions? We can make this question precise by introducing the notion of margin of partial functions.
Definition. The margin of the partial function
for some is the minimum Hamming distance between inputs in that take different values under :
Let us say that the partial function
The answer is no: having a linear margin is not sufficient to guarantee having small randomized query complexity.
Theorem. There exists a partial function
, with and randomized query complexity .
Proof sketch. Following Goldreich, Goldwasser, and Ron (1998), one way to prove the theorem is via error-correcting codes.
We know that there is a set
of points in the hypercube such that every pair of points in has Hamming distance between them. (Taking to be a random set of points and removing any pair that does not satisfy the condition suffices to establish this claiim; there are also more explict constructions that can be used.) Let us define
by setting its value independently at random for each point in . The minimum distance between points in guarantees that has linear margin. And we can show that for any decision tree with query complexity , the uniform distribution on points in leads to leaves that have almost the same probability of being reached by points in and in , so that no such decision tree can compute with bounded error.
Building on this result, we can obtain a full (asymptotic) hierarchy theorem for partial functions with linear margin.
Once again, we can ask about particular sets of partial functions with linear margins. Doing so leads to property testing, which considers the partial functions that arise when considering properties of combinatorial objects (such as graphs or functions) and the number of queries required to distinguish the objects that have the property from those that are far from having the same property.