Sunday, May 3, 2009

Wrapup

Thanks again to all the hardy souls who hung in there for the semester!

Please remember to fill in the course evaluation at my.cmu.edu, under Academics.

Final grades are more or less tabulated; please email us if you want to learn them early.


best,
Ryan

Friday, May 1, 2009

Parameterized Complexity and ETH

We look into Parameterized Complexity theory (introduced by Downey and Fellows via many papers) and its connections to classical complexity theory specifically via the Exponential Time Hypothesis (of Impaggliazzo and Paturi 2001) in this post. Parameterized complexity provides a framework for a more refined analysis of hard problems. Heuristics, parallel algorithms, approximation schemes, randomized algorithms are some of the approaches used to counter such problems. But these approaches suffer from many defects ranging from no hard bounds on the quality of the solution (heuristics), to not being applicable to really large instances (parallel algorithms) to impractical solutions (like some PTAS's etc.).

Introduction:

Classical complexity classifies problems using the concept of some resource (time or space). This leads to a good theory but it also ignores any structural information in the input which makes problems appear harder then they really are. There is a wide variation in the worst-case complexities of known exact algorithms for the NP-complete problems. For e.g., there are pretty good 3SAT solvers present right now which scale to a large number of variables.

Parameterized complexity tries to explain these differences by assuming that there is some part of the problem (the `parameter') will be small and allow us to develop efficient poly-time algorithms. The classic example for this is to consider a database query - it has two parts the database and the query of which query is usually much smaller than the database. The query size k would be the natural paramter for a parameterized complexity analysis by admitting algorithms whose non-polynomial behavior is restricted by the parameter: if k is small and n is large, O(2k.n) is better than O(nk).

The main contribution of the theory is establishing the intractability of certain problems by classifying problems into complexity classes by reductions which are inherently 2-dimensional depending on the problem size as well the parameter. A problem can have different parameterizations also, each leading to different results.


Complexity Classes - FPT and the W[t] hierarchy:

A parameterized language L is a subset of Σ*xΣ* where Σ is a finite alphabet. Let (x,k)L, then we call x the main part and k the parameter.

a) Tractability - Fixed-parameter tractable (FPT): LFPT if it can be decided in time at most f(k).xO(1) for an arbitary function f. For a fixed k it is in P, moreover for every k it is in the same polynomial class via the same machine. As an example, consider the p-SAT problem where given the formula φ and the parameter k= number of variables in φ decide whether φ is satisfiable. This is clearly in FPT as the obvious brute force approach for formula of size n with k variables will take time O(2k.n). There are many real world problems which are in FPT like parameterized Vertex Cover which uses the kernelization technique to get a O(k.n+1.286k) algorithm (due to Chen, Kanj and Jia, 2001). Note that the classical Vertex Cover is NP complete - yet this result shows that it
is not as `hard' as say Independent Set.

b) Parametric Intractability: Analogously to classical complexity theory, Downey and Fellows developed a completeness program for intractable parameterized problems. In order to compare the hardness of parameterized problems, we need a 2-dimensional reduction (called `parameterized-reduction'). Roughly, a language L is parameterically reducible to Lʹ if there is an FPT algorithm that transforms (x,k) to (xʹ,kʹ) so that (x,k)L iff (xʹ,kʹ)Lʹ and kʹ=g(k) where g is an unrestricted function. Note that Karp reductions are rarely parameterized reductions (e.g. Clique to Independent Set is one of the exceptions).


1) W[1]: The lowest class of parameterized intractability can be defined as the set of languages that are reducible to the Short-Turing Machine Acceptance problem (also called the k-step halting problem) where we want to determine for an input consisting of a nondeterministic Turing machine M and a string x, whether M has a computation path accepting x in at most k steps. In some sense, this is the parameterized analogue to the Turing machine acceptance problem: the basic generic NP-complete problem in classical complexity theory. Canonical problems include Independent Set (does G have an independent of size k), Weighted 3SAT (does φ have a satisfying assignment of weight k) etc. We will also give an alternative definition of W[1] afterwards which is actually used for the basic results.


2) The W[t] hierarchy: Interestingly, while general k-SAT and 3SAT are equivalent for NP-hardness in classical complexity, there is no known parameterized reduction computing general satisfiability from 3-CNF formulae. This leads to a realization that the logical depth of a formula affects its parameterized complexity. Intuitively it is related to the number of alternations between unbounded fan-in AND and OR gates. Hence, we can base the hierarchy on the complexity of circuits required to check a solution.

The W[t] is based on Circuit-SAT problem (parameterized by the weight of input like for Weighted 3SAT) for family of circuits Fh,t having:
1. Small gates: with constant fan-in
2. Large gates: with unbounded fan-in

and of depth h (max. number of gates in the path) and weft (max. number of large gates in the path) at most t. Clearly, we have the containments:

FPTW[1]W[2]

It is not known whether the containments are strict or not. Note that W[1] can now be defined as the class that can be reduced to an parameterized Circuit SAT on the family of constant depth weft 1 circuits. Downey and Fellows (1995) proved that parameterized Independent Set is W[1]-complete. The hardness proof is very intricate and we omit it here. Note that using this definition it is easy to see why Independent Set is in W[1] (sketch):
1. Each vertex in G corresponds to one input gate in the circuit
2. For every edge (v,u)G build a small OR gate: (1-v)(1-u))
3. Output from small gates are given as input to a single large AND gate.

Interestingly unlike NP-hardness results, W[1]-hardness for the k-Halting Problem uses Independent Set completeness as intermediate results. There have been many papers demonstrating naturally occuring W[t]-complete problems like Dominating Set for W[2] etc.)- for an extensive list of already classified problems see the Downey and Fellows monograph.

Exponential Time Hypothesis (ETH)

ETH was first studied by Impagliazzo, Paturi and Zane 2001 and it states that 3SAT DTIME(2o(n)), where n is the number of variables. They also proved that the hypothesis is robust to analogous hypotheses for other NP-complete problems like Independent Set etc. In fact, they also showed that the ETH is independent of size measure: 3SAT is solvable in time 2o(n) iff it is solvable in time 2o(m) for input size m. Note that this suggests that weighted 3SAT should also be intractable for `any parameter'.

Connections with Classical complexity:

The above paragraph suggests that ETH and parameterized complexity might be related. In fact they are and Chen and Grohe recently proved that ETH and paramterized complexity theory are isomorphic by an explicit reduction preserving isomorphism. We don't show that here, instead we prove a simpler result that yet provides a strong link to ETH (proof is a version from Downey, Castro et. al 2003 and uses parameterized miniaturization and is different from the original proof by Abhramson, Downey and Fellows):

Theorem: If FPT = W[1], then ETH fails i.e. 3SAT DTIME(2o(n)).

Proof: We use the equivalent definition of ETH based on simple Circuit SAT. The idea is to capture the ETH perfectly in a parameterized complexity class. The starting point is the Mini-Circuit SAT problem which is a parameterized miniaturization of simple Circuit SAT:
Input: Positive integers k and n, and a Boolean circuit of total size at most klogn.
Decision: Does there exist any x for which C(x)=1.

Note that the parameter here is k. Also, trying all possible inputs gives a brute force O(nk) algorithm. Next we give a cruicial lemma due to Cai and Juedes 2001. It essentially fully characterizes the ETH with a complexity class in parameterized theory.

Lemma 1: Mini-Circuit SAT is in FPT iff ETH fails.
Proof: One direction follows from the brute force algorithm and noting that 2o(klogn) is a FPT function.

Now suppose we are given a boolean circuit C of size N and that Mini-Cirsuit SAT is solvable in FPT time f(k)nc. Set k=f-1(N) and n=2N/k. In general k=f-1(N) will be some slowly growing function of N; so N/k=o(N) and also cN/k=o(N). Hence using the FPT algorithm for Mini-Circuit SAT we have a running time for Circuit SAT as: f(f-1(N))(2N/k)c=N2cN/k=2cN/k+logN=2o(N).
Thus ETH fails. Proved.

Now lets define the complexity class MINI[1] to be the set of languages that are FPT reducible to Mini-Circuit SAT. It turns out that many klogn miniatures of familiar NP-complete problems are MINI[1] complete (Downey, Fellows et al. 2002). It is easy to see this because essentially all the usual NP-complete reductions of Circuit SAT to these problems work as FPT reductions because they were also linear size reductions.

We concentrate on the Mini-Independent Set problem. Surprisingly, it can be reduced to the usual parameterized Independent Set problem.

Lemma 2: Independent Set parameterized by the size of independent set is MINI[1]-hard.
Proof: We give a Turing reduction. Let graph G=(V,E) be the miniature, for which we wish to determine whether G has an independent set of size r with Vklogn. We can think of verices of G as organized in k blocks V1,V2,,Vk each of size logn. So for each possible way of writing r as a sum of k terms r=r1+r2+r3++rk with each rilogn, we have a turing reduction branch which represnts a commitment to choose ri vertices from the corresponding block Vi to be in the independent set. The total number of branches are (logn)k and again it is a FPT-function.

For each branch, we produce a graph Gʹ that has an independent set of size k iff the miniature G has an independent set of size r distributed as indicated by the commitment made on that branch. The graph Gʹ consists of k cliques with some cross edges. The ith clique consists of vertices in correspondence with the subsets of Vi of size ri. An edge connects a vertex x in the ith clique and a vertex y in the jth clique iff there is a vertex u in the subset SxVi represented by x and a vertex v in the subset SyVj represented by y, such that (u,v)E.

From Lemma 2 we can conclude that FPTMINI[1]W[1]. We just need to observe now that if FPT = W[1] it implies M[1]FPT. By Lemma 1, ETH fails.
Proved.

Discussion:

The above result hints that FPT vs W[1] problem is like the P vs NP problem of classical complexity theory. In fact it was also proved by Downey and Fellows that FPT W[1] implies P NP. But it is not known that if FPT W[1] implies anything about the rest of W[t] hierarchy. Importantly, practical intracbility of problems in NP, which are unlikely to be complete for NP, can be shown using W[t] hardness results. Parameterized complexity has deep connections to other complexity results as well. Just recently (a week back) Galesi and Lauria showed a connection with Proof Complexity based on randomized algorithms for W[t]-hard problems. The field is very active and there are many papers and surveys being published in it.

Lower bounds for locally decodable codes

In this blog, we will look at upper bounds on the rates of locally decodable codes and their relation to a problem called Private Information Retrieval. Some of the proofs presented here are unique in that they relate the complexity of classical algorithms to quantum algorithms and prove lower bounds on quantum algorithms!

Locally decodable codes are codes that can probabilistically recover a bits from corrupted codewords by querying a small number of bits, Hadamard code being a straightforward example. While an ideal code would have have rate, be resilient to large number of errors and would be locally decodable, it as been shown that it is not possible to do extremely well on all the criteria. Specifically, locally decodable codes imply that the code rate is o(1).

The first paper in this direction ([Katz, Trevisan STOC'00]) showed that any locally decodable code with constant number of query bits has codewords of superlinear length. The main ideas behind these bounds is that smooth codes (codes which are queried uniformly for local decoding) are not much worse than non-smooth ones and that such smooth codes need super-linear encoding lengths.

Def 1: Suppose that a code maps C:{0,1}nΣm. For fixed δ,ε and integer q, C is (q,δ,ε)-locally decodable code if there exists a probabilistic algorithm that recover an arbitrary bit xi of the input x with probability at least 1/2+ε from a corrupted codeword y which is with in a distance d(y,C(x))<δm of C(x) after querying no more than q indices of y.

Intuitively, smooth code words are those for which there exist probabilistic decoding
algorithms that query the codeword (roughly) uniformly, i.e., are not heavily biased
towards querying few indices of codeword. More formally:

Def 2: A codeword C:{0,1}nΣm is (q,c,ε)-smooth
for fixed c,ε and integer q if there exists a local decoding algorithm A
that can recover an aribitrary bit of input x from codeword C(x) with probability at least 1/2+ε using no more than q indices of y with the added condition that \prob[A(x,i)reads indexj]c/m. This definition does not require A to recover bits from corrupted codewords.

Lemma 1: If C is a (q,δ,ε)-locally decodable code, then it is also a (q,q/δ,ε)-smooth.
Proof sketch: Take the decoder for C and identify all the locations that are queried with probability greater than q/δm while trying to decode m. Construct a new decoder that just assumes the value 0 for queries to all such locations. Since there no more than δm of these, the new decodder can still recover from errors. This new decoder has the smoothness properties we want.

Lets start with case q=1, and see why it is not possible to construct locally decodable codes that encode inputs of length greater than a constant. Suppose that the algorithm is trying to decode the i-th bit xi by reading just a single index of the (corrupt) codeword y. Then, j[m]Prx[A(C(x),i)=xiA(,i)readsj]Pr[A(,i)readsj]
=Prx[A(C(x),i)=xi]1/2+ε,

beacuse of which there must exist a good index j1 such that Prx[A(C(x),i)=xiA(,i)readsj1]1/2+ε. Suppose that the code got randomly corrupt at index j1. Then yj1 has no correlation with xi and the algorithm A can not infer any information from index j1. But since A can recover from δm errors, there must exist another index j2 which also \textit{good} in the sense that it can be queried to get information about xi. By extending the above reasoning, we can see that we can also corrupt yj2 randomly along with yj1 and can expect to find another index j3 with reasonable correlation to xi. We can extend the same argument δm times to reason that there are at least δm indices j[m] such that A can compute xi from C(x)j with probability at least 1/2+ε. This is true of all inputs x, which by the pigeon hole principle implies that there is at least one index jʹ[m] such that at least δn of xis can extracted with 1/2+ε probability by querying yjʹ. Now this means that the code should put an enormous amount of information about the input in to one index of the output. Consider the following quantitive
lemma about limit of information recovery from a function:

Lemma 2: If C:{0,1}nR, and there exists a A such that Pr[A(C(x),i)=xi]1/2+ε,
then logR(1-H(1/2+ε))n, where H is the binary entropy function.
Idea: if a decoder has an any hope of recovering bits with reasonable probability,
the compression should not e too high.

Setting R=Σ, we have nlogΣδ(1-H(1/2+ε)) (Therorem 3).

Now, consider the case q>1. A generalization of this same argument can be used
to show superlinear bounds on the length of the codewords as follows:

Suppose that A is a (q,c,ε)-smooth decoder for code C.
Similar to the q=1 case, call a set S (S[m],Sq) ε-good for i if Pr[A(C(x),i)=xiAreadss]1/2+ε. Define hypergraph Hi with vertices labelled [m] and egde set Ei defined by the ε/2-good sets for i. A matching of a hypergraph is an edge set with no common vertex and a vertex cover is a vertex set such that every edge contains at least one vertex from the vertex set.

Lemma 4: Hi has a matching Mi of size at least εm/cq.
Proof:
1/2+ε
Prx[A(C(x),i)=xiA(,i)readsEi]Pr[A(,i)readsEi]
+Prx[A(C(x),i)=xiA(,i)readsEc_]PrA(,i)readsEc_]
<Pr[A(,i)readsEi]+(1/2+ε/2)(1-Pr[A(,i)readsEi]),
which implies that Pr[A(,i)reads fromEi]>ε. If Pe denotes the probability that A(,i) reads eEi, then we have ε<eEiPe. Also for every j[m], eEijec/m
by smoothness condition. Now, if V is vertex cover of Hi, eV for all eEi. Putting this together with the earlier fact, we have ε<sEieVPejVeEijePeVc/m which implies V>εm/c, and therefore, Hi has a matching Mi of size at least εm/cq.

Say that a set S hits matching Mi if there is some set sS such that sMi. The following lemma gives a bound on the number of vertices to be selected from Hi so that selected set hits a constant fraction of Mis.

Lemma 5: If H is a hypergraph on m vertices containing hyperedges of at most than q vertices. Suppose H has a matching of size γm(γ<1/q). There exists t=Θ(γ-1/qm(q-1)/q) so that for a randomly chosen (from H) set of t elements, such that the probability of this set hitting an arbitrary matching is 3/4.

Theorem 6: If C:{0,1}nΣm is a (q,c,ε)-locally decodable codes, then:
m(εδ/q2)1/(q-1)(3n(1-H(1/2+ε))4logΣ)(qq-1).
Proof: Lemma 1 shows that C is (q,q/δ,ε)-smooth. Lemma 4 shows that for every i, there exists a set Mi consisting of disjoint set of size at most q such, each mMi is ε/2-good for i and Miεδm/q2. Lemma 5 says that there exists a set of t=Θ((εδ/q2)-1/qm(q-1)/q)) indices from [n] so that the values at these locations have enough information to help decode at least 3/4 of the inputs bits with advantage at least ε/2. Applying Lemma 2 which gives us a lower bound on t proves the theorem.

While this is some start, this still leaves a large gap between existing locally decodable codes (which are exponential long) and the lower bound. Goldreich, Karloff, Schulman and Trevisan improve the lower bound for the specific case of linear codes and decoding algorithms that query only 2 bits. They show that in such a case, m=2Ω(n). Using new techniques (reduction to quantum queries), Kerenidis and Wolf (arXiv: quant-ph/0208062v2) show that any 2-query LDC (not necessarily linear) is exponentially long. Their work is as follows.


Quantum queries
---------------

In an earlier post, Matt talked about the fact that QM systems exist in a linear superposition of several states. The choice of basis vectors used to describe can of course be chosen according to convenience (the choice is usually the eigenstates of the measurement). A superposition of states does not mean a statistical mix of different states, rather it means that the state of the particle itself is a complex vector. Such a state is called a pure state (can be expressed as φ=iαii). The density matrix of a pure state (or qubit) is the outer product M=φφ, where φ=φ=φT* is the complex conjugate transpose of φ. Note that an arbitrary matrix M need not in general be the density matrix of a pure state. Matrices of the form ρ=ipiφiφi are called mixed states. As against a pure system, such mixed states are statistical ensembles of different pure states. Obviously, such a mixed state does not represent any pure state. However, the set of pure states of which ρ is an ensemble is not uniquely determined by ρ alone. ρ represents both {(pi,vi}i and {(pʹi,vʹi as long there is a unitary matrix U such that vi=jpj/pʹiUijvj. Another context where mixed systems are useful for us is to describe a subsystem of an entangled state. For example, in the entangled state (00+11)/2, the second qubit alone is not a pure system. It can be viewed as the mixed state (00+11)/2.

Measurement of a qubit can be thought of as projecting the qubit in to a subspace. Based on the set of subspaces we are trying to project our qubit in to, the qubit has different probability of collapsing in to these subspaces. We can generalise this notion to arbitrary positive operators (not just orthonormal projectors) and mixed states -- such a measurement system is called positive operator valued measurement (POVM). A POVM is a set of positive operators Ei=MiMi such that i=1kEi=I. When such a measurement is made on state ρ, the outcome is any one of the states MiρMi/Tr(MiρMi), the probabilities of the respective outcomes being pi=Tr(MiρMi).

In what follows we usually deal with k of the form 2m and systems of m qubits.
If B={\ketψi} is an orthonormal basis for the system, measuring in B-basis means using the POVM: Ei=\ketψi\braψi. Then, the outcome of measuring a pure state \ketφ is simply \braketφψ2 as we expect.

Now, we are ready to define a quantum query:
A query to j-th bit of a length m string y is the unitary operation (quantum mechanics mandates that all state transformations are unitary):
cj(-1)cyjcj.
Of course, the fact that we can apply this query transformation to superposition states to (indirectly) read off several values of the function is what makes quantum queries powerful (for that matter, this ability to manipulate superposition states is what gives extra power to quantum algorithms). The following lemma illustrates this power right away:

Lemma 7: Let f:{0,1}2{0,1} and suppose that we have a=a1a2{0,1}2 whose bits are to be queried in order to compute f(a). There exists a quantum algorithm that uses just one query, and outputs f(a) with probability exactly 11/14, and outputs 1-f(a) otherwise.
Proof: Note that a classical algorithm has to query both bits before it can compute f(a) with any accuracy. The quantum algorithm is as follows: query (01
+ |11\rangle + |12\rangle)/\sqrt{3}.Theoutputofthequerywouldbe:
φa=(01+(-1)a111+(-1)-a212)/3, the mutual phase shifts now holding information about a1,a2. To extract this information, we measure this state (i.e., cause it to collapse to basis) with the basis {ψb}(b{0,1}2), where ψb=(01+(-1)b101+(-1)b210+(-1)b1+b211)/2.

φa gives outcome a with probability φaψa2=3/4, the other three outcomes being equally probable (1/12). Suppose that the measurement outcome is m. The following procedure does exactly what we want:

  • f is a constant. Output the constant with probability 11/14.

  • f(1)-1=1. If f(m)=1, then output 1. If f(m)=0, then output 0 with probability 6/7 and 1 with probability 1/7. Now, if f(a)=1, then probability of the algorithm outputting 1 is (3/4).1+(1/4)(1/7)=11/14. If f(a)=0, then probability of algorithm outputting 0 is (11/12)(6/7)=11/14.

  • f(1)-1=2. Output f(m) with probability 13/14 and 1-f(m) with probability 1/14.

  • f(1)-1=3. Similar to case f(1)-1=1.



Definition: A (q,δ,ε)-LQDC (locally quantum-decoable code) is the same as a LDC except that we replace the probabilistic decoder with a quantum decoder and queries are quantum-queries that probe superpositions. Now we can use the earlier lemma to show that:

Theorem 8: A (2,δ,ε)-LDC is a (1,δ,4ε/7)-LQDC.
Proof: The idea is to replace the two classical queries with one quantum query (allowing for a bit of inaccuracy) based algorithm. The quantum decoder will choose two target bits the same way the classical decoder works, and instead of reading both bits and applying a function f, the LQDC does a quantum query on this two bits and outputs the value of f on this bits (with 11/14 accuracy). This suffices to get a 4ε/7 predictor.

Now we want to show lower bounds on 1-query LQDCs. The starting point as in the
case of 1-query LDCs is a theorem that gives a relation between the space
used to store information and fidelity of information reconstruction:

Theeorem 9: If f maps n bit strings to m-qubit states (mixed) states with
recovery probability at least p, then m(1-H(p))n. This leads to:

Theeorem 10: If C:{0,1}n{0,1}m is a (1,δ,ε)-LQDC, then m2cn-1, for c=1-H(1/2+δε/4).


Wehner and Wolf (arXiv: quant-ph/0403140v2) consider the case where the LDC encode
over larger alphabets and when the decoder uses only a part of the information from each query. Briet and Wolf (arXiv: quant-ph/0806.2101v1) further study the relation between LQDC and LDC and conclude that their powers are roughly the same for small constant number of queries.

Despite all this work, the best lower bounds on the overheads for >3 query LDC is only polynomial while known >3-query LDCs are eponentially long. This gap leaves a big open question.

LDCs and Private Information Retrieval (PIR)
--------------------------------------
The relation between LDCs and PIRs is close and worth mentioning since PIR systems yield some of the best LDCs. PIR is the problem of distribution a database among different servers so that a client can query a particular bit (by sending randomized queries to different servers) so that individual servers have almost no clue about the particular query bit that the client is accessing (information theoretic security, not computational). Here is the definition of a 2-server PIR system.

Def [GKST CCC'02]: A one-round, database size n, (1-δ) secure, 2-server PIR with recovery probability p, query size q, answer size a is the set of algorithms (Q,(S1,S2),D), where:
- Q on input i uses a randomness (r) to generate two q bit long queries (q1,q2)=Q(i,r). q1 and q2 are sent off to the two servers which use algorithms S1 and S2 to respond with length a-answers to these queries. The decoder algorithm D uses i,r and the answers of the servers to decode xi. Given that the random string for input to Q is selected from uniform distribution, the decoder D should succeed with probability at least p. The secrecy condition is that the distribution of queries send to servers should be different for different target bits by more than a distance of δ.

The same authors show that if there is a 1-round, (1-δ)-secure 2-server PIR of database size n, query size t, answer size a with recovery probability at least 1/2+ε, then there exists a (2,3,ε-δ)-smooth code C:{0,1}n({0,1}a)m, where m6.2t. This reduction can be used to show that a query sizes Θ(nc)(c>0) are needed for PIR with constant answer sizes.

Kolmogorov Complexity and derandomization

Motivation
------------

The sets Rμ of random strings with high Kolmogorov complexity and bounded resources are good examples of sets with a lot of information content that is difficult to access. Many of these sets have been studied [BM97,Ko91,KC00] as possible examples of intractable sets that are not complete for any of the standard complexity classes. Based on [Al05], we now show that these sets can, in fact, be exploited by efficient reductions.

The completeness results are obtained via derandomization techniques, mostly relativizing hardness vs randomness tradeoffs in the contrapositive: If there exists a problem in complexity class C that is hard when given oracle access to A, then there exists a pseudorandom generator secure against A that is computable within C. We argue that no pseudorandom generator computable in C can be secure against Rμ, and thus every problem in C is easy given oracle access to Rμ. In other words, C reduces to Rμ.

Definitions
-----------

The computational model that we use is the multi-tape Turing machine with random-access to its input tape (the ideas work in any general model). Wlg, we pick a single universal machine U.
Kolmogorov complexity CA(x) is usually defined as the length of the shortest description d{0,1}* such that UA(d)=x, where U is a universal Turing machine with oracle access to A. The definition we use is slightly different: instead of requiring UA(d) to produce x, it should recognize the correct value of xi for every i.
We now define several resource-bounded Kolmogorov complexity measures:


Definition 1 [Kt,KT,KS]
KtA(x)=min{d+logt:UA(d,i,b) accepts in t steps iff xi=b}
KTA(x)=min{d+t:UA(d,i,b) accepts in t steps iff xi=b}

The only difference is that in KT the time bound is given exponentially more weight. Considering space-bounded notions, too, will yield complete problems for space-bounded
classes:

KSA(x)=min{d+s:UA(d,i,b) accepts in s space iff xi=b}

Both Kt and KS can be approximated in terms of KTA, for appropriate oracle choices:

Theorem 2:
There exist a complete set A for E, a complete set B for DSPACE(n), and a constant c s.t.
Kt(x)/cKTA(x)
KS(x)/cKTB(x)

Proof sketch: Let AE and let x be given, s.t. KTA(x)=m. Thus, there is a description dx of length m, s.t. UA(dx,i,b) accepts iff xi=b in time at most m. During computation, U asks queries of length at most m. Since AE, each query can be answered in time 2O(m).
Let M denote the algorithm simulating the computation of UA(dx,i,b) for every i by directly computing the answers of A, then the description dʹx=M,dx is sufficient for U to compute U(dʹx,i,b) in time 2O(m). As dʹx=m+O(1), we can conclude that Kt(x)m+O(1)+log(2O(m))=O(m).
The proof for KS is similar.

We focus on sets containing strings of high complexity, for various measures.

Definition 3:
For any Kolmogorov complexity measure μ, define
Rμ={x:μ(x)x/2}.

The bound of x/2 is arbitrary. Essentially, all we need is that the set Rμ has polynomial density (it contains at least 2n/nk strings of each length n, for some k).
The following propositions are straightforward.

Proposition 4:
RKtE, RKSDSPACE(n), and RKTcoNP.

Proposition 5:
The sets RKt, RKS all have polynomial density.


Nonuniform Hardness Results
-------------------------------

We show that strings of high Kolmogorov complexity are very useful as oracles. We will argue that an appropriate set Rμ of Kolmogorov random strings can be used to distinguish the output of a pseudorandom generator Gf from truly random strings. This in turn will enable us to efficiently reduce f to Rμ.

Recall that A is PSPACE-robust if PSPACEA=PA (machines are allowed to ask oracle
queries of only polynomial size). The complete sets for many large complexity classes (PSPACE, EXP, EXPSPACE) have this property, as well as the complete sets (under linear-time
reductions) for classes like DSPACE(n) and E.

We will build a pseudorandom generator based on the Nisan-Wigderson paradigm [NW94]. [BFNW93] construct, for any ε>0, a variant GBFNW_:{0,1}nε{0,1}n such that for any x of size nε, the function is computable in space O(nε) given access to the Boolean function f on inputs of size at most nε. If f is PSPACE-robust, there is a constant c independent of ε, such that each bit is computable in time nεc with oracle access to f. The following hardness versus randomness tradeoff holds.

Theorem 6 [BFNW93]
Let f be a Boolean function, ε>0, and GBFNW_ be the pseudorandom generator described above. Let T be a set and p(n) a polynomial. If
PrUn[rT]-PxUnε[GBFNW_(x)T]1/p(n)
for all large n, then there exists a polynomial size oracle circuit family {Cn} with oracle T that computes f and queries T non-adaptively.

We use the notation AP/poly_tt to denote that there exists a truth-table (i.e., nonadaptive) reduction from A to B that is computable by a family of polynomial-size circuits.
(A truth-table reduction is a reduction from one set of natural numbers to another. Truth-table reductions are more powerful than Turing reductions in a mathematical sense, since they provide finer equivalent class, but they are a weaker tool.)

Theorem 7:
Let A be any PSPACE-robust set. Let L be a set of polynomial density s.t. for every xL, KTA(x)>xγ for some constant γ>0. Then A is reducible to L via P/poly_tt reductions.

Proof Sketch: Let f be the characteristic function of A. Consider GBFNW_,
choose ε as follows. We know that every bit of GBFNW_ is computable in time nεc for some constant c independent of ε, given access to A. Set ε=γ/2c (wlg c>1).

Any string in the range of GBFNW_ has small KTA complexity: Let y=GBFNW_(x), for some x. On x, every bit of GBFNW_ is computable in time nγ/2 with access to oracle A. Hence, KTA(y)x+O(nγ/2logn)+O(1)nγ. Thus, L distinguishes the output of GBFNW_ from random, and by Theorem 6, f is P/poly_tt reducible to L.

By Theorem 2 and Proposition 5, we can apply Theorem 7 to
the set A from Theorem 2, RKt, and the set B from Theorem 2, RKS.
(A and B are clearly PSPACE-robust, since they are complete for EXP and PSPACE, respectively.) Combining this with Proposition 4, we get:

Corollary 8:
RKt is complete for EXP under P/poly_tt reductions. RKS is complete for PSPACE under P/poly_tt reductions.


Those results also obtain natural examples that witness the difference in power of various reducibilities, as some of the sets which we show are complete under truth-table reductions are provably not complete under polynomial-time many-one reductions.