NC (complexity)
Unsolved problem in computer science: [math]\displaystyle{ \mathsf{NC} \overset{?}{=} \mathsf{P} }[/math] (more unsolved problems in computer science)
|
In computational complexity theory, the class NC (for "Nick's Class") is the set of decision problems decidable in polylogarithmic time on a parallel computer with a polynomial number of processors. In other words, a problem with input size n is in NC if there exist constants c and k such that it can be solved in time O((log n)c) using O(nk) parallel processors. Stephen Cook[1][2] coined the name "Nick's class" after Nick Pippenger, who had done extensive research[3] on circuits with polylogarithmic depth and polynomial size.[4]
Just as the class P can be thought of as the tractable problems (Cobham's thesis), so NC can be thought of as the problems that can be efficiently solved on a parallel computer.[5] NC is a subset of P because polylogarithmic parallel computations can be simulated by polynomial-time sequential ones. It is unknown whether NC = P, but most researchers suspect this to be false, meaning that there are probably some tractable problems that are "inherently sequential" and cannot significantly be sped up by using parallelism. Just as the class NP-complete can be thought of as "probably intractable", so the class P-complete, when using NC reductions, can be thought of as "probably not parallelizable" or "probably inherently sequential".
The parallel computer in the definition can be assumed to be a parallel, random-access machine (PRAM). That is a parallel computer with a central pool of memory, and any processor can access any bit of memory in constant time. The definition of NC is not affected by the choice of how the PRAM handles simultaneous access to a single bit by more than one processor. It can be CRCW, CREW, or EREW. See PRAM for descriptions of those models.
Equivalently, NC can be defined as those decision problems decidable by a uniform Boolean circuit (which can be calculated from the length of the input, for NC, we suppose we can compute the Boolean circuit of size n in logarithmic space in n) with polylogarithmic depth and a polynomial number of gates with a maximum fan-in of 2.
RNC is a class extending NC with access to randomness.
Problems in NC
As with P, by a slight abuse of language, one might classify function problems and search problems as being in NC. NC is known to include many problems, including
- Integer addition, multiplication and division;
- Matrix multiplication, determinant, inverse, rank;
- Polynomial GCD, by a reduction to linear algebra using Sylvester matrix
- Finding a maximal matching.
Often algorithms for those problems had to be separately invented and could not be naïvely adapted from well-known algorithms – Gaussian elimination and Euclidean algorithm rely on operations performed in sequence. One might contrast ripple carry adder with a carry-lookahead adder.
Example
An example of problem in NC1 is the parity check on a bit string.[6] The problem consists in counting the number of 1s in a string made of 1 and 0. A simple solution consists in summing all the string's bits. Since addition is associative, [math]\displaystyle{ x_1 + \cdots + x_n = (x_1 + \cdots + x_{\frac{n}{2}}) + (x_{\frac{n}{2} + 1} + \cdots + x_n) }[/math]. Recursively applying such property, it is possible to build a binary tree of length [math]\displaystyle{ O(log(n)) }[/math] in which every sum between two bits [math]\displaystyle{ x_i }[/math] and [math]\displaystyle{ x_j }[/math] is expressible by means of basic logical operators, e.g. through the boolean expression [math]\displaystyle{ (x_i \land \neg x_j) \lor (\neg x_i \land x_j) }[/math].
The NC hierarchy
NCi is the class of decision problems decidable by uniform boolean circuits with a polynomial number of gates of at most two inputs and depth O((log n)i), or the class of decision problems solvable in time O((log n)i) on a parallel computer with a polynomial number of processors. Clearly, we have
- [math]\displaystyle{ \mathsf{NC}^1 \subseteq \mathsf{NC}^2 \subseteq \cdots \subseteq \mathsf{NC}^i \subseteq \cdots \mathsf{NC} }[/math]
which forms the NC-hierarchy.
We can relate the NC classes to the space classes L and NL[7] and AC.[8]
- [math]\displaystyle{ \mathsf{NC}^1 \subseteq \mathsf{L} \subseteq \mathsf{NL} \subseteq \mathsf{AC}^1 \subseteq \mathsf{NC}^2 \subseteq \mathsf{P}. }[/math]
The NC classes are related to the AC classes, which are defined similarly, but with gates having unbounded fan-in. For each i, we have[5][8]
- [math]\displaystyle{ \mathsf{NC}^i \subseteq \mathsf{AC}^i \subseteq \mathsf{NC}^{i+1}. }[/math]
As an immediate consequence of this, we have that NC = AC.[9] It is known that both inclusions are strict for i = 0.[5]
Similarly, we have that NC is equivalent to the problems solvable on an alternating Turing machine restricted to at most two options at each step with O(log n) space and [math]\displaystyle{ (\log n)^{O(1)} }[/math] alternations.[10]
Open problem: Is NC proper?
One major open question in complexity theory is whether or not every containment in the NC hierarchy is proper. It was observed by Papadimitriou that, if NCi = NCi+1 for some i, then NCi = NCj for all j ≥ i, and as a result, NCi = NC. This observation is known as NC-hierarchy collapse because even a single equality in the chain of containments
- [math]\displaystyle{ \mathsf{NC}^1 \subseteq \mathsf{NC}^2 \subseteq \cdots }[/math]
implies that the entire NC hierarchy "collapses" down to some level i. Thus, there are 2 possibilities:
- [math]\displaystyle{ \mathsf{NC}^1 \subset \cdots \subset \mathsf{NC}^i \subset \cdots \subset \mathsf{NC}^{i+j} \subset \cdots \mathsf{NC} }[/math]
- [math]\displaystyle{ \mathsf{NC}^1 \subset \cdots \subset \mathsf{NC}^i = \cdots = \mathsf{NC}^{i+j} = \cdots \mathsf{NC} }[/math]
It is widely believed that (1) is the case, although no proof as to the truth of either statement has yet been discovered.
NC0
The special class NC0 operates only on a constant length of input bits. It is therefore described as the class of functions definable by uniform boolean circuits with constant depth and bounded fan-in.
Barrington's theorem
A branching program with n variables of width k and length m consists of a sequence of m instructions. Each of the instructions is a tuple (i, p, q) where i is the index of variable to check (1 ≤ i ≤ n), and p and q are functions from {1, 2, ..., k} to {1, 2, ..., k}. Numbers 1, 2, ..., k are called states of the branching program. The program initially starts in state 1, and each instruction (i, p, q) changes the state from x to p(x) or q(x), depending on whether the ith variable is 0 or 1. The function mapping an input to a final state of the program is called the yield of the program (more precisely, the yield on an input is the function mapping any initial state to the corresponding final state). The program accepts a set [math]\displaystyle{ A \subset 2^n }[/math] of variable values when there is some set of functions [math]\displaystyle{ F \subset k^k }[/math] such that a variable sequence [math]\displaystyle{ x \in 2^n }[/math] is in A precisely when its yield is in F.
A family of branching programs consists of a branching program with n variables for each n. It accepts a language when the n variable program accepts the language restricted to length n inputs.
It is easy to show that every language L on {0,1} can be recognized by a family of branching programs of width 5 and exponential length, or by a family of exponential width and linear length.
Every regular language on {0,1} can be recognized by a family of branching programs of constant width and linear number of instructions (since a DFA can be converted to a branching program). BWBP denotes the class of languages recognizable by a family of branching programs of bounded width and polynomial length.[11]
Barrington's theorem[12] says that BWBP is exactly nonuniform NC1. The proof uses the nonsolvability of the symmetric group S5.[11]
The theorem is rather surprising. For instance, it implies that the majority function can be computed by a family of branching programs of constant width and polynomial size, while intuition might suggest that to achieve polynomial size, one needs a linear number of states.
Proof of Barrington's theorem
A branching program of constant width and polynomial size can be easily converted (via divide-and-conquer) to a circuit in NC1.
Conversely, suppose a circuit in NC1 is given. Without loss of generality, assume it uses only AND and NOT gates.
Lemma 1 — If there exists a branching program that sometimes works as a permutation P and sometimes as a permutation Q, by right-multiplying permutations in the first instruction by α, and in the last instruction left-multiplying by β, we can make a circuit of the same length that behaves as βPα or βQα, respectively.
Call a branching program α-computing a circuit C if it works as identity when C's output is 0, and as α when C's output is 1.
As a consequence of Lemma 1 and the fact that all cycles of length 5 are conjugate, for any two 5-cycles α, β, if there exists a branching program α-computing a circuit C, then there exists a branching program β-computing the circuit C, of the same length.
Lemma 2 — There exist 5-cycles γ, δ such that their commutator ε=γδγ−1δ−1 is a 5-cycle. For example, γ = (1 2 3 4 5), δ = (1 3 5 4 2) giving ε = (1 3 2 5 4).
We will now prove Barrington's theorem by induction:
Suppose we have a circuit C which takes inputs x1,...,xn and assume that for all subcircuits D of C and 5-cycles α, there exists a branching program α-computing D. We will show that for all 5-cycles α, there exists a branching program α-computing C.
- If the circuit C simply outputs some input bit xi, the branching program we need has just one instruction: checking xi's value (0 or 1), and outputting the identity or α (respectively).
- If the circuit C outputs ¬A for some different circuit A, create a branching program α−1-computing A and then multiply the output of the program by α. By Lemma 1, we get a branching program for A outputting the identity or α, i.e. α-computing ¬A=C.
- If the circuit C outputs A∧B for circuits A and B, join the branching programs that γ-compute A, δ-compute B, γ−1-compute A, and δ−1-compute B for a choice of 5-cycles γ and δ such that their commutator ε=γδγ−1δ−1 is also a 5-cycle. (The existence of such elements was established in Lemma 2.) If one or both of the circuits outputs 0, the resulting program will be the identity due to cancellation; if both circuits output 1, the resulting program will output the commutator ε. In other words, we get a program ε-computing A∧B. Because ε and α are two 5-cycles, they are conjugate, and hence there exists a program α-computing A∧B by Lemma 1.
By assuming the subcircuits have branching programs so that they are α-computing for all 5-cycles α∈S5, we have shown C also has this property, as required.
The size of the branching program is at most 4d, where d is the depth of the circuit. If the circuit has logarithmic depth, the branching program has polynomial length.
Notes
- ↑ Cook, S.A. (1981). "Towards a complexity theory of synchronous parallel computation" (in en). L'Enseignement Mathématique 27: 99–124. http://citeseerx.ist.psu.edu/showciting?cid=1672592.
- ↑ Cook, Stephen A. (1985-01-01). "A taxonomy of problems with fast parallel algorithms" (in en). Information and Control. International Conference on Foundations of Computation Theory 64 (1): 2–22. doi:10.1016/S0019-9958(85)80041-3. ISSN 0019-9958.
- ↑ Pippenger, Nicholas (1979). "On simultaneous resource bounds" (in en). 20th Annual Symposium on Foundations of Computer Science (SFCS 1979): 307–311. doi:10.1109/SFCS.1979.29. ISSN 0272-5428. https://www.infona.pl//resource/bwmeta1.element.ieee-art-000004568025.
- ↑ Arora & Barak (2009) p.120
- ↑ 5.0 5.1 5.2 Arora & Barak (2009) p.118
- ↑ "Lecture 2: The Complexity of Some Problems". IAS/PCMI Summer Session 2000 - Clay Mathematics Undergraduate Program - Basic Course on Computational Complexity. Clarkson University. 2000-07-18. https://lin-web.clarkson.edu/~alexis/PCMI/Notes/lectureB02.pdf. Retrieved 2021-11-11.
- ↑ Papadimitriou (1994) Theorem 16.1
- ↑ 8.0 8.1 Clote & Kranakis (2002) p.437
- ↑ Clote & Kranakis (2002) p.12
- ↑ S. Bellantoni and I. Oitavem (2004). "Separating NC along the delta axis". Theoretical Computer Science 318 (1–2): 57–78. doi:10.1016/j.tcs.2003.10.021.
- ↑ 11.0 11.1 Clote & Kranakis (2002) p.50
- ↑ Barrington, David A. (1989). "Bounded-Width Polynomial-Size Branching Programs Recognize Exactly Those Languages in NC1". J. Comput. Syst. Sci. 38 (1): 150–164. doi:10.1016/0022-0000(89)90037-8. ISSN 0022-0000. http://www.cs.umass.edu/~barring/publications/bwbp.pdf.
References
- Arora, Sanjeev; Barak, Boaz (2009). Computational complexity. A modern approach. Cambridge University Press. ISBN 978-0-521-42426-4.
- Clote, Peter; Kranakis, Evangelos (2002). Boolean functions and computation models. Texts in Theoretical Computer Science. An EATCS Series. Berlin: Springer-Verlag. ISBN 3-540-59436-1.
- Greenlaw, Raymond, James Hoover, and Walter Ruzzo. Limits To Parallel computation; P-Completeness Theory. ISBN 0-19-508591-4
- Kozen, Dexter C. (1992). The design and analysis of algorithms. Lectures 28 - 34 and 36
- Kozen, Dexter C. (2006). Theory of Computation. Texts in Computer Science. Springer-Verlag. ISBN 1-84628-297-7. Lecture 12: Relation of NC to Time-Space Classes
- Papadimitriou, Christos (1993). "Section 15.3: The class NC". Computational Complexity (1st ed.). Addison Wesley. pp. 375–381. ISBN 0-201-53082-1.
- Straubing, Howard (1994). Finite automata, formal logic, and circuit complexity. Progress in Theoretical Computer Science. Basel: Birkhäuser. ISBN 3-7643-3719-2. https://archive.org/details/finiteautomatafo0000stra.
- Vollmer, Heribert (1998). Introduction to circuit complexity. A uniform approach. Texts in Theoretical Computer Science. Berlin: Springer-Verlag. ISBN 3-540-64310-9. https://books.google.com/books?id=55ZTgOJs8bsC.
Original source: https://en.wikipedia.org/wiki/NC (complexity).
Read more |