Parallel computation thesis explained
In computational complexity theory, the parallel computation thesis is a hypothesis which states that the time used by a (reasonable) parallel machine is polynomially related to the space used by a sequential machine. The parallel computation thesis was set forth by Chandra and Stockmeyer in 1976.[1]
In other words, for a computational model which allows computations to branch and run in parallel without bound, a formal language which is decidable under the model using no more than
steps for inputs of length
n is decidable by a non-branching machine using no more than
units of storage for some constant
k. Similarly, if a machine in the unbranching model decides a language using no more than
storage, a machine in the parallel model can decide the language in no more than
steps for some constant
k.
The parallel computation thesis is not a rigorous formal statement, as it does not clearly define what constitutes an acceptable parallel model. A parallel machine must be sufficiently powerful to emulate the sequential machine in time polynomially related to the sequential space; compare Turing machine, non-deterministic Turing machine, and alternating Turing machine. N. Blum (1983) introduced a model for which the thesis does not hold.[2] However, the model allows
parallel threads of computation after
steps. (See
Big O notation.) Parberry (1986) suggested a more "reasonable" bound would be
or
, in defense of the thesis.
[3] Goldschlager (1982) proposed a model which is sufficiently universal to emulate all "reasonable" parallel models, which adheres to the thesis.
[4] Chandra and Stockmeyer originally formalized and proved results related to the thesis for deterministic and alternating Turing machines, which is where the thesis originated.
[5] Notes and References
- Chandra. Ashok K.. Stockmeyer. Larry J.. Alternation. FOCS'76: Proceedings of the 17th Annual Symposium on Foundations of Computer Science. 98–108. 10.1109/SFCS.1976.4. 1976.
- Information Processing Letters. Blum. Norbert. A note on the 'parallel computation thesis'. 17. 4. 203–205. 1983. 10.1016/0020-0190(83)90041-8.
- 10.1145/8312.8317. Parberry. I.. Parallel speedup of sequential machines: a defense of parallel computation thesis. ACM SIGACT News. 18. 1. 54–67. 1986. free.
- 10.1145/322344.322353. Goldschlager. Leslie M.. A universal interconnection pattern for parallel computers. Journal of the ACM. 29. 3. 1073–1086. 1982. free.
- 10.1145/322234.322243. Chandra. Ashok K.. Kozen. Dexter C.. Stockmeyer. Larry J.. Alternation. Journal of the ACM. 28. 1. 114–133. 1981. free.