Hypercomputation Explained

Hypercomputation or super-Turing computation is a set of hypothetical models of computation that can provide outputs that are not Turing-computable. For example, a machine that could solve the halting problem would be a hypercomputer; so too would one that could correctly evaluate every statement in Peano arithmetic.

The Church–Turing thesis states that any "computable" function that can be computed by a mathematician with a pen and paper using a finite set of simple algorithms, can be computed by a Turing machine. Hypercomputers compute functions that a Turing machine cannot and which are, hence, not computable in the Church–Turing sense.

Technically, the output of a random Turing machine is uncomputable; however, most hypercomputing literature focuses instead on the computation of deterministic, rather than random, uncomputable functions.

History

A computational model going beyond Turing machines was introduced by Alan Turing in his 1938 PhD dissertation Systems of Logic Based on Ordinals.[1] This paper investigated mathematical systems in which an oracle was available, which could compute a single arbitrary (non-recursive) function from naturals to naturals. He used this device to prove that even in those more powerful systems, undecidability is still present. Turing's oracle machines are mathematical abstractions, and are not physically realizable.[2]

State space

In a sense, most functions are uncomputable: there are

\aleph0

computable functions, but there are an uncountable number (

\aleph0
2
) of possible super-Turing functions.[3]

Models

Hypercomputer models range from useful but probably unrealizable (such as Turing's original oracle machines), to less-useful random-function generators that are more plausibly "realizable" (such as a random Turing machine).

Uncomputable inputs or black-box components

A system granted knowledge of the uncomputable, oracular Chaitin's constant (a number with an infinite sequence of digits that encode the solution to the halting problem) as an input can solve a large number of useful undecidable problems; a system granted an uncomputable random-number generator as an input can create random uncomputable functions, but is generally not believed to be able to meaningfully solve "useful" uncomputable functions such as the halting problem. There are an unlimited number of different types of conceivable hypercomputers, including:

"Infinite computational steps" models

In order to work correctly, certain computations by the machines below literally require infinite, rather than merely unlimited but finite, physical space and resources; in contrast, with a Turing machine, any given computation that halts will require only finite physical space and resources.

A Turing machine that can complete infinitely many steps in finite time, a feat known as a supertask. Simply being able to run for an unbounded number of steps does not suffice. One mathematical model is the Zeno machine (inspired by Zeno's paradox). The Zeno machine performs its first computation step in (say) 1 minute, the second step in ½ minute, the third step in ¼ minute, etc. By summing 1 + ½ + ¼ + ... (a geometric series) we see that the machine performs infinitely many steps in a total of 2 minutes. According to Oron Shagrir, Zeno machines introduce physical paradoxes and its state is logically undefined outside of one-side open period of [0, 2), thus undefined exactly at 2 minutes after beginning of the computation.<ref>These models have been independently developed by many different authors, including {{cite book|author=Hermann Weyl| year=1927 | title=Philosophie der Mathematik und Naturwissenschaft| author-link=Hermann Weyl }}; the model is discussed in {{cite journal |author=[[Oron Shagrir|Shagrir, O.]] |title=Super-tasks, accelerating Turing machines and uncomputability |journal= Theoretical Computer Science |date=June 2004 |pages=105–114 |doi=10.1016/j.tcs.2003.12.007 |volume=317 |issue=1–3 |doi-access=free }}, Petrus H. Potgieter. Zeno machines and hypercomputation. Theoretical Computer Science. 358 . 1 . July 2006 . 23–33. 10.1016/j.tcs.2005.11.040. cs/0412022. 6749770. and Vincent C. Müller . On the possibilities of hypercomputing supertasks . Minds and Machines . 21 . 1 . 2011 . 83–96 . 10.1007/s11023-011-9222-6. 10.1.1.225.3696 . 253434 .

It seems natural that the possibility of time travel (existence of closed timelike curves (CTCs)) makes hypercomputation possible by itself. However, this is not so since a CTC does not provide (by itself) the unbounded amount of storage that an infinite computation would require. Nevertheless, there are spacetimes in which the CTC region can be used for relativistic hypercomputation.[13] According to a 1992 paper,[14] a computer operating in a Malament–Hogarth spacetime or in orbit around a rotating black hole[15] could theoretically perform non-Turing computations for an observer inside the black hole.[16] [17] Access to a CTC may allow the rapid solution to PSPACE-complete problems, a complexity class which, while Turing-decidable, is generally considered computationally intractable.[18] [19]

Quantum models

Some scholars conjecture that a quantum mechanical system which somehow uses an infinite superposition of states could compute a non-computable function.[20] This is not possible using the standard qubit-model quantum computer, because it is proven that a regular quantum computer is PSPACE-reducible (a quantum computer running in polynomial time can be simulated by a classical computer running in polynomial space).[21]

"Eventually correct" systems

Some physically realizable systems will always eventually converge to the correct answer, but have the defect that they will often output an incorrect answer and stick with the incorrect answer for an uncomputably large period of time before eventually going back and correcting the mistake.

In mid 1960s, E Mark Gold and Hilary Putnam independently proposed models of inductive inference (the "limiting recursive functionals"[22] and "trial-and-error predicates",[23] respectively). These models enable some nonrecursive sets of numbers or languages (including all recursively enumerable sets of languages) to be "learned in the limit"; whereas, by definition, only recursive sets of numbers or languages could be identified by a Turing machine. While the machine will stabilize to the correct answer on any learnable set in some finite time, it can only identify it as correct if it is recursive; otherwise, the correctness is established only by running the machine forever and noting that it never revises its answer. Putnam identified this new interpretation as the class of "empirical" predicates, stating: "if we always 'posit' that the most recently generated answer is correct, we will make a finite number of mistakes, but we will eventually get the correct answer. (Note, however, that even if we have gotten to the correct answer (the end of the finite sequence) we are never sure that we have the correct answer.)"[23] L. K. Schubert's 1974 paper "Iterated Limiting Recursion and the Program Minimization Problem"[24] studied the effects of iterating the limiting procedure; this allows any arithmetic predicate to be computed. Schubert wrote, "Intuitively, iterated limiting identification might be regarded as higher-order inductive inference performed collectively by an ever-growing community of lower order inductive inference machines."

A symbol sequence is computable in the limit if there is a finite, possibly non-halting program on a universal Turing machine that incrementally outputs every symbol of the sequence. This includes the dyadic expansion of π and of every other computable real, but still excludes all noncomputable reals. The 'Monotone Turing machines' traditionally used in description size theory cannot edit their previous outputs; generalized Turing machines, as defined by Jürgen Schmidhuber, can. He defines the constructively describable symbol sequences as those that have a finite, non-halting program running on a generalized Turing machine, such that any output symbol eventually converges; that is, it does not change any more after some finite initial time interval. Due to limitations first exhibited by Kurt Gödel (1931), it may be impossible to predict the convergence time itself by a halting program, otherwise the halting problem could be solved. Schmidhuber ([25] [26]) uses this approach to define the set of formally describable or constructively computable universes or constructive theories of everything. Generalized Turing machines can eventually converge to a correct solution of the halting problem by evaluating a Specker sequence.

Analysis of capabilities

Many hypercomputation proposals amount to alternative ways to read an oracle or advice function embedded into an otherwise classical machine. Others allow access to some higher level of the arithmetic hierarchy. For example, supertasking Turing machines, under the usual assumptions, would be able to compute any predicate in the truth-table degree containing

0
\Sigma
1
or
0
\Pi
1
. Limiting-recursion, by contrast, can compute any predicate or function in the corresponding Turing degree, which is known to be
0
\Delta
2
. Gold further showed that limiting partial recursion would allow the computation of precisely the
0
\Sigma
2
predicates.
ModelComputable predicatesNotesRefs
supertasking
0
\operatorname{tt}\left(\Sigma
1,
0
\Pi
1\right)
dependent on outside observer[27]
limiting/trial-and-error
0
\Delta
2
iterated limiting (k times)
0
\Delta
k+1

Blum–Shub–Smale machineincomparable with traditional computable real functions[28]
Malament–Hogarth spacetimeHYPdependent on spacetime structure[29]
analog recurrent neural network
0
\Delta
1[f]
f is an advice function giving connection weights; size is bounded by runtime[30] [31]
infinite time Turing machine

AQI

Arithmetical Quasi-Inductive sets[32]
classical fuzzy Turing machine
0
\Sigma
1

\cup

0
\Pi
1
for any computable t-norm
increasing function oracle
1
\Delta
1
for the one-sequence model;
1
\Pi
1
are r.e.

Criticism

Martin Davis, in his writings on hypercomputation,[33] [34] refers to this subject as "a myth" and offers counter-arguments to thephysical realizability of hypercomputation. As for its theory, he argues againstthe claims that this is a new field founded in the 1990s. This point of view relieson the history of computability theory (degrees of unsolvability, computability overfunctions, real numbers and ordinals), as also mentioned above.In his argument, he makes a remark that all of hypercomputation is little more than: "if non-computable inputs are permitted, then non-computable outputs are attainable."[35]

See also

Further reading

Notes and References

  1. 10.1112/plms/s2-45.1.161. Systems of Logic Based on Ordinals†. 1939. Turing. A. M.. Proceedings of the London Mathematical Society. 45. 161–228. 21.11116/0000-0001-91CE-3. free.
  2. "Let us suppose that we are supplied with some unspecified means of solving number-theoretic problems; a kind of oracle as it were. We shall not go any further into the nature of this oracle apart from saying that it cannot be a machine" (Undecidable p. 167, a reprint of Turing's paper Systems of Logic Based On Ordinals)
  3. The Computational Power of Interactive Recurrent Neural Networks . J. Cabessa . H.T. Siegelmann . Neural Computation . 24 . 4 . 996–1019 . Apr 2012 . 10.1162/neco_a_00263. 22295978 . 10.1.1.411.7540 . 5826757 .
  4. [Arnold Schönhage]
  5. Web site: The Professors and the Brainstorms . Andrew Hodges . 23 September 2011 . The Alan Turing Home Page .
  6. H.T. Siegelmann . E.D. Sontag . Analog Computation via Neural Networks . Theoretical Computer Science . 131 . 2 . 331–360 . 1994 . 10.1016/0304-3975(94)90178-3 . free .
  7. Biacino. L.. Gerla, G.. 2002. Fuzzy logic, continuity and effectiveness. Archive for Mathematical Logic. 0933-5846. 41. 7. 643–667. 10.1007/s001530100128. 10.1.1.2.8029. 12513452.
  8. Wiedermann. Jiří . 2004. Characterizing the super-Turing computing power and efficiency of classical fuzzy Turing machines. Theoretical Computer Science. 317. 1–3. 61–69. 10.1016/j.tcs.2003.12.004. Their (ability to solve the halting problem) is due to their acceptance criterion in which the ability to solve the halting problem is indirectly assumed.. free.
  9. Nondeterminism, Fairness and a Fundamental Analogy. EATCS Bulletin. 37. 186–193. 1989. Edith Spaan . Leen Torenvliet . Peter van Emde Boas .
  10. 10.1016/j.amc.2005.09.076. The many forms of hypercomputation. 2006. Ord. Toby. Applied Mathematics and Computation. 178. 143–153.
  11. Web site: Dmytro Taranovsky . July 17, 2005 . Finitism and Hypercomputation . Apr 26, 2011.
  12. Hewitt, Carl. "What Is Commitment." Physical, Organizational, and Social (Revised), Coordination, Organizations, Institutions, and Norms in Agent Systems II: AAMAS (2006).
  13. 1105.0047. 10.1142/S0129626412400105. Closed Timelike Curves in Relativistic Computation. 2012. Andréka. Hajnal. Németi. István. Székely. Gergely. Parallel Processing Letters. 22. 3. 16816151.
  14. 10.1007/BF00682813. Does general relativity allow an observer to view an eternity in a finite time?. 1992. Hogarth. Mark L.. Foundations of Physics Letters. 5. 2. 173–181. 1992FoPhL...5..173H. 120917288.
  15. Book: Can General Relativistic Computers Break the Turing Barrier? . István Neméti . Hajnal Andréka . Hajnal Andréka . Logical Approaches to Computational Barriers, Second Conference on Computability in Europe, CiE 2006, Swansea, UK, June 30-July 5, 2006. Proceedings . Springer . Lecture Notes in Computer Science . 3988 . 10.1007/11780342 . 2006 . 978-3-540-35466-6 . registration .
  16. gr-qc/0104023. Etesi. Gabor. Nemeti. Istvan. Non-Turing computations via Malament-Hogarth space-times. International Journal of Theoretical Physics. 2002. 41. 2. 341–370. 10.1023/A:1014019225365. 17081866.
  17. 10.1086/289716. Forever is a Day: Supertasks in Pitowsky and Malament-Hogarth Spacetimes. 1993. Earman. John. Norton. John D.. Philosophy of Science. 60. 22–42. 122764068.
  18. gr-qc/0209061. Brun. Todd A.. Computers with closed timelike curves can solve hard problems. Found. Phys. Lett.. 2003. 16. 3. 245–253. 10.1023/A:1025967225931. 16136314.
  19. [Scott Aaronson|S. Aaronson]
  20. There have been some claims to this effect; see Tien Kieu . Quantum Algorithm for the Hilbert's Tenth Problem . Int. J. Theor. Phys. . 2003 . 42 . quant-ph/0110136 . 1461–1478 . 10.1023/A:1025780028846 . 7. Hilbert problems . 6634980 . or M. Ziegler . Computational Power of Infinite Quantum Parallelism . 2005 . . 44 . 11 . 2059–2071 . 10.1007/s10773-005-8984-0. quant-ph/0410141 . 2005IJTP...44.2059Z . 9879859 . and the ensuing literature. For a retort see Warren D. Smith . 10.1016/j.amc.2005.09.078 . Three counterexamples refuting Kieu's plan for "quantum adiabatic hypercomputation"; and some uncomputable quantum mechanical tasks . Applied Mathematics and Computation . 178 . 1 . 184–193. 2006 . .
  21. Quantum Complexity Theory. 10.1137/S0097539796300921. 1997. Bernstein. Ethan. Vazirani. Umesh. SIAM Journal on Computing. 26. 5. 1411–1473.
  22. E. M. Gold . Limiting Recursion . Journal of Symbolic Logic . 30 . 1 . 28–48 . 1965 . 2270580 . 10.2307/2270580. 33811657 ., E. Mark Gold . Language identification in the limit . Information and Control . 10 . 447–474 . 1967 . 10.1016/S0019-9958(67)91165-5 . 5. free .
  23. Hilary Putnam . Trial and Error Predicates and the Solution to a Problem of Mostowksi . Journal of Symbolic Logic . 30 . 1 . 49–57 . 1965 . 2270581 . 10.2307/2270581. 44655062 .
  24. L. K. Schubert . Iterated Limiting Recursion and the Program Minimization Problem . Journal of the ACM . 21 . 3 . July 1974 . 10.1145/321832.321841. 436–445. 2071951 . free .
  25. quant-ph/0011122. Schmidhuber. Juergen. Algorithmic Theories of Everything. 2000.
  26. J. Schmidhuber . Hierarchies of generalized Kolmogorov complexities and nonenumerable universal measures computable in the limit . International Journal of Foundations of Computer Science . 13 . 4 . 587–612 . 2002 . 10.1142/S0129054102001291 . quant-ph/0011122 . 2000quant.ph.11122S.
  27. Petrus H. Potgieter. Zeno machines and hypercomputation. Theoretical Computer Science. 358 . 1 . July 2006 . 23–33. 10.1016/j.tcs.2005.11.040. cs/0412022. 6749770.
  28. Book: Lenore Blum, Felipe Cucker, Michael Shub, and Stephen Smale. Complexity and Real Computation. Complexity and Real Computation . 978-0-387-98281-6. 1998. Springer .
  29. P.D. Welch . The extent of computation in Malament-Hogarth spacetimes . gr-qc/0609035 . British Journal for the Philosophy of Science . 59 . 4 . 2008 . 659–674 . 10.1093/bjps/axn031. P.D. Welch .
  30. 10.1126/science.268.5210.545 . 17756722 . Computation Beyond the Turing Limit . H.T. Siegelmann . Science . 268 . 5210 . 545–548 . Apr 1995 . 1995Sci...268..545S . 17495161 .
  31. Hava Siegelmann . Eduardo Sontag . Analog Computation via Neural Networks . Theoretical Computer Science . 131 . 1994 . 331–360 . 10.1016/0304-3975(94)90178-3 . 2 . Eduardo Sontag. Hava Siegelmann . free .
  32. P.D. Welch . Characteristics of discrete transfinite time Turing machine models: Halting times, stabilization times, and Normal Form theorems . Theoretical Computer Science . 2009 . 410 . 4–5 . 426–442 . 10.1016/j.tcs.2008.09.050 . P.D. Welch . free .
  33. Davis, Martin . Why there is no such discipline as hypercomputation . Applied Mathematics and Computation . 178 . 1 . 2006 . 4–7 . 10.1016/j.amc.2005.09.066.
  34. Book: Davis, Martin. Alan Turing: Life and Legacy of a Great Thinker. Springer. 2004 . The Myth of Hypercomputation.
  35. Book: Miniworkshop: Hilbert's Tenth Problem, Mazur's Conjecture and Divisibility Sequences . Martin Davis . The Myth of Hypercomputation . Alexandra Shlapentokh . Mathematisches Forschungsinstitut Oberwolfach . MFO Report . 3 . 2 . Jan 2003 .