The number (; spelled out as "pi") is a mathematical constant that is the ratio of a circle's circumference to its diameter, approximately equal to 3.14159. The number appears in many formulae across mathematics and physics. It is an irrational number, meaning that it cannot be expressed exactly as a ratio of two integers, although fractions such as
\tfrac{22}{7}
For thousands of years, mathematicians have attempted to extend their understanding of, sometimes by computing its value to a high degree of accuracy. Ancient civilizations, including the Egyptians and Babylonians, required fairly accurate approximations of for practical computations. Around 250BC, the Greek mathematician Archimedes created an algorithm to approximate with arbitrary accuracy. In the 5th century AD, Chinese mathematicians approximated to seven digits, while Indian mathematicians made a five-digit approximation, both using geometrical techniques. The first computational formula for, based on infinite series, was discovered a millennium later.[1] The earliest known use of the Greek letter π to represent the ratio of a circle's circumference to its diameter was by the Welsh mathematician William Jones in 1706.[2]
The invention of calculus soon led to the calculation of hundreds of digits of, enough for all practical scientific computations. Nevertheless, in the 20th and 21st centuries, mathematicians and computer scientists have pursued new approaches that, when combined with increasing computational power, extended the decimal representation of to many trillions of digits.[3] [4] These computations are motivated by the development of efficient algorithms to calculate numeric series, as well as the human quest to break records.[5] The extensive computations involved have also been used to test supercomputers as well as stress testing consumer computer hardware.
Because its definition relates to the circle, is found in many formulae in trigonometry and geometry, especially those concerning circles, ellipses and spheres. It is also found in formulae from other topics in science, such as cosmology, fractals, thermodynamics, mechanics, and electromagnetism. It also appears in areas having little to do with geometry, such as number theory and statistics, and in modern mathematical analysis can be defined without any reference to geometry. The ubiquity of makes it one of the most widely known mathematical constants inside and outside of science. Several books devoted to have been published, and record-setting calculations of the digits of often result in news headlines.
The symbol used by mathematicians to represent the ratio of a circle's circumference to its diameter is the lowercase Greek letter , sometimes spelled out as pi. In English, is pronounced as "pie" .[6] In mathematical use, the lowercase letter is distinguished from its capitalized and enlarged counterpart, which denotes a product of a sequence, analogous to how denotes summation.
The choice of the symbol is discussed in the section Adoption of the symbol .
is commonly defined as the ratio of a circle's circumference to its diameter :
The ratio is constant, regardless of the circle's size. For example, if a circle has twice the diameter of another circle, it will also have twice the circumference, preserving the ratio . This definition of implicitly makes use of flat (Euclidean) geometry; although the notion of a circle can be extended to any curve (non-Euclidean) geometry, these new circles will no longer satisfy the formula .
Here, the circumference of a circle is the arc length around the perimeter of the circle, a quantity which can be formally defined independently of geometry using limits—a concept in calculus.[7] For example, one may directly compute the arc length of the top half of the unit circle, given in Cartesian coordinates by the equation , as the integral:
An integral such as this was adopted as the definition of by Karl Weierstrass, who defined it directly as an integral in 1841.
Integration is no longer commonly used in a first analytical definition because, as explains, differential calculus typically precedes integral calculus in the university curriculum, so it is desirable to have a definition of that does not rely on the latter. One such definition, due to Richard Baltzer[8] and popularized by Edmund Landau,[9] is the following: is twice the smallest positive number at which the cosine function equals 0.[10] is also the smallest positive number at which the sine function equals zero, and the difference between consecutive zeroes of the sine function. The cosine and sine can be defined independently of geometry as a power series,[11] or as the solution of a differential equation.
In a similar spirit, can be defined using properties of the complex exponential,, of a complex variable . Like the cosine, the complex exponential can be defined in one of several ways. The set of complex numbers at which is equal to one is then an (imaginary) arithmetic progression of the form:and there is a unique positive real number with this property.[12]
A variation on the same idea, making use of sophisticated mathematical concepts of topology and algebra, is the following theorem:[13] there is a unique (up to automorphism) continuous isomorphism from the group R/Z of real numbers under addition modulo integers (the circle group), onto the multiplicative group of complex numbers of absolute value one. The number is then defined as half the magnitude of the derivative of this homomorphism.[14]
is an irrational number, meaning that it cannot be written as the ratio of two integers. Fractions such as and are commonly used to approximate, but no common fraction (ratio of whole numbers) can be its exact value. Because is irrational, it has an infinite number of digits in its decimal representation, and does not settle into an infinitely repeating pattern of digits. There are several proofs that is irrational; they generally require calculus and rely on the reductio ad absurdum technique. The degree to which can be approximated by rational numbers (called the irrationality measure) is not precisely known; estimates have established that the irrationality measure is larger than the measure of or but smaller than the measure of Liouville numbers.[15]
The digits of have no apparent pattern and have passed tests for statistical randomness, including tests for normality; a number of infinite length is called normal when all possible sequences of digits (of any given length) appear equally often. The conjecture that is normal has not been proven or disproven.
Since the advent of computers, a large number of digits of have been available on which to perform statistical analysis. Yasumasa Kanada has performed detailed statistical analyses on the decimal digits of, and found them consistent with normality; for example, the frequencies of the ten digits 0 to 9 were subjected to statistical significance tests, and no evidence of a pattern was found. Any random sequence of digits contains arbitrarily long subsequences that appear non-random, by the infinite monkey theorem. Thus, because the sequence of 's digits passes statistical tests for randomness, it contains some sequences of digits that may appear non-random, such as a sequence of six consecutive 9s that begins at the 762nd decimal place of the decimal representation of . This is also called the "Feynman point" in mathematical folklore, after Richard Feynman, although no connection to Feynman is known.
See also: Lindemann–Weierstrass theorem. In addition to being irrational, is also a transcendental number, which means that it is not the solution of any non-constant polynomial equation with rational coefficients, such as .
The transcendence of has two important consequences: First, cannot be expressed using any finite combination of rational numbers and square roots or n-th roots (such as
\sqrt[3]{31}
\sqrt{10}
As an irrational number, cannot be represented as a common fraction. But every number, including, can be represented by an infinite series of nested fractions, called a continued fraction:
Truncating the continued fraction at any point yields a rational approximation for ; the first four of these are,,, and . These numbers are among the best-known and most widely used historical approximations of the constant. Each approximation generated in this way is a best rational approximation; that is, each is closer to than any other fraction with the same or a smaller denominator. Because is transcendental, it is by definition not algebraic and so cannot be a quadratic irrational. Therefore, cannot have a periodic continued fraction. Although the simple continued fraction for (shown above) also does not exhibit any other obvious pattern,[18] several generalized continued fractions do, such as:[19]
The middle of these is due to the mid-17th century mathematician William Brouncker, see § Brouncker's formula.
Some approximations of pi include:
Digits in other number systems
Any complex number, say, can be expressed using a pair of real numbers. In the polar coordinate system, one number (radius or) is used to represent 's distance from the origin of the complex plane, and the other (angle or) the counter-clockwise rotation from the positive real line:where is the imaginary unit satisfying
i2=-1
\varphi=\pi
There are different complex numbers satisfying
zn=1
See also: Chronology of computation of π.
The best-known approximations to dating before the Common Era were accurate to two decimal places; this was improved upon in Chinese mathematics in particular by the mid-first millennium, to an accuracy of seven decimal places.After this, no further progress was made until the late medieval period.
The earliest written approximations of are found in Babylon and Egypt, both within one percent of the true value. In Babylon, a clay tablet dated 1900–1600 BC has a geometrical statement that, by implication, treats as = 3.125. In Egypt, the Rhind Papyrus, dated around 1650 BC but copied from a document dated to 1850 BC, has a formula for the area of a circle that treats as .[18] Although some pyramidologists have theorized that the Great Pyramid of Giza was built with proportions related to, this theory is not widely accepted by scholars.[22] In the Shulba Sutras of Indian mathematics, dating to an oral tradition from the first or second millennium BC, approximations are given which have been variously interpreted as approximately 3.08831, 3.08833, 3.004, 3, or 3.125.[23]
The first recorded algorithm for rigorously calculating the value of was a geometrical approach using polygons, devised around 250 BC by the Greek mathematician Archimedes, implementing the method of exhaustion. This polygonal algorithm dominated for over 1,000 years, and as a result is sometimes referred to as Archimedes's constant. Archimedes computed upper and lower bounds of by drawing a regular hexagon inside and outside a circle, and successively doubling the number of sides until he reached a 96-sided regular polygon. By calculating the perimeters of these polygons, he proved that (that is,).[24] Archimedes' upper bound of may have led to a widespread popular belief that is equal to . Around 150 AD, Greek-Roman scientist Ptolemy, in his Almagest, gave a value for of 3.1416, which he may have obtained from Archimedes or from Apollonius of Perga. Mathematicians using polygonal algorithms reached 39 digits of in 1630, a record only broken in 1699 when infinite series were used to reach 71 digits.[25]
In ancient China, values for included 3.1547 (around 1 AD),
\sqrt{10}
3.1415926<\pi<3.1415927
The Indian astronomer Aryabhata used a value of 3.1416 in his Āryabhaṭīya (499 AD). Fibonacci in computed 3.1418 using a polygonal method, independent of Archimedes. Italian author Dante apparently employed the value .
The Persian astronomer Jamshīd al-Kāshī produced nine sexagesimal digits, roughly the equivalent of 16 decimal digits, in 1424, using a polygon with sides,[26] [27] which stood as the world record for about 180 years. French mathematician François Viète in 1579 achieved nine digits with a polygon of sides. Flemish mathematician Adriaan van Roomen arrived at 15 decimal places in 1593. In 1596, Dutch mathematician Ludolph van Ceulen reached 20 digits, a record he later increased to 35 digits (as a result, was called the "Ludolphian number" in Germany until the early 20th century). Dutch scientist Willebrord Snellius reached 34 digits in 1621, and Austrian astronomer Christoph Grienberger arrived at 38 digits in 1630 using 1040 sides.[28] Christiaan Huygens was able to arrive at 10 decimal places in 1654 using a slightly different method equivalent to Richardson extrapolation.[29] [30]
The calculation of was revolutionized by the development of infinite series techniques in the 16th and 17th centuries. An infinite series is the sum of the terms of an infinite sequence. Infinite series allowed mathematicians to compute with much greater precision than Archimedes and others who used geometrical techniques. Although infinite series were exploited for most notably by European mathematicians such as James Gregory and Gottfried Wilhelm Leibniz, the approach also appeared in the Kerala school sometime in the 14th or 15th century. Around 1500 AD, a written description of an infinite series that could be used to compute was laid out in Sanskrit verse in Tantrasamgraha by Nilakantha Somayaji.[31] The series are presented without proof, but proofs are presented in a later work, Yuktibhāṣā, from around 1530 AD. Several infinite series are described, including series for sine (which Nilakantha attributes to Madhava of Sangamagrama), cosine, and arctangent which are now sometimes referred to as Madhava series. The series for arctangent is sometimes called Gregory's series or the Gregory–Leibniz series. Madhava used infinite series to estimate to 11 digits around 1400.[32]
In 1593, François Viète published what is now known as Viète's formula, an infinite product (rather than an infinite sum, which is more typically used in calculations):[33]
In 1655, John Wallis published what is now known as Wallis product, also an infinite product:
In the 1660s, the English scientist Isaac Newton and German mathematician Gottfried Wilhelm Leibniz discovered calculus, which led to the development of many infinite series for approximating . Newton himself used an arcsine series to compute a 15-digit approximation of in 1665 or 1666, writing, "I am ashamed to tell you to how many figures I carried these computations, having no other business at the time."[34]
In 1671, James Gregory, and independently, Leibniz in 1673, discovered the Taylor series expansion for arctangent:[31] [35]
This series, sometimes called the Gregory–Leibniz series, equals when evaluated with
z=1
z=1
In 1699, English mathematician Abraham Sharp used the Gregory–Leibniz series for to compute to 71 digits, breaking the previous record of 39 digits, which was set with a polygonal algorithm.
In 1706, John Machin used the Gregory–Leibniz series to produce an algorithm that converged much faster:[2] [37]
Machin reached 100 digits of with this formula. Other mathematicians created variants, now known as Machin-like formulae, that were used to set several successive records for calculating digits of .[38]
Isaac Newton accelerated the convergence of the Gregory–Leibniz series in 1684 (in an unpublished work; others independently discovered the result):[39]
\arctanx =
x | |
1+x2 |
+
| ||||
+ |
2 ⋅ 4 | |
3 ⋅ 5 |
x5 | |
(1+x2)3 |
+ …
Leonhard Euler popularized this series in his 1755 differential calculus textbook, and later used it with Machin-like formulae, including with which he computed 20 digits of in one hour.[40]
Machin-like formulae remained the best-known method for calculating well into the age of computers, and were used to set records for 250 years, culminating in a 620-digit approximation in 1946 by Daniel Ferguson – the best approximation achieved without the aid of a calculating device.
In 1844, a record was set by Zacharias Dase, who employed a Machin-like formula to calculate 200 decimals of in his head at the behest of German mathematician Carl Friedrich Gauss.
In 1853, British mathematician William Shanks calculated to 607 digits, but made a mistake in the 528th digit, rendering all subsequent digits incorrect. Though he calculated an additional 100 digits in 1873, bringing the total up to 707, his previous mistake rendered all the new digits incorrect as well.[41]
Some infinite series for converge faster than others. Given the choice of two infinite series for, mathematicians will generally use the one that converges more rapidly because faster convergence reduces the amount of computation needed to calculate to any given accuracy.[42] A simple infinite series for is the Gregory–Leibniz series:
As individual terms of this infinite series are added to the sum, the total gradually gets closer to, and – with a sufficient number of terms – can get as close to as desired. It converges quite slowly, though – after 500,000 terms, it produces only five correct decimal digits of .[43]
An infinite series for (published by Nilakantha in the 15th century) that converges more rapidly than the Gregory–Leibniz series is:[44]
The following table compares the convergence rates of these two series:
Infinite series for | After 1st term | After 2nd term | After 3rd term | After 4th term | After 5th term | Converges to: | |||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
\pi=
-
+
-
+
-
+
+ … | 4.0000 | 2.6666 ... | 3.4666 ... | 2.8952 ... | 3.3396 ... | = 3.1415 ... | |||||||||||||||||||||
\pi={{3}}+
| 3.0000 | 3.1666 ... | 3.1333 ... | 3.1452 ... | 3.1396 ... |
After five terms, the sum of the Gregory–Leibniz series is within 0.2 of the correct value of, whereas the sum of Nilakantha's series is within 0.002 of the correct value. Nilakantha's series converges faster and is more useful for computing digits of . Series that converge even faster include Machin's series and Chudnovsky's series, the latter producing 14 correct decimal digits per term.
Not all mathematical advances relating to were aimed at increasing the accuracy of approximations. When Euler solved the Basel problem in 1735, finding the exact value of the sum of the reciprocal squares, he established a connection between and the prime numbers that later contributed to the development and study of the Riemann zeta function:
Swiss scientist Johann Heinrich Lambert in 1768 proved that is irrational, meaning it is not equal to the quotient of any two integers. Lambert's proof exploited a continued-fraction representation of the tangent function.[45] French mathematician Adrien-Marie Legendre proved in 1794 that 2 is also irrational. In 1882, German mathematician Ferdinand von Lindemann proved that is transcendental,[46] confirming a conjecture made by both Legendre and Euler.[47] Hardy and Wright states that "the proofs were afterwards modified and simplified by Hilbert, Hurwitz, and other writers".[48]
In the earliest usages, the Greek letter was used to denote the semiperimeter (semiperipheria in Latin) of a circle[49] and was combined in ratios with (for diameter or semidiameter) or (for radius) to form circle constants.[50] [51] [52] [53] (Before then, mathematicians sometimes used letters such as or instead.) The first recorded use is Oughtred's, to express the ratio of periphery and diameter in the 1647 and later editions of Latin: Clavis Mathematicae|italic=yes.[54] Barrow likewise used to represent the constant,[55] while Gregory instead used to represent .[56]
The earliest known use of the Greek letter alone to represent the ratio of a circle's circumference to its diameter was by Welsh mathematician William Jones in his 1706 work Latin: Synopsis Palmariorum Matheseos|italic=unset; or, a New Introduction to the Mathematics.[2] The Greek letter appears on p. 243 in the phrase " Periphery ", calculated for a circle with radius one. However, Jones writes that his equations for are from the "ready pen of the truly ingenious Mr. John Machin", leading to speculation that Machin may have employed the Greek letter before Jones. Jones' notation was not immediately adopted by other mathematicians, with the fraction notation still being used as late as 1767.[57]
Euler started using the single-letter form beginning with his 1727 Essay Explaining the Properties of Air, though he used, the ratio of periphery to radius, in this and some later writing.[58] [59] Euler first used in his 1736 work Mechanica,[60] and continued in his widely read 1748 work Latin: [[Introductio in analysin infinitorum]]|italic=yes (he wrote: "for the sake of brevity we will write this number as ; thus is equal to half the circumference of a circle of radius ").[61] Because Euler corresponded heavily with other mathematicians in Europe, the use of the Greek letter spread rapidly, and the practice was universally adopted thereafter in the Western world, though the definition still varied between and as late as 1761.[62]
The development of computers in the mid-20th century again revolutionized the hunt for digits of . Mathematicians John Wrench and Levi Smith reached 1,120 digits in 1949 using a desk calculator. Using an inverse tangent (arctan) infinite series, a team led by George Reitwiesner and John von Neumann that same year achieved 2,037 digits with a calculation that took 70 hours of computer time on the ENIAC computer.[63] The record, always relying on an arctan series, was broken repeatedly (3089 digits in 1955,[64] 7,480 digits in 1957; 10,000 digits in 1958; 100,000 digits in 1961) until 1 million digits were reached in 1973.
Two additional developments around 1980 once again accelerated the ability to compute . First, the discovery of new iterative algorithms for computing, which were much faster than the infinite series; and second, the invention of fast multiplication algorithms that could multiply large numbers very rapidly. Such algorithms are particularly important in modern computations because most of the computer's time is devoted to multiplication. They include the Karatsuba algorithm, Toom–Cook multiplication, and Fourier transform-based methods.
The iterative algorithms were independently published in 1975–1976 by physicist Eugene Salamin and scientist Richard Brent. These avoid reliance on infinite series. An iterative algorithm repeats a specific calculation, each iteration using the outputs from prior steps as its inputs, and produces a result in each step that converges to the desired value. The approach was actually invented over 160 years earlier by Carl Friedrich Gauss, in what is now termed the arithmetic–geometric mean method (AGM method) or Gauss–Legendre algorithm. As modified by Salamin and Brent, it is also referred to as the Brent–Salamin algorithm.
The iterative algorithms were widely used after 1980 because they are faster than infinite series algorithms: whereas infinite series typically increase the number of correct digits additively in successive terms, iterative algorithms generally multiply the number of correct digits at each step. For example, the Brent–Salamin algorithm doubles the number of digits in each iteration. In 1984, brothers John and Peter Borwein produced an iterative algorithm that quadruples the number of digits in each step; and in 1987, one that increases the number of digits five times in each step.[65] Iterative methods were used by Japanese mathematician Yasumasa Kanada to set several records for computing between 1995 and 2002. This rapid convergence comes at a price: the iterative algorithms require significantly more memory than infinite series.[66]
For most numerical calculations involving, a handful of digits provide sufficient precision. According to Jörg Arndt and Christoph Haenel, thirty-nine digits are sufficient to perform most cosmological calculations, because that is the accuracy necessary to calculate the circumference of the observable universe with a precision of one atom. Accounting for additional digits needed to compensate for computational round-off errors, Arndt concludes that a few hundred digits would suffice for any scientific application. Despite this, people have worked strenuously to compute to thousands and millions of digits. This effort may be partly ascribed to the human compulsion to break records, and such achievements with often make headlines around the world.[67] [68] They also have practical benefits, such as testing supercomputers, testing numerical analysis algorithms (including high-precision multiplication algorithms); and within pure mathematics itself, providing data for evaluating the randomness of the digits of .
Modern calculators do not use iterative algorithms exclusively. New infinite series were discovered in the 1980s and 1990s that are as fast as iterative algorithms, yet are simpler and less memory intensive. The fast iterative algorithms were anticipated in 1914, when Indian mathematician Srinivasa Ramanujan published dozens of innovative new formulae for, remarkable for their elegance, mathematical depth and rapid convergence. One of his formulae, based on modular equations, is
This series converges much more rapidly than most arctan series, including Machin's formula. Bill Gosper was the first to use it for advances in the calculation of, setting a record of 17 million digits in 1985. Ramanujan's formulae anticipated the modern algorithms developed by the Borwein brothers (Jonathan and Peter) and the Chudnovsky brothers. The Chudnovsky formula developed in 1987 is
It produces about 14 digits of per term and has been used for several record-setting calculations, including the first to surpass 1 billion (109) digits in 1989 by the Chudnovsky brothers, 10 trillion (1013) digits in 2011 by Alexander Yee and Shigeru Kondo,[69] and 100 trillion digits by Emma Haruka Iwao in 2022.[70] For similar formulae, see also the Ramanujan–Sato series.
In 2006, mathematician Simon Plouffe used the PSLQ integer relation algorithm[71] to generate several new formulae for, conforming to the following template:where is (Gelfond's constant), is an odd number, and are certain rational numbers that Plouffe computed.[72]
Monte Carlo methods, which evaluate the results of multiple random trials, can be used to create approximations of . Buffon's needle is one such technique: If a needle of length is dropped times on a surface on which parallel lines are drawn units apart, and if of those times it comes to rest crossing a line ( > 0), then one may approximate based on the counts:[73]
Another Monte Carlo method for computing is to draw a circle inscribed in a square, and randomly place dots in the square. The ratio of dots inside the circle to the total number of dots will approximately equal .
Another way to calculate using probability is to start with a random walk, generated by a sequence of (fair) coin tosses: independent random variables such that with equal probabilities. The associated random walk isso that, for each, is drawn from a shifted and scaled binomial distribution. As varies, defines a (discrete) stochastic process. Then can be calculated by[74]
This Monte Carlo method is independent of any relation to circles, and is a consequence of the central limit theorem, discussed below.
These Monte Carlo methods for approximating are very slow compared to other methods, and do not provide any information on the exact number of digits that are obtained. Thus they are never used to approximate when speed or accuracy is desired.
Two algorithms were discovered in 1995 that opened up new avenues of research into . They are called spigot algorithms because, like water dripping from a spigot, they produce single digits of that are not reused after they are calculated.[75] This is in contrast to infinite series or iterative algorithms, which retain and use all intermediate digits until the final result is produced.
Mathematicians Stan Wagon and Stanley Rabinowitz produced a simple spigot algorithm in 1995.[76] Its speed is comparable to arctan algorithms, but not as fast as iterative algorithms.
Another spigot algorithm, the BBP digit extraction algorithm, was discovered in 1995 by Simon Plouffe:[77]
This formula, unlike others before it, can produce any individual hexadecimal digit of without calculating all the preceding digits. Individual binary digits may be extracted from individual hexadecimal digits, and octal digits can be extracted from one or two hexadecimal digits. An important application of digit extraction algorithms is to validate new claims of record computations: After a new record is claimed, the decimal result is converted to hexadecimal, and then a digit extraction algorithm is used to calculate several randomly selected hexadecimal digits near the end; if they match, this provides a measure of confidence that the entire computation is correct.
Between 1998 and 2000, the distributed computing project PiHex used Bellard's formula (a modification of the BBP algorithm) to compute the quadrillionth (1015th) bit of, which turned out to be 0.[78] In September 2010, a Yahoo! employee used the company's Hadoop application on one thousand computers over a 23-day period to compute 256 bits of at the two-quadrillionth (2×1015th) bit, which also happens to be zero.[79]
In 2022, Plouffe found a base-10 algorithm for calculating digits of .[80]
Because is closely related to the circle, it is found in many formulae from the fields of geometry and trigonometry, particularly those concerning circles, spheres, or ellipses. Other branches of science, such as statistics, physics, Fourier analysis, and number theory, also include in some of their important formulae.
appears in formulae for areas and volumes of geometrical shapes based on circles, such as ellipses, spheres, cones, and tori. Below are some of the more common formulae that involve .
Some of the formulae above are special cases of the volume of the n-dimensional ball and the surface area of its boundary, the (n−1)-dimensional sphere, given below.
Apart from circles, there are other curves of constant width. By Barbier's theorem, every curve of constant width has perimeter times its width. The Reuleaux triangle (formed by the intersection of three circles with the sides of an equilateral triangle as their radii) has the smallest possible area for its width and the circle the largest. There also exist non-circular smooth and even algebraic curves of constant width.[81]
Definite integrals that describe circumference, area, or volume of shapes generated by circles typically have values that involve . For example, an integral that specifies half the area of a circle of radius one is given by:[82]
In that integral, the function
\sqrt{1-x2}
x
The trigonometric functions rely on angles, and mathematicians generally use radians as units of measurement. plays an important role in angles measured in radians, which are defined so that a complete circle spans an angle of 2 radians. The angle measure of 180° is equal to radians, and .
Common trigonometric functions have periods that are multiples of ; for example, sine and cosine have period 2, so for any angle and any integer,
Many of the appearances of in the formulae of mathematics and the sciences have to do with its close relationship with geometry. However, also appears in many natural situations having apparently nothing to do with geometry.
f''(x)+λf(x)=0
f''(t)=-λf(x)
f\mapstof''
The value is, in fact, the least such value of the wavenumber, and is associated with the fundamental mode of vibration of the string. One way to show this is by estimating the energy, which satisfies Wirtinger's inequality: for a function
f:[0,1]\to\Complex
1 | |
H | |
0[0,1] |
The number serves appears in similar eigenvalue problems in higher-dimensional analysis. As mentioned above, it can be characterized via its role as the best constant in the isoperimetric inequality: the area enclosed by a plane Jordan curve of perimeter satisfies the inequalityand equality is clearly achieved for the circle, since in that case and .[84]
Ultimately, as a consequence of the isoperimetric inequality, appears in the optimal constant for the critical Sobolev inequality in n dimensions, which thus characterizes the role of in many physical phenomena as well, for example those of classical potential theory.[85] [86] [87] In two dimensions, the critical Sobolev inequality isfor f a smooth function with compact support in,
\nablaf
\|f\|2
\|\nablaf\|1
Wirtinger's inequality also generalizes to higher-dimensional Poincaré inequalities that provide best constants for the Dirichlet energy of an n-dimensional membrane. Specifically, is the greatest constant such thatfor all convex subsets of of diameter 1, and square-integrable functions u on of mean zero.[88] Just as Wirtinger's inequality is the variational form of the Dirichlet eigenvalue problem in one dimension, the Poincaré inequality is the variational form of the Neumann eigenvalue problem, in any dimension.
The constant also appears as a critical spectral parameter in the Fourier transform. This is the integral transform, that takes a complex-valued integrable function on the real line to the function defined as:
Although there are several different conventions for the Fourier transform and its inverse, any such convention must involve somewhere. The above is the most canonical definition, however, giving the unique unitary operator on that is also an algebra homomorphism of to .[89]
The Heisenberg uncertainty principle also contains the number . The uncertainty principle gives a sharp lower bound on the extent to which it is possible to localize a function both in space and in frequency: with our conventions for the Fourier transform,
The physical consequence, about the uncertainty in simultaneous position and momentum observations of a quantum mechanical system, is discussed below. The appearance of in the formulae of Fourier analysis is ultimately a consequence of the Stone–von Neumann theorem, asserting the uniqueness of the Schrödinger representation of the Heisenberg group.[90]
The fields of probability and statistics frequently use the normal distribution as a simple model for complex phenomena; for example, scientists generally assume that the observational error in most experiments follows a normal distribution.[91] The Gaussian function, which is the probability density function of the normal distribution with mean and standard deviation, naturally contains :
The factor of
\tfrac{1}{\sqrt{2\pi}}
The central limit theorem explains the central role of normal distributions, and thus of, in probability and statistics. This theorem is ultimately connected with the spectral characterization of as the eigenvalue associated with the Heisenberg uncertainty principle, and the fact that equality holds in the uncertainty principle only for the Gaussian function. Equivalently, is the unique constant making the Gaussian normal distribution equal to its own Fourier transform.[92] Indeed, according to, the "whole business" of establishing the fundamental theorems of Fourier analysis reduces to the Gaussian integral.[90]
The constant appears in the Gauss–Bonnet formula which relates the differential geometry of surfaces to their topology. Specifically, if a compact surface has Gauss curvature K, thenwhere is the Euler characteristic, which is an integer.[93] An example is the surface area of a sphere S of curvature 1 (so that its radius of curvature, which coincides with its radius, is also 1.) The Euler characteristic of a sphere can be computed from its homology groups and is found to be equal to two. Thus we havereproducing the formula for the surface area of a sphere of radius 1.
The constant appears in many other integral formulae in topology, in particular, those involving characteristic classes via the Chern–Weil homomorphism.[94]
One of the key tools in complex analysis is contour integration of a function over a positively oriented (rectifiable) Jordan curve . A form of Cauchy's integral formula states that if a point is interior to, then[95]
Although the curve is not a circle, and hence does not have any obvious connection to the constant, a standard proof of this result uses Morera's theorem, which implies that the integral is invariant under homotopy of the curve, so that it can be deformed to a circle and then integrated explicitly in polar coordinates. More generally, it is true that if a rectifiable closed curve does not contain, then the above integral is times the winding number of the curve.
The general form of Cauchy's integral formula establishes the relationship between the values of a complex analytic function on the Jordan curve and the value of at any interior point of :[96] provided is analytic in the region enclosed by and extends continuously to . Cauchy's integral formula is a special case of the residue theorem, that if is a meromorphic function the region enclosed by and is continuous in a neighbourhood of, thenwhere the sum is of the residues at the poles of .
The constant is ubiquitous in vector calculus and potential theory, for example in Coulomb's law,[97] Gauss's law, Maxwell's equations, and even the Einstein field equations.[98] [99] Perhaps the simplest example of this is the two-dimensional Newtonian potential, representing the potential of a point source at the origin, whose associated field has unit outward flux through any smooth and oriented closed surface enclosing the source:The factor of
1/2\pi
\Phi
R2
\delta
In higher dimensions, factors of are present because of a normalization by the n-dimensional volume of the unit n sphere. For example, in three dimensions, the Newtonian potential is:which has the 2-dimensional volume (i.e., the area) of the unit 2-sphere in the denominator.
The factorial function
n!
\Gamma(n)=(n-1)!
\Gamma(1/2)=\sqrt{\pi}
The gamma function is defined by its Weierstrass product development:[100] where is the Euler–Mascheroni constant. Evaluated at and squared, the equation reduces to the Wallis product formula. The gamma function is also connected to the Riemann zeta function and identities for the functional determinant, in which the constant plays an important role.
The gamma function is used to calculate the volume of the n-dimensional ball of radius r in Euclidean n-dimensional space, and the surface area of its boundary, the (n−1)-dimensional sphere:[101]
Further, it follows from the functional equation that
The gamma function can be used to create a simple approximation to the factorial function for large : which is known as Stirling's approximation. Equivalently,
As a geometrical application of Stirling's approximation, let denote the standard simplex in n-dimensional Euclidean space, and denote the simplex having all of its sides scaled up by a factor of . Then
Ehrhart's volume conjecture is that this is the (optimal) upper bound on the volume of a convex body containing only one lattice point.[102]
The Riemann zeta function is used in many areas of mathematics. When evaluated at it can be written as
Finding a simple solution for this infinite series was a famous problem in mathematics called the Basel problem. Leonhard Euler solved it in 1735 when he showed it was equal to . Euler's result leads to the number theory result that the probability of two random numbers being relatively prime (that is, having no shared factors) is equal to .[103] This probability is based on the observation that the probability that any number is divisible by a prime is (for example, every 7th integer is divisible by 7.) Hence the probability that two numbers are both divisible by this prime is, and the probability that at least one of them is not is . For distinct primes, these divisibility events are mutually independent; so the probability that two numbers are relatively prime is given by a product over all primes:[104]
This probability can be used in conjunction with a random number generator to approximate using a Monte Carlo approach.
The solution to the Basel problem implies that the geometrically derived quantity is connected in a deep way to the distribution of prime numbers. This is a special case of Weil's conjecture on Tamagawa numbers, which asserts the equality of similar such infinite products of arithmetic quantities, localized at each prime p, and a geometrical quantity: the reciprocal of the volume of a certain locally symmetric space. In the case of the Basel problem, it is the hyperbolic 3-manifold .[105]
The zeta function also satisfies Riemann's functional equation, which involves as well as the gamma function:
Furthermore, the derivative of the zeta function satisfies
A consequence is that can be obtained from the functional determinant of the harmonic oscillator. This functional determinant can be computed via a product expansion, and is equivalent to the Wallis product formula.[106] The calculation can be recast in quantum mechanics, specifically the variational approach to the spectrum of the hydrogen atom.[107]
The constant also appears naturally in Fourier series of periodic functions. Periodic functions are functions on the group of fractional parts of real numbers. The Fourier decomposition shows that a complex-valued function on can be written as an infinite linear superposition of unitary characters of . That is, continuous group homomorphisms from to the circle group of unit modulus complex numbers. It is a theorem that every character of is one of the complex exponentials
en(x)=e2\pi
There is a unique character on, up to complex conjugation, that is a group isomorphism. Using the Haar measure on the circle group, the constant is half the magnitude of the Radon–Nikodym derivative of this character. The other characters have derivatives whose magnitudes are positive integral multiples of 2. As a result, the constant is the unique number such that the group T, equipped with its Haar measure, is Pontrjagin dual to the lattice of integral multiples of 2. This is a version of the one-dimensional Poisson summation formula.
The constant is connected in a deep way with the theory of modular forms and theta functions. For example, the Chudnovsky algorithm involves in an essential way the j-invariant of an elliptic curve.
SL2(Z)
SL2(R)
q=e\pi
The constant is the unique constant making the Jacobi theta function an automorphic form, which means that it transforms in a specific way. Certain identities hold for all automorphic forms. An example iswhich implies that transforms as a representation under the discrete Heisenberg group. General modular forms and other theta functions also involve, once again because of the Stone–von Neumann theorem.
The Cauchy distributionis a probability density function. The total probability is equal to one, owing to the integral:
The Shannon entropy of the Cauchy distribution is equal to, which also involves .
The Cauchy distribution plays an important role in potential theory because it is the simplest Furstenberg measure, the classical Poisson kernel associated with a Brownian motion in a half-plane.[109] Conjugate harmonic functions and so also the Hilbert transform are associated with the asymptotics of the Poisson kernel. The Hilbert transform H is the integral transform given by the Cauchy principal value of the singular integral
The constant is the unique (positive) normalizing factor such that H defines a linear complex structure on the Hilbert space of square-integrable real-valued functions on the real line.[110] The Hilbert transform, like the Fourier transform, can be characterized purely in terms of its transformation properties on the Hilbert space : up to a normalization factor, it is the unique bounded linear operator that commutes with positive dilations and anti-commutes with all reflections of the real line.[111] The constant is the unique normalizing factor that makes this transformation unitary.
An occurrence of in the fractal called the Mandelbrot set was discovered by David Boll in 1991.[112] He examined the behaviour of the Mandelbrot set near the "neck" at . When the number of iterations until divergence for the point is multiplied by, the result approaches as approaches zero. The point at the cusp of the large "valley" on the right side of the Mandelbrot set behaves similarly: the number of iterations until divergence multiplied by the square root of tends to .[113]
Let be the set of all twice differentiable real functions
f:R\toR
f''(x)+f(x)=0
t\inR
et:V\toR
f\inV
et(f)=f(t)
et
t\mapsto\keret
R\toP(V)
Although not a physical constant, appears routinely in equations describing fundamental principles of the universe, often because of 's relationship to the circle and to spherical coordinate systems. A simple formula from the field of classical mechanics gives the approximate period of a simple pendulum of length, swinging with a small amplitude (is the earth's gravitational acceleration):[115]
One of the key formulae of quantum mechanics is Heisenberg's uncertainty principle, which shows that the uncertainty in the measurement of a particle's position (Δ) and momentum (Δ) cannot both be arbitrarily small at the same time (where is the Planck constant):[116]
The fact that is approximately equal to 3 plays a role in the relatively long lifetime of orthopositronium. The inverse lifetime to lowest order in the fine-structure constant is[117] where is the mass of the electron.
is present in some structural engineering formulae, such as the buckling formula derived by Euler, which gives the maximum axial load that a long, slender column of length, modulus of elasticity, and area moment of inertia can carry without buckling:[118]
The field of fluid dynamics contains in Stokes' law, which approximates the frictional force exerted on small, spherical objects of radius, moving with velocity in a fluid with dynamic viscosity :[119]
In electromagnetics, the vacuum permeability constant μ0 appears in Maxwell's equations, which describe the properties of electric and magnetic fields and electromagnetic radiation. Before 20 May 2019, it was defined as exactly
See main article: Piphilology. Piphilology is the practice of memorizing large numbers of digits of, and world-records are kept by the Guinness World Records. The record for memorizing digits of, certified by Guinness World Records, is 70,000 digits, recited in India by Rajveer Meena in 9 hours and 27 minutes on 21 March 2015.[120] In 2006, Akira Haraguchi, a retired Japanese engineer, claimed to have recited 100,000 decimal places, but the claim was not verified by Guinness World Records.[121]
One common technique is to memorize a story or poem in which the word lengths represent the digits of : The first word has three letters, the second word has one, the third has four, the fourth has one, the fifth has five, and so on. Such memorization aids are called mnemonics. An early example of a mnemonic for pi, originally devised by English scientist James Jeans, is "How I want a drink, alcoholic of course, after the heavy lectures involving quantum mechanics." When a poem is used, it is sometimes referred to as a piem.[122] Poems for memorizing have been composed in several languages in addition to English. Record-setting memorizers typically do not rely on poems, but instead use methods such as remembering number patterns and the method of loci.[123]
A few authors have used the digits of to establish a new form of constrained writing, where the word lengths are required to represent the digits of . The Cadaeic Cadenza contains the first 3835 digits of in this manner,[124] and the full-length book Not a Wake contains 10,000 words, each representing one digit of .[125]
Perhaps because of the simplicity of its definition and its ubiquitous presence in formulae, has been represented in popular culture more than other mathematical constructs.[126]
In the Palais de la Découverte (a science museum in Paris) there is a circular room known as the pi room. On its wall are inscribed 707 digits of . The digits are large wooden characters attached to the dome-like ceiling. The digits were based on an 1873 calculation by English mathematician William Shanks, which included an error beginning at the 528th digit. The error was detected in 1946 and corrected in 1949.
In Carl Sagan's 1985 novel Contact it is suggested that the creator of the universe buried a message deep within the digits of . This part of the story was omitted from the film adaptation of the novel.[127] The digits of have also been incorporated into the lyrics of the song "Pi" from the 2005 album Aerial by Kate Bush.[128] In the 1967 Star Trek episode "Wolf in the Fold", an out-of-control computer is contained by being instructed to "Compute to the last digit the value of ".[24]
In the United States, Pi Day falls on 14 March (written 3/14 in the US style), and is popular among students.[24] and its digital representation are often used by self-described "math geeks" for inside jokes among mathematically and technologically minded groups. A college cheer variously attributed to the Massachusetts Institute of Technology or the Rensselaer Polytechnic Institute includes "3.14159".[129] [130] Pi Day in 2015 was particularly significant because the date and time 3/14/15 9:26:53 reflected many more digits of pi.[131] [132] In parts of the world where dates are commonly noted in day/month/year format, 22 July represents "Pi Approximation Day", as 22/7 = 3.142857.[133]
Some have proposed replacing by,[134] arguing that, as the number of radians in one turn or the ratio of a circle's circumference to its radius, is more natural than and simplifies many formulae.[135] [136] This use of has not made its way into mainstream mathematics,[137] but since 2010 this has led to people celebrating Two Pi Day or Tau Day on June 28.[138]
In 1897, an amateur mathematician attempted to persuade the Indiana legislature to pass the Indiana Pi Bill, which described a method to square the circle and contained text that implied various incorrect values for, including 3.2. The bill is notorious as an attempt to establish a value of mathematical constant by legislative fiat. The bill was passed by the Indiana House of Representatives, but rejected by the Senate, and thus it did not become a law.[139]
In contemporary internet culture, individuals and organizations frequently pay homage to the number . For instance, the computer scientist Donald Knuth let the version numbers of his program TeX approach . The versions are 3, 3.1, 3.14, and so forth.[140] has been added to several programming languages as a predefined constant.[141] [142]
A Biography of the World's Most Mysterious Number
. registration . Prometheus Books . 2004 . 978-1-59102-200-8 .\overline{\tfrac{16}5-\tfrac4{239}} -\tfrac13\overline{\tfrac{16}{53}-\tfrac4{2393}} +\tfrac15\overline{\tfrac{16}{55}-\tfrac4{2395}} -,\&c.=
Reprinted in Book: Smith, David Eugene . 1929 . A Source Book in Mathematics . McGraw–Hill . William Jones: The First Use of for the Circle Ratio . https://archive.org/details/sourcebookinmath1929smit/page/346/ . 346–347 .
Book: Newton, Isaac . Isaac Newton
. Isaac Newton . 1971 . Whiteside . Derek Thomas . Tom Whiteside . The Mathematical Papers of Isaac Newton . 4, 1674–1684 . Cambridge University Press . 526–653 .
Book: Euler, Leonhard . Leonhard Euler
. Leonhard Euler . 1755 . . §2.2.30 . 318 . Academiae Imperialis Scientiarium Petropolitanae . la . https://archive.org/details/institutiones-calculi-differentialis-cum-eius-vsu-in-analysi-finitorum-ac-doctri/page/318 . E 212.
Euler . Leonhard . Leonhard Euler . 1798 . written 1779 . Investigatio quarundam serierum, quae ad rationem peripheriae circuli ad diametrum vero proxime definiendam maxime sunt accommodatae . Nova Acta Academiae Scientiarum Petropolitinae . 11 . 133–149, 167–168 . E 705 .
Chien-Lih . Hwang . 2004 . 88.38 Some Observations on the Method of Arctangents for the Calculation of . Mathematical Gazette . 88 . 512 . 270–278 . 10.1017/S0025557200175060 . 123532808 .
Chien-Lih . Hwang . 2005 . 89.67 An elementary derivation of Euler's series for the arctangent function . Mathematical Gazette . 89 . 516 . 469–470 . 10.1017/S0025557200178404 . 123395287 .