In mathematics, the resultant of two polynomials is a polynomial expression of their coefficients that is equal to zero if and only if the polynomials have a common root (possibly in a field extension), or, equivalently, a common factor (over their field of coefficients). In some older texts, the resultant is also called the eliminant.
The resultant is widely used in number theory, either directly or through the discriminant, which is essentially the resultant of a polynomial and its derivative. The resultant of two polynomials with rational or polynomial coefficients may be computed efficiently on a computer. It is a basic tool of computer algebra, and is a built-in function of most computer algebra systems. It is used, among others, for cylindrical algebraic decomposition, integration of rational functions and drawing of curves defined by a bivariate polynomial equation.
The resultant of n homogeneous polynomials in n variables (also called multivariate resultant, or Macaulay's resultant for distinguishing it from the usual resultant) is a generalization, introduced by Macaulay, of the usual resultant. It is, with Gröbner bases, one of the main tools of elimination theory.
The resultant of two univariate polynomials and is commonly denoted
\operatorname{res}(A,B)
\operatorname{Res}(A,B).
In many applications of the resultant, the polynomials depend on several indeterminates and may be considered as univariate polynomials in one of their indeterminates, with polynomials in the other indeterminates as coefficients. In this case, the indeterminate that is selected for defining and computing the resultant is indicated as a subscript:
\operatorname{res}x(A,B)
\operatorname{Res}x(A,B).
The degrees of the polynomials are used in the definition of the resultant. However, a polynomial of degree may also be considered as a polynomial of higher degree where the leading coefficients are zero. If such a higher degree is used for the resultant, it is usually indicated as a subscript or a superscript, such as
\operatorname{res}d,e(A,B)
d,e | |
\operatorname{res} | |
x |
(A,B).
The resultant of two univariate polynomials over a field or over a commutative ring is commonly defined as the determinant of their Sylvester matrix. More precisely, let andbe nonzero polynomials of degrees and respectively. Let us denote by
l{P}i
The resultant of and is thus the determinantwhich has columns of and columns of (the fact that the first column of 's and the first column of 's have the same length, that is, is here only for simplifying the display of the determinant).For instance, taking and we get
If the coefficients of the polynomials belong to an integral domain, then where
λ1,...,λd
\mu1,...,\mue
In this section and its subsections, and are two polynomials in of respective degrees and, and their resultant is denoted
\operatorname{res}(A,B).
The following properties hold for the resultant of two polynomials with coefficients in a commutative ring . If is a field or more generally an integral domain, the resultant is the unique function of the coefficients of two polynomials that satisfies these properties.
\operatorname{res}R(A,B)=\operatorname{res}S(A,B).
A=a0
\operatorname{res}(A,B)=
e. | |
a | |
0 |
\operatorname{res}(A,B)=
d. | |
b | |
0 |
\operatorname{res}(x+a1,x+b1)=b1-a1
\operatorname{res}(B,A)=(-1)de\operatorname{res}(A,B)
\operatorname{res}(AB,C)=\operatorname{res}(A,C)\operatorname{res}(B,C)
\operatorname{res}(A,B)=AP+BQ.
Let and be two polynomials of respective degrees and with coefficients in a commutative ring, and
\varphi\colonR\toS
\varphi
\varphi
R[x]\toS[x]
\varphi.
\varphi
\deg(\varphi(A))=d
\deg(\varphi(B))=e
\deg(\varphi(A))<d
\deg(\varphi(B))<e,
\deg(\varphi(A))=d
\deg(\varphi(B))=f<e,
a0
\deg(\varphi(A))=f<d
\deg(\varphi(B))=e,
b0
These properties are easily deduced from the definition of the resultant as a determinant. They are mainly used in two situations. For computing a resultant of polynomials with integer coefficients, it is generally faster to compute it modulo several primes and to retrieve the desired resultant with Chinese remainder theorem. When is a polynomial ring in other indeterminates, and is the ring obtained by specializing to numerical values some or all indeterminates of, these properties may be restated as if the degrees are preserved by the specialization, the resultant of the specialization of two polynomials is the specialization of the resultant. This property is fundamental, for example, for cylindrical algebraic decomposition.
\operatorname{res}(A(x+a),B(x+a))=\operatorname{res}(A(x),B(x))
\operatorname{res}(A(ax),B(ax))=ade\operatorname{res}(A(x),B(x))
Ar(x)=xdA(1/x)
Br(x)=xeB(1/x)
This means that the property of the resultant being zero is invariant under linear and projective changes of the variable.
These properties imply that in the Euclidean algorithm for polynomials, and all its variants (pseudo-remainder sequences), the resultant of two successive remainders (or pseudo-remainders) differs from the resultant of the initial polynomials by a factor which is easy to compute. Conversely, this allows one to deduce the resultant of the initial polynomials from the value of the last remainder or pseudo-remainder. This is the starting idea of the subresultant-pseudo-remainder-sequence algorithm, which uses the above formulae for getting subresultant polynomials as pseudo-remainders, and the resultant as the last nonzero pseudo-remainder (provided that the resultant is not zero). This algorithm works for polynomials over the integers or, more generally, over an integral domain, without any division other than exact divisions (that is, without involving fractions). It involves
O(de)
O((d+e)3)
In this section, we consider two polynomialsandwhose coefficients are distinct indeterminates. Letbe the polynomial ring over the integers defined by these indeterminates.The resultant
\operatorname{res}(A,B)
\operatorname{res}(A,B)
I
R[x]
I\capR
\operatorname{res}(A,B)
The generic resultant for the degrees and is homogeneous in various ways. More precisely:
a0,\ldots,ad.
b0,\ldots,be.
ai
bj.
ai
bi
d,e | |
\operatorname{res} | |
x |
(P,Q)
Let
I=\langleA,B\rangle
R[x],
R=k[y1,\ldots,yn]
\operatorname{res}x(A,B)\inI\capR
I\capR
R\operatorname{res}x(A,B)
I\capR
\operatorname{res}x(A,B).
I\capR
R\operatorname{res}x(A,B).
I\capR
\operatorname{res}x(A,B).
\operatorname{res}x(A,B)
I\capR.
The first assertion is a basic property of the resultant. The other assertions are immediate corollaries of the second one, which can be proved as follows.
As at least one of and is monic, a tuple
(\beta1,\ldots,\betan)
\operatorname{res}x(A,B)
\alpha
(\beta1,\ldots,\betan,\alpha)
I\capR.
(\beta1,\ldots,\betan)
I\capR,
\alpha
(\beta1,\ldots,\betan,\alpha)
I\capR
R\operatorname{res}x(A,B)
Theoretically, the resultant could be computed by using the formula expressing it as a product of roots differences. However, as the roots may generally not be computed exactly, such an algorithm would be inefficient and numerically unstable. As the resultant is a symmetric function of the roots of each polynomial, it could also be computed by using the fundamental theorem of symmetric polynomials, but this would be highly inefficient.
As the resultant is the determinant of the Sylvester matrix (and of the Bézout matrix), it may be computed by using any algorithm for computing determinants. This needs
O(n3)
It follows from that the computation of a resultant is strongly related to the Euclidean algorithm for polynomials. This shows that the computation of the resultant of two polynomials of degrees and may be done in
O(de)
However, when the coefficients are integers, rational numbers or polynomials, these arithmetic operations imply a number of GCD computations of coefficients which is of the same order and make the algorithm inefficient.The subresultant pseudo-remainder sequences were introduced to solve this problem and avoid any fraction and any GCD computation of coefficients. A more efficient algorithm is obtained by using the good behavior of the resultant under a ring homomorphism on the coefficients: to compute a resultant of two polynomials with integer coefficients, one computes their resultants modulo sufficiently many prime numbers and then reconstructs the result with the Chinese remainder theorem.
The use of fast multiplication of integers and polynomials allows algorithms for resultants and greatest common divisors that have a better time complexity, which is of the order of the complexity of the multiplication, multiplied by the logarithm of the size of the input (
log(s(d+e)),
Resultants were introduced for solving systems of polynomial equations and provide the oldest proof that there exist algorithms for solving such systems. These are primarily intended for systems of two equations in two unknowns, but also allow solving general systems.
Consider the system of two polynomial equationswhere and are polynomials of respective total degrees and . Then
d,e | |
R=\operatorname{res} | |
y |
(P,Q)
\alpha
\beta
P(\alpha,\beta)=Q(\alpha,\beta)=0
\deg(P(\alpha,y))<d
\deg(Q(\alpha,y))<e
x=\alpha
Therefore, solutions to the system are obtained by computing the roots of, and for each root
\alpha,
P(\alpha,y),
Q(\alpha,y),
\operatorname{res}x(P,Q).
Bézout's theorem results from the value of
\deg\left(\operatorname{res}y(P,Q)\right)\lede
At first glance, it seems that resultants may be applied to a general polynomial system of equationsby computing the resultants of every pair
(Pi,Pj)
xn
A method, introduced at the end of the 19th century, works as follows: introduce new indeterminates
U2,\ldots,Uk
U2,\ldots,Uk
x1,\ldots,xn-1,
\alpha1,\ldots,\alphan-1
Pi(\alpha1,\ldots,\alphan-1,xn)
To get a correct algorithm two complements have to be added to the method. Firstly, at each step, a linear change of variable may be needed in order that the degrees of the polynomials in the last variable are the same as their total degree. Secondly, if, at any step, the resultant is zero, this means that the polynomials have a common factor and that the solutions split in two components: one where the common factor is zero, and the other which is obtained by factoring out this common factor before continuing.
This algorithm is very complicated and has a huge time complexity. Therefore, its interest is mainly historical.
The discriminant of a polynomial, which is a fundamental tool in number theory, is
-1 | |
a | |
0 |
(-1)n(n\operatorname{res}x(f(x),f'(x))
a0
f(x)
n
If
\alpha
\beta
P(\alpha)=Q(\beta)=0
\gamma=\alpha+\beta
\operatorname{res}x(P(x),Q(z-x)),
\tau=\alpha\beta
nQ(z/x)) | |
\operatorname{res} | |
x(P(x),x |
n
Q(y)
1/\beta
ynQ(1/y)=0
Let
K(\alpha)
\alpha,
P(x)
\beta\inK(\alpha)
\beta=Q(\alpha),
Q
\beta
\operatorname{res}x(P(x),z-Q(x)),
\beta.
Given two plane algebraic curves defined as the zeros of the polynomials and, the resultant allows the computation of their intersection. More precisely, the roots of
\operatorname{res}y(P,Q)
\operatorname{res}x(P,Q)
A rational plane curve may be defined by a parametric equationwhere, and are polynomials. An implicit equation of the curve is given byThe degree of this curve is the highest degree of, and, which is equal to the total degree of the resultant.
In symbolic integration, for computing the antiderivative of a rational fraction, one uses partial fraction decomposition for decomposing the integral into a "rational part", which is a sum of rational fractions whose antiprimitives are rational fractions, and a "logarithmic part" which is a sum of rational fractions of the formwhere is a square-free polynomial and is a polynomial of lower degree than . The antiderivative of such a function involves necessarily logarithms, and generally algebraic numbers (the roots of). In fact, the antiderivative is where the sum runs over all complex roots of .
The number of algebraic numbers involved by this expression is generally equal to the degree of, but it occurs frequently that an expression with less algebraic numbers may be computed. The Lazard–Rioboo–Trager method produces an expression, where the number of algebraic numbers is minimal, without any computation with algebraic numbers.
Let be the square-free factorization of the resultant which appears on the right. Trager proved that the antiderivative is where the internal sums run over the roots of the
Si
Si=1
Ti(r,x)
Ti(r,x)
rQ'(x)-P(x)
Q(x).
All preceding applications, and many others, show that the resultant is a fundamental tool in computer algebra. In fact most computer algebra systems include an efficient implementation of the computation of resultants.
The resultant is also defined for two homogeneous polynomial in two indeterminates. Given two homogeneous polynomials and of respective total degrees and, their homogeneous resultant is the determinant of the matrix over the monomial basis of the linear mapwhere runs over the bivariate homogeneous polynomials of degree, and runs over the homogeneous polynomials of degree . In other words, the homogeneous resultant of and is the resultant of and when they are considered as polynomials of degree and (their degree in may be lower than their total degree):(The capitalization of "Res" is used here for distinguishing the two resultants, although there is no standard rule for the capitalization of the abbreviation).
The homogeneous resultant has essentially the same properties as the usual resultant, with essentially two differences: instead of polynomial roots, one considers zeros in the projective line, and the degree of a polynomial may not change under a ring homomorphism.That is:
\varphi\colonR\toS
\varphi
Any property of the usual resultant may similarly extended to the homogeneous resultant, and the resulting property is either very similar or simpler than the corresponding property of the usual resultant.
Macaulay's resultant, named after Francis Sowerby Macaulay, also called the multivariate resultant, or the multipolynomial resultant,[1] is a generalization of the homogeneous resultant to homogeneous polynomials in indeterminates. Macaulay's resultant is a polynomial in the coefficients of these homogeneous polynomials that vanishes if and only if the polynomials have a common non-zero solution in an algebraically closed field containing the coefficients, or, equivalently, if the hyper surfaces defined by the polynomials have a common zero in the dimensional projective space. The multivariate resultant is, with Gröbner bases, one of the main tools of effective elimination theory (elimination theory on computers).
Like the homogeneous resultant, Macaulay's may be defined with determinants, and thus behaves well under ring homomorphisms. However, it cannot be defined by a single determinant. It follows that it is easier to define it first on generic polynomials.
A homogeneous polynomial of degree in variables may have up to coefficients; it is said to be generic, if these coefficients are distinct indeterminates.
Let
P1,\ldots,Pn
d1,...,dn.
P1,\ldots,Pn
C[x1,\ldots,xn],
The Macaulay degree is the integer
D=d1+ … +dn-n+1,
Qi
D-di,
If, the Macaulay matrix is the Sylvester matrix, and is a square matrix, but this is no longer true for . Thus, instead of considering the determinant, one considers all the maximal minors, that is the determinants of the square submatrices that have as many rows as the Macaulay matrix. Macaulay proved that the -ideal generated by these principal minors is a principal ideal, which is generated by the greatest common divisor of these minors. As one is working with polynomials with integer coefficients, this greatest common divisor is defined up to its sign. The generic Macaulay resultant is the greatest common divisor which becomes, when, for each, zero is substituted for all coefficients of
Pi,
di | |
x | |
i |
,
B/di
Pi,
B=d1 … dn
x1,...,xn
C[x1,...,xn]
P1,...,Pn.
From now on, we consider that the homogeneous polynomials
P1,\ldots,Pn
d1,\ldots,dn
k[x1,...,xn].
Pi.
The main property of the resultant is that it is zero if and only if
P1,\ldots,Pn
The "only if" part of this theorem results from the last property of the preceding paragraph, and is an effective version of Projective Nullstellensatz: If the resultant is nonzero, thenwhere
D=d1+ … +dn-n+1
\langlex1,\ldots,xn\rangle
P1,\ldots,Pn
x1,\ldots,xn.
As the computation of a resultant may be reduced to computing determinants and polynomial greatest common divisors, there are algorithms for computing resultants in a finite number of steps.
However, the generic resultant is a polynomial of very high degree (exponential in) depending on a huge number of indeterminates. It follows that, except for very small and very small degrees of input polynomials, the generic resultant is, in practice, impossible to compute, even with modern computers. Moreover, the number of monomials of the generic resultant is so high, that, if it would be computable, the result could not be stored on available memory devices, even for rather small values of and of the degrees of the input polynomials.
Therefore, computing the resultant makes sense only for polynomials whose coefficients belong to a field or are polynomials in few indeterminates over a field.
dO(n),
Another case where the computation of the resultant may provide useful information is when the coefficients of the input polynomials are polynomials in a small number of indeterminates, often called parameters. In this case, the resultant, if not zero, defines a hypersurface in the parameter space. A point belongs to this hyper surface, if and only if there are values of
x1,\ldots,xn
x1,\ldots,xn
Macaulay's resultant provides a method, called "U-resultant" by Macaulay, for solving systems of polynomial equations.
Given homogeneous polynomials
P1,\ldots,Pn-1,
d1,\ldots,dn-1,
x1,\ldots,xn,
P1,\ldots,Pn-1,Pn,
u1,\ldots,un.
ui
Ui
The U-resultant is a homogeneous polynomial in
k[u1,\ldots,un].
P1,\ldots,Pn-1
d1 … dn-1.
\alpha1u1+\ldots+\alphanun
\alpha1,\ldots,\alphan
P1,\ldots,Pn-1.
Pi
The U-resultant as defined by Macaulay requires the number of homogeneous polynomials in the system of equations to be
n-1
n
n-1
Let
P1,\ldots,Pk
x1,\ldots,xn,
d1,\ldots,dk,
d1\ged2\ge … \gedk.
di=1
D=d1+ … +dn-n+1.
Let
u1,\ldots,un
Pk+1=u1x1+ … +unxn.
x1,\ldots,xn,
Qi
D-di
Reducing the Macaulay matrix by a variant of Gaussian elimination, one obtains a square matrix of linear forms in
u1,\ldots,un.
P1,\ldots,Pk
P1,\ldots,Pk
P1,\ldots,Pk,
The number of rows of the Macaulay matrix is less than
(ed)n,
Pi.
dO(n).