Vandermonde matrix explained
In linear algebra, a Vandermonde matrix, named after Alexandre-Théophile Vandermonde, is a matrix with the terms of a geometric progression in each row: an
matrix
V=V(x0,x1, … ,xm)=
\begin{bmatrix}
1&x0&
&...&
&x1&
&...&
&x2&
&...&
&\vdots&\vdots&\ddots&\vdots\\
1&xm&
&...&
with entries
, the
jth power of the number
, for all
zero-based indices
and
.
[1] Some authors define the Vandermonde matrix as the
transpose of the above matrix.
[2] The determinant of a square Vandermonde matrix (when
) is called a
Vandermonde determinant or
Vandermonde polynomial. Its value is:
This is non-zero if and only if all
are distinct (no two are equal), making the Vandermonde matrix
invertible.
Applications
The polynomial interpolation problem is to find a polynomial
p(x)=a0+a1x+a2x2+...+anxn
which satisfies
for given data points
. This problem can be reformulated in terms of
linear algebra by means of the Vandermonde matrix, as follows.
computes the values of
at the points
via a matrix multiplication
, where
is the vector of coefficients and
y=(y0,\ldots,ym)=(p(x0),\ldots,p(xm))
is the vector of values (both written as column vectors):
If
and
are distinct, then
V is a square matrix with non-zero determinant, i.e. an
invertible matrix. Thus, given
V and
y, one can find the required
by solving for its coefficients
in the equation
:
[3]
.
That is, the map from coefficients to values of polynomials is a bijective linear mapping with matrix
V, and the interpolation problem has a unique solution. This result is called the unisolvence theorem, and is a special case of the Chinese remainder theorem for polynomials.
In statistics, the equation
means that the Vandermonde matrix is the
design matrix of
polynomial regression.
In numerical analysis, solving the equation
naïvely by Gaussian elimination results in an algorithm with
time complexity O(
n3). Exploiting the structure of the Vandermonde matrix, one can use
Newton's divided differences method[4] (or the Lagrange interpolation formula
[5] [6]) to solve the equation in O(
n2) time, which also gives the
UL factorization of
. The resulting algorithm produces extremely accurate solutions, even if
is
ill-conditioned. (See
polynomial interpolation.)
The Vandermonde determinant is used in the representation theory of the symmetric group.[7]
When the values
belong to a
finite field, the Vandermonde determinant is also called the
Moore determinant, and has properties which are important in the theory of
BCH codes and
Reed–Solomon error correction codes.
The discrete Fourier transform is defined by a specific Vandermonde matrix, the DFT matrix, where the
are chosen to be th roots of unity. The
Fast Fourier transform computes the product of this matrix with a vector in
time.
[8] In the physical theory of the quantum Hall effect, the Vandermonde determinant shows that the Laughlin wavefunction with filling factor 1 is equal to a Slater determinant. This is no longer true for filling factors different from 1 in the fractional quantum Hall effect.
In the geometry of polyhedra, the Vandermonde matrix gives the normalized volume of arbitrary
-faces of
cyclic polytopes. Specifically, if
is a
-face of the cyclic polytope
corresponding to
, then
Determinant
The determinant of a square Vandermonde matrix is called a Vandermonde polynomial or Vandermonde determinant. Its value is the polynomial
which is non-zero if and only if all
are distinct.
The Vandermonde determinant was formerly sometimes called the discriminant, but in current terminology the discriminant of a polynomial
is the
square of the Vandermonde determinant of the roots
. The Vandermonde determinant is an alternating form in the
, meaning that exchanging two
changes the sign, and
thus depends on order for the
. By contrast, the discriminant
does not depend on any order, so that
Galois theory implies that the discriminant is a polynomial function of the coefficients of
.
The determinant formula is proved below in three ways. The first uses polynomial properties, especially the unique factorization property of multivariate polynomials. Although conceptually simple, it involves non-elementary concepts of abstract algebra. The second proof is based on the linear algebra concepts of change of basis in a vector space and the determinant of a linear map. In the process, it computes the LU decomposition of the Vandermonde matrix. The third proof is more elementary but more complicated, using only elementary row and column operations.
First proof: Polynomial properties
The first proof relies on properties of polynomials.
By the Leibniz formula,
is a polynomial in the
, with integer coefficients. All entries of the
-th column have total degree
. Thus, again by the Leibniz formula, all terms of the determinant have total degree
(that is, the determinant is a
homogeneous polynomial of this degree).
If, for
, one substitutes
for
, one gets a matrix with two equal rows, which has thus a zero determinant. Thus, considering the determinant as univariate in
the
factor theorem implies that
is a divisor of
It thus follows that for all
and
,
is a divisor of
This will now be strengthened to show that the product of all those divisors of
is a divisor of
Indeed, let
be a polynomial with
as a factor, then
for some polynomial
If
is another factor of
then
becomes zero after the substitution of
for
If
the factor
becomes zero after this substitution, since the factor
remains nonzero. So, by the factor theorem,
divides
and
divides
Iterating this process by starting from
one gets that
is divisible by the product of all
with
that is
\det(V)=Q\prod0\le(xj-xi),
where
is a polynomial. As the product of all
and
have the same degree
, the polynomial
is, in fact, a constant. This constant is one, because the product of the diagonal entries of
is
, which is also the
monomial that is obtained by taking the first term of all factors in
This proves that
and finishes the proof.
\det(V)=\prod0\le(xj-xi).
Second proof: linear maps
Let be a field containing all
and
the
vector space of the polynomials of degree less than or equal to with coefficients in . Let
be the
linear map defined by
p(x)\mapsto(p(x0),p(x1),\ldots,p(xn))
.
The Vandermonde matrix is the matrix of
with respect to the
canonical bases of
and
Changing the basis of
amounts to multiplying the Vandermonde matrix by a change-of-basis matrix (from the right). This does not change the determinant, if the determinant of is .
The polynomials
,
,
, …,
are
monic of respective degrees 0, 1, …, . Their matrix on the
monomial basis is an upper-triangular matrix (if the monomials are ordered in increasing degrees), with all diagonal entries equal to one. This matrix is thus a change-of-basis matrix of determinant one. The matrix of
on this new basis is
\begin{bmatrix}
1&0&0&\ldots&0\
1&x1-x0&0&\ldots&0\
1&x2-x0&(x2-x0)(x2-x1)&\ldots&0\
\vdots&\vdots&\vdots&\ddots&\vdots\
1&xn-x0&(xn-x0)(xn-x1)&\ldots&(xn-x0)(xn-x1) … (xn-xn-1)
\end{bmatrix}
.Thus Vandermonde determinant equals the determinant of this matrix, which is the product of its diagonal entries.
This proves the desired equality. Moreover, one gets the LU decomposition of as
.
Third proof: row and column operations
The third proof is based on the fact that if one adds to a column of a matrix the product by a scalar of another column then the determinant remains unchanged.
So, by subtracting to each column – except the first one – the preceding column multiplied by
, the determinant is not changed. (These subtractions must be done by starting from last columns, for subtracting a column that has not yet been changed). This gives the matrix
V=\begin{bmatrix}
1&0&0&0& … &0\\
1&x1-x0&x1(x1-x0)&x
0)& … &x
(x1-x0)\\
1&x2-x0&x2(x2-x0)&x
0)& … &x
(x2-x0)\\
\vdots&\vdots&\vdots&\vdots&\ddots&\vdots\\
1&xn-x0&xn(xn-x0)&x
0)& … &x
(xn-x0)\\
\end{bmatrix}
Applying the Laplace expansion formula along the first row, we obtain
, with
B=\begin{bmatrix}
x1-x0&x1(x1-x0)&x
0)& … &x
(x1-x0)\\
x2-x0&x2(x2-x0)&x
0)& … &x
(x2-x0)\\
\vdots&\vdots&\vdots&\ddots&\vdots\\
xn-x0&xn(xn-x0)&x
0)& … &x
(xn-x0)\\
\end{bmatrix}
As all the entries in the
-th row of
have a factor of
, one can take these factors out and obtain
\det(V)=(x1-x0)(x2-x0) … (xn-x0)\begin{vmatrix}
1&x1&x
\\
1&x2&x
\\
\vdots&\vdots&\vdots&\ddots&\vdots\\
1&xn&x
\\
\end{vmatrix}=\prod1<i\leq(xi-x0)\det(V')
,where
is a Vandermonde matrix in
. Iterating this process on this smaller Vandermonde matrix, one eventually gets the desired expression of
as the product of all
such that
.
Rank of the Vandermonde matrix
- An rectangular Vandermonde matrix such that has rank if and only if all are distinct.
- An rectangular Vandermonde matrix such that has rank if and only if there are of the that are distinct.
- A square Vandermonde matrix is invertible if and only if the are distinct. An explicit formula for the inverse is known (see below).[9] [10] [11]
Inverse Vandermonde matrix
As explained above in Applications, the polynomial interpolation problem for
p(x)=a0+a1x+a2x2+...+anxn
satisfying
is equivalent to the matrix equation
, which has the unique solution
. There are other known formulas which solve the interpolation problem, which must be equivalent to the unique
, so they must give explicit formulas for the inverse matrix
. In particular,
Lagrange interpolation shows that the columns of the inverse matrix
are the coefficients of the Lagrange polynomials
Lj(x)=L0j+L1jx+ … +Lnjxn=\prod0\leq
=
,
where
. This is easily demonstrated: the polynomials clearly satisfy
for
while
, so we may compute the product
, the identity matrix.
Confluent Vandermonde matrices
As described before, a Vandermonde matrix describes the linear algebra interpolation problem of finding the coefficients of a polynomial
of degree
based on the values
, where
are
distinct points. If
are not distinct, then this problem does not have a unique solution (and the corresponding Vandermonde matrix is singular). However, if we specify the values of the derivatives at the repeated points, then the problem can have a unique solution. For example, the problem
\begin{cases}
p(0)=y1\\
p'(0)=y2\\
p(1)=y3\end{cases}
where
, has a unique solution for all
with
. In general, suppose that
are (not necessarily distinct) numbers, and suppose for simplicity that equal values are adjacent:
x1= … =
,
= … =
,
\ldots,
= … =
where
and
are distinct. Then the corresponding interpolation problem is
\begin{cases}
)=y1,&
)=y2,&\ldots,&
)=
,\\
)=
,&
,&\ldots,&
)=
,\\
\vdots&&& \vdots\\
)=
,&
)=
,&\ldots,&
)=
.
\end{cases}
The corresponding matrix for this problem is called a confluent Vandermonde matrix, given as follows. If
, then
for a unique
(denoting
). We let
Vi,j=\begin{cases}
0&ifj<i-m\ell,\\[6pt]
\dfrac{(j-1)!}{(j-(i-m\ell))!}
&ifj\geqi-m\ell.
\end{cases}
This generalization of the Vandermonde matrix makes it non-singular, so that there exists a unique solution to the system of equations, and it possesses most of the other properties of the Vandermonde matrix. Its rows are derivatives (of some order) of the original Vandermonde rows.
Another way to derive this formula is by taking a limit of the Vandermonde matrix as the
's approach each other. For example, to get the case of
, take subtract the first row from second in the original Vandermonde matrix, and let
: this yields the corresponding row in the confluent Vandermonde matrix. This derives the generalized interpolation problem with given values and derivatives as a limit of the original case with distinct points: giving
is similar to giving
for small
. Geometers have studied the problem of tracking confluent points along their tangent lines, known as
compacitification of configuration space.
See also
Further reading
Notes and References
- Roger A. Horn and Charles R. Johnson (1991), Topics in matrix analysis, Cambridge University Press. See Section 6.1.
- Book: Golub . Gene H. . Matrix Computations . Van Loan . Charles F. . The Johns Hopkins University Press . 2013 . 978-1-4214-0859-0 . 4th . 203–207.
- François Viète (1540-1603), Vieta's formulas, https://en.wikipedia.org/wiki/Vieta%27s_formulas
- Björck . Å. . Pereyra . V. . 1970 . Solution of Vandermonde Systems of Equations . . 24 . 112 . 893–903 . 10.1090/S0025-5718-1970-0290541-1. 122006253 . free .
- Book: Press . WH . Numerical Recipes: The Art of Scientific Computing . Teukolsky . SA . Vetterling . WT . Flannery . BP . Cambridge University Press . 2007 . 978-0-521-88068-8 . 3rd . New York . Section 2.8.1. Vandermonde Matrices . http://apps.nrbook.com/empanel/index.html?pg=94.
- Inverse of Vandermonde Matrix (2018), https://proofwiki.org/wiki/Inverse_of_Vandermonde_Matrix
- Lecture 4 reviews the representation theory of symmetric groups, including the role of the Vandermonde determinant.
- Gauthier, J. "Fast Multipoint Evaluation On n Arbitrary Points." Simon Fraser University, Tech. Rep (2017).
- Book: Turner
, L. Richard
. Inverse of the Vandermonde matrix with applications. August 1966.
- 65. 2. 95–100. Macon. N.. A. Spitzbart . Inverses of Vandermonde Matrices. The American Mathematical Monthly. February 1958. 2308881. 10.2307/2308881.
- Web site: Inverse of Vandermonde Matrix. 2018.