Row echelon form explained

In linear algebra, a matrix is in row echelon form if it can be obtained as the result of Gaussian elimination. Every matrix can be put in row echelon form by applying a sequence of elementary row operations. The term echelon comes from the French échelon ("level" or step of a ladder), and refers to the fact that the nonzero entries of a matrix in row echelon form look like an inverted staircase.

For square matrices, an upper triangular matrix with nonzero entries on the diagonal is in row echelon form, and a matrix in row echelon form is (weakly) upper triangular. Thus, the row echelon form can be viewed as a generalization of upper triangular form for rectangular matrices.

A matrix is in reduced row echelon form if it is in row echelon form, with the additional property that the first nonzero entry of each row is equal to

1

and is the only nonzero entry of its column. The reduced row echelon form of a matrix is unique and does not depend on the sequence of elementary row operations used to obtain it. The variant of Gaussian elimination that transforms a matrix to reduced row echelon form is sometimes called Gauss–Jordan elimination.

A matrix is in column echelon form if its transpose is in row echelon form. Since all properties of column echelon forms can therefore immediately be deduced from the corresponding properties of row echelon forms, only row echelon forms are considered in the remainder of the article.

(General) row echelon form

A matrix is in row echelon form if

Some texts add the condition that the leading coefficient must be 1[3] while others require this only in reduced row echelon form.

These two conditions imply that all entries in a column below a leading coefficient are zeros.

The following is an example of a

4 x 5

matrix in row echelon form, but not in reduced row echelon form (see below):

\left[\begin{array}{ccccc} 1&a0&a1&a2&a3\\ 0&0&2&a4&a5\\ 0&0&0&1&a6\\ 0&0&0&0&0 \end{array}\right]

Many properties of matrices may be easily deduced from their row echelon form, such as the rank and the kernel.

Reduced row echelon form

A matrix is in reduced row echelon form (also called row canonical form) if it satisfies the following conditions:

If the first two conditions are verified, the last condition is equivalent to:

While a matrix may have several echelon forms, its reduced echelon form is unique.

Given a matrix in reduced row echelon form, if one permutes the columns in order to have the leading of the th row in the th column, one gets a matrix of the form

\begin{pmatrix} I&X\ 0&0 \end{pmatrix},

where is the identity matrix of dimension

j

equal to the rank of the entire matrix, is a matrix with

j

rows and

n-j

columns, and the two 's are zero matrices of appropriate size. Since a permutation of columns is not a row operation, the resulting matrix is inequivalent under elementary row operations. In the Gaussian elimination method, this corresponds to a permutation of the unknowns in the original linear system that allows a linear parametrization of the row space, in which the first

j

coefficients are unconstrained, and the remaining

n-j

are determined as linear combinations of these.

Systems of linear equations

A system of linear equations is said to be in row echelon form if its augmented matrix is in row echelon form. Similarly, a system of linear equations is said to be in reduced row echelon form or in canonical form if its augmented matrix is in reduced row echelon form.

The canonical form may be viewed as an explicit solution of the linear system. In fact, the system is inconsistent if and only if one of the equations of the canonical form is reduced to 0 = 1; that is if there is a leading in the column of the constant terms[4] Otherwise, regrouping in the right hand side all the terms of the equations but the leading ones, expresses the variables corresponding to the pivots as constants or linear functions of the other variables, if any.

Transformation to row echelon form

See main article: Gaussian elimination.

Gaussian elimination is the main algorithm for transforming every matrix into a matrix in row echelon form. A variant, sometimes called Gauss–Jordan elimination produces a reduced row echelon form. Both consist of a finite sequence of elementary row operations; the number of required elementary row operations is at most for an -by- matrix.[5] For a given matrix, despite the row echelon form not being unique, all row echelon forms, including the reduced row echelon form, have the same number of zero rows and the pivots are located in the same positions.

This is an example of a matrix in reduced row echelon form, which shows that the left part of the matrix is not always an identity matrix:

\left[\begin{array}{ccccc} 1&0&a1&0&b1\\ 0&1&a2&0&b2\\ 0&0&0&1&b3 \end{array}\right]

For a matrix with integer coefficients, the Hermite normal form is a row echelon form that can be calculated without introducing any denominator, by using Euclidean division or Bézout's identity. The reduced echelon form of a matrix with integer entries generally contains non-integer entries, because of the need of dividing by its leading coefficient each row of the echelon form.

The non-uniqueness of the row echelon form of a matrix follows from the fact that some elementary row operations transform a matrix in row echelon form into another (equivalent) matrix that is also in row echelon form. These elementary row operations include the multiplication of a row by a nonzero scalar and the addition of a scalar multiple of a row to one of the rows above it. For example:

\begin{bmatrix}1&3&-1\ 0&1&7\\end{bmatrix}\xrightarrow{addrow2torow1

}\begin 1 & 4 & 6 \\ 0 & 1 & 7 \\ \end. In this example, the unique reduced row echelon form can be obtained by subtracting three times the second row from the first row :

\begin{bmatrix}1&3&-1\ 0&1&7\\end{bmatrix}\xrightarrow{subtract3 x (row2)fromrow1

}\begin 1 & 0 & -22 \\ 0 & 1 & 7 \\ \end.

Affine spaces of reduced echelon forms

In this section and the following one, we denote the location of the columns containing the leading entries of the successive rows of a

k x n

matrix

A

in reduced row echelon form (the pivots) as

(L1,...,Lj)

, with

0<L1<Lj\len,

where

j\lek

is the dimension of the row space of the matrix. The data

(k,n,L1,\ldots,Lj)

will be called the shape of

A

, which has leading non-zero entries

\{

A
i,Li

=1\}i=1,

, the entries in the column

Li

above and below it vanish, and so do all those to the left of it within the same row, as well as all entries in the

i

th row for

i>j

:
\begin{align} A
i,Li

=1    &fori=1,...,

j,\\ A
l,Li

=0    &forl\nei,\Ai,=0    &forl<Li,\\ Ai,=0    &fori>j \end{align}.

Since all other entries are arbitrary elements of the base field

K

, the set

A(k,n,L1,\ldots,Lj)

of all reduced echelon form matrices with shape

(k,n,L1,\ldots,Lj)

is a -affine space of dimension[6] [7]

dim(A(k,n,L1,...,Lj))=nj-

1
2

j(j-1)-

j
\sum
i=1

Li.

To see this, note that, of the

nj

possible matrix entries within the first

j

rows,

j2

are determined as

0

's and

1

's because they are in the columns

(L1,...,Lj)

containing the pivots. A further
j(L
\sum
i-1)
are also required to be

0

, because they are to the left of the pivots, but of these,
j-1
\sum
i=0

i=

1
2

j(j-1)

are also in the columns

(L1,...,Lj)

. Therefore, the total number of entries that are not fixed to be equal to

0

or

1

is

nj-j2+

1
2

j(j-1)-

j
\sum
i=1

Li+j =nj-

1
2

j(j-1)-

j
\sum
i=1

Li.

Maximal rank: Schubert cells

The row echelon form can be used to give a concrete description of the Schubert cells associated with the Grassmannian of

k

-dimensional subspaces of a vector space.

If

j=k\len

, the matrices

A\inA(k,n,L1,...,Lk)

are of maximal rank

k

, and determine

k

-dimensional subspaces

w\subsetV

of the free

K

-module

V:=Kn

, as the span

w=span\{W1,...,Wk\}

of linear combinations

Wi:=

n
\sum
l=1

Ailel,i=1,...,k

of the elementary basis vectors

(e1,...,en)

,with coefficients equal to the row vectors. In this case, the affine space

A(k,n,L1,...,Lk)

is the Schubert cell[6] [7]

Xλ(l{V})

of the Grassmannian

Grk(V)

, consisting of

k

-dimensional subspaces of

V

corresponding to the integer partition

λ=(λ1\ge\geλk\ge0)

with parts equal to

λi:=n-k-Li+i,1\lej\lek,

relative to the complete flag

l{V}=(V1\subsetV2\subsetVn=V),

where

Vi=span\{e1,...,ei\},i=1,...n.

This means that

Xλ(l{V})\subsetGrk(V)

consists of those

k

-dimensional subspaces

w\subsetV

whose intersections with the subspaces

\{Vj\}j=1,

have dimensions

dim(w\capVj)=i,forn-ki+i\lej\len-ki+1+i,i=1,...,k.

Its dimension is then equal to the weight

|λ|=

k
\sum
i=1

λi

of the partition[6]

\dim({Xλ(l{V})})=|λ|.

An equivalent, but simpler characterization of the Schubert cell

Xλ(l{V})

may be given in terms of the dual complete flag

\tilde{l{V}}=(\tilde{V}1\subset\tilde{V}2\subset\tilde{V}n=V),

where

\tilde{V}i=span\{en,...,en-i+1\},i=1,...n.

Then

Xλ(l{V})\subsetGrk(V)

consists of those

k

-dimensional subspaces

w\subsetV

that have a basis

(\tilde{W}1,...,\tilde{W}k)

consisting of elements

\tilde{W}i\in

\tilde{V}
n-Li+1
=\tilde{V}
ki-i+1

,i=1,...,k

of the subspaces

\{\tilde{V}
ki-i+1

\}i=1,

which, relative to the standard basis, are the row vectors

(Wk,...,W1)

of the row echelon form, written in reverse order.

References

External links

Notes and References

  1. Phrased in terms of each individual zero row in :"A matrix is said to be in row echelon form ... (iii) If there are rows whose entries are all zero, they are below the rows having nonzero entries."
  2. "A matrix is said to be in row echelon form ... (ii) If row does not consist entirely of zeros, the number of leading zero entries in row

    k+1

    is greater than the number of leading zero entries in row ."
  3. See, for instance, the first clause of the definition of row echelon form in : "A matrix is said to be in row echelon form (i) If the first nonzero entry in each nonzero row is 1."
  4. Book: Cheney, Ward. Linear Algebra: Theory and Applications. Kincaid. David R.. 2010-12-29. Jones & Bartlett Publishers. 9781449613525. 47–50. en.
  5. Book: Anton, Howard. Elementary Linear Algebra: Applications Version, 11th Edition. Rorres. Chris. 2013-10-23. Wiley Global Education. 9781118879160. 21. en.
  6. Book: Fulton, William . Young Tableaux. With Applications to Representation Theory and Geometry, Chapt. 9.4 . 1997 . London Mathematical Society Student Texts . 35 . Cambridge University Press . Cambridge, U.K.. 9780521567244 . 10.1017/CBO9780511626241 .
  7. Kleiman . S.L.. Laksov . Dan . Schubert Calculus . American Mathematical Society. American Mathematical Monthly . 79. 10 . 1972 . 0377-9017 . 10.1080/00029890.1972.11993188 . 1061-1082 .