Rank factorization explained

In mathematics, given a field

F

, nonnegative integers

m,n

, and a matrix

A\inFm x

, a rank decomposition or rank factorization of is a factorization of of the form, where

C\inFm x

and

F\inFr x

, where

r=\operatorname{rank}A

is the rank of

A

.

Existence

Every finite-dimensional matrix has a rank decomposition: Let A be an m\times n matrix whose column rank is r. Therefore, there are r linearly independent columns in A; equivalently, the dimension of the column space of A is r. Let \mathbf_1, \mathbf_2, \ldots, \mathbf_r be any basis for the column space of A and place them as column vectors to form the m\times r matrix C = \begin\mathbf_1 & \mathbf_2 & \cdots & \mathbf_r\end. Therefore, every column vector of A is a linear combination of the columns of C. To be precise, if A = \begin\mathbf_1 & \mathbf_2 & \cdots & \mathbf_n\end is an m\times n matrix with \mathbf_j as the j-th column, then

aj=f1jc1+f2jc2++frjcr,

where f_'s are the scalar coefficients of \mathbf_j in terms of the basis \mathbf_1, \mathbf_2, \ldots, \mathbf_r. This implies that A = CF, where f_ is the (i,j)-th element of F.

Non-uniqueness

If A = C_1 F_1 is a rank factorization, taking C_2 = C_1 R and F_2 = R^ F_1 gives another rank factorization for any invertible matrix R of compatible dimensions.

Conversely, if A = F_G_ = F_G_ are two rank factorizations of A, then there exists an invertible matrix R such that F_1 = F_2 R and G_1 = R^ G_.[1]

Construction

Rank factorization from reduced row echelon forms

In practice, we can construct one specific rank factorization as follows: we can compute B, the reduced row echelon form of A. Then C is obtained by removing from A all non-pivot columns (which can be determined by looking for columns in B which do not contain a pivot), and F is obtained by eliminating any all-zero rows of B.

Note: For a full-rank square matrix (i.e. when n=m=r), this procedure will yield the trivial result C=A and F=B=I_n (the n\times n identity matrix).

Example

Consider the matrix

A=\begin{bmatrix}1&3&1&4\ 2&7&3&9\ 1&5&3&1\ 1&2&0&8\end{bmatrix} \sim\begin{bmatrix}1&0&-2&0\ 0&1&1&0\ 0&0&0&1\ 0&0&0&0\end{bmatrix}=B.

B is in reduced echelon form.

Then C is obtained by removing the third column of A, the only one which is not a pivot column, and F by getting rid of the last row of zeroes from B, so

C=\begin{bmatrix}1&3&4\ 2&7&9\ 1&5&1\ 1&2&8\end{bmatrix},    F=\begin{bmatrix}1&0&-2&0\ 0&1&1&0\ 0&0&0&1\end{bmatrix}.

It is straightforward to check that

A=\begin{bmatrix}1&3&1&4\ 2&7&3&9\ 1&5&3&1\ 1&2&0&8\end{bmatrix} =\begin{bmatrix}1&3&4\ 2&7&9\ 1&5&1\ 1&2&8\end{bmatrix} \begin{bmatrix}1&0&-2&0\ 0&1&1&0\ 0&0&0&1\end{bmatrix} =CF.

Proof

Let P be an n\times n permutation matrix such that AP = (C, D) in block partitioned form, where the columns of C are the r pivot columns of A. Every column of D is a linear combination of the columns of C, so there is a matrix G such that D = CG, where the columns of G contain the coefficients of each of those linear combinations. So AP = (C, CG) = C(I_r, G), I_r being the r\times r identity matrix. We will show now that (I_r, G) = FP.

Transforming A into its reduced row echelon form B amounts to left-multiplying by a matrix E which is a product of elementary matrices, so EAP = BP = EC(I_r, G), where EC = \begin I_r \\ 0 \end. We then can write BP = \begin I_r & G \\ 0 & 0 \end, which allows us to identify (I_r, G) = FP, i.e. the nonzero r rows of the reduced echelon form, with the same permutation on the columns as we did for A. We thus have AP = CFP, and since P is invertible this implies A = CF, and the proof is complete.

Singular value decomposition

If

F\in\{R,C\},

then one can also construct a full-rank factorization of A via a singular value decomposition

A=U\SigmaV* =\begin{bmatrix}U1&U2\end{bmatrix}\begin{bmatrix}\Sigmar&0\ 0&0\end{bmatrix}\begin{bmatrix}

*
V
1
*
\V
2

\end{bmatrix} =U1\left(\Sigmar

*\right)
V
1

.

Since U_1 is a full-column-rank matrix and \Sigma_r V_1^* is a full-row-rank matrix, we can take C = U_1 and F = \Sigma_r V_1^*.

Consequences

rank(A) = rank(AT)

An immediate consequence of rank factorization is that the rank of A is equal to the rank of its transpose A^\textsf. Since the columns of A are the rows of A^\textsf, the column rank of A equals its row rank.

Proof: To see why this is true, let us first define rank to mean column rank. Since A = CF, it follows that A^\textsf = F^\textsfC^\textsf. From the definition of matrix multiplication, this means that each column of A^\textsf is a linear combination of the columns of F^\textsf. Therefore, the column space of A^\textsf is contained within the column space of F^\textsf and, hence, \operatorname\left(A^\textsf\right) \leq \operatorname\left(F^\textsf\right).

Now, F^\textsf is n \times r, so there are r columns in F^\textsf and, hence, \operatorname\left(A^\textsf\right) \leq r = \operatorname\left(A\right). This proves that \operatorname\left(A^\textsf\right) \leq \operatorname\left(A\right).

Now apply the result to A^\textsf to obtain the reverse inequality: since \left(A^\textsf\right)^\textsf = A, we can write \operatorname\left(A\right)= \operatorname\left(\left(A^\textsf\right)^\textsf\right) \leq \operatorname\left(A^\textsf\right). This proves \operatorname\left(A\right) \leq \operatorname\left(A^\textsf\right).

We have, therefore, proved \operatorname\left(A^\textsf\right) \leq \operatorname\left(A\right) and \operatorname\left(A\right) \leq \operatorname\left(A^\textsf\right), so \operatorname\left(A\right) = \operatorname\left(A^\textsf\right).

References

Notes and References

  1. Piziak. R.. Odell. P. L.. Full Rank Factorization of Matrices. Mathematics Magazine. 1 June 1999. 72. 3. 193. 10.2307/2690882. 2690882.