Matrix product state explained

A matrix product state (MPS) is a representation of a quantum many-body state. It is at the core of the one of the most effective algorithms for solving one dimensional strongly correlated quantum systems – the density matrix renormalization group (DMRG) algorithm.

For a system of

N

spins of dimension

d

, the general form of the MPS for periodic boundary conditions (PBS) can be written in the following form:|\Psi\rangle = \sum_ \operatorname\left[A_1^{(s_1)} A_2^{(s_2)} \cdots A_N^{(s_N)}\right] |s_1 s_2 \ldots s_N\rangle.For open boundary conditions (OBC),

|\Psi\rangle

takes the form

|\Psi\rangle=\sum\{s\

} A_1^ A_2^ \cdots A_N^ |s_1 s_2 \ldots s_N\rangle.

Here

(si)
A
i

are the

Di x Di+1

matrices (

D

is the dimension of the virtual subsystems) and

|si\rangle

are the single-site basis states. For periodic boundary conditions, we consider

DN+1=D1

, and for open boundary conditions

D1=1

. The parameter

D

  is related to the entanglement between particles. In particular, if the state is a product state (i.e. not entangled at all), it can be described as a matrix product state with

D=1

.

\{si\}

represents a

d

-dimensional local space on site

i=1,2,...,N

. For qubits,

si\in\{0,1\}

. For qudits (-level systems),

si\in\{0,1,\ldots,d-1\}

.

For states that are translationally symmetric, we can choose: A_1^ = A_2^ = \cdots = A_N^ \equiv A^.In general, every state can be written in the MPS form (with

D

growing exponentially with the particle number ). Note that the MPS decomposition is not unique. MPS are practical when

D

is small – for example, does not depend on the particle number. Except for a small number of specific cases (some mentioned in the section Examples), such a thing is not possible, though in many cases it serves as a good approximation.

For introductions see, and.[1] In the context of finite automata see.[2] For emphasis placed on the graphical reasoning of tensor networks, see the introduction.

Wave function as a Matrix Product State

For a system of

N

lattice sites each of which has a

d

-dimensional Hilbert space, the completely general state can be written as

|\Psi\rangle=\sum\{s\

} \psi_ |s_1 \ldots s_N\rangle,

where

\psi
s1...sN
is a

dN

-dimensional tensor. For example, the wave function of the system described by the Heisenberg model is defined by the

2N

dimensional tensor, whereas for the Hubbard model the rank is

4N

.

The main idea of the MPS approach is to separate physical degrees of freedom of each site, so that the wave function can be rewritten as the product of

N

matrices, where each matrix corresponds to one particular site. The whole procedure includes the series of reshaping and singular value decompositions (SVD).[3]

There are three ways to represent wave function as an MPS: left-canonical decomposition, right-canonical decomposition, and mixed-canonical decomposition.

Left-Canonical Decomposition

The decomposition of the

dN

-dimensional tensor starts with the separation of the very left index, i.e., the first index

s1

, which describes physical degrees of freedom of the first site. It is performed by reshaping

|\Psi\rangle

as follows

|\Psi\rangle=\sum\{s\

} \psi_ |s_1 \ldots s_N\rangle.

In this notation,

s1

is treated as a row index,

(s2\ldotssN)

as a column index, and the coefficient
\psi
s1,(s2...sN)
is of dimension

(d x dN-1)

. The SVD procedure yields
\psi
s1,(s2...sN)

=

r1
\sum
\alpha1
U
s1,\alpha1
D
\alpha1,\alpha1

(V\dagger

)
\alpha1,(s2...sN)

=

r1
\sum
\alpha1
U
s1,\alpha1
\psi
\alpha1,(s2...sN)

=

r1
\sum
\alpha1
s1
A
\alpha1
\psi
\alpha1,(s2...sN)

.

In the relation above, matrices

D

and

V\dagger

are multiplied and form the matrix
\psi
\alpha1,(s2...sN)

and

r1\leqd

.
s1
A
\alpha1
stores the information about the first lattice site. It was obtained by decomposing matrix

U

into

d

row vectors
s1
A
with entries
s1
A
\alpha1
=U
s1,\alpha1

. So, the state vector takes the form

|\Psi\rangle=\sum\{s\

} \sum_ A^_ \psi_ |s_1 \ldots s_N\rangle.

The separation of the second site is performed by grouping

s2

and

\alpha1

, and representing
\psi
\alpha1,(s2...sN)

as a matrix
\psi
(\alpha1s2),(s3...sN)
of dimension

(r1d x dN-2)

. The subsequent SVD of
\psi
(\alpha1s2),(s3...sN)
can be performed as follows:
\psi
(\alpha1s2),(s3...sN)

=

r2
\sum
\alpha2
U
(\alpha1s2),\alpha2
D
\alpha2,\alpha2

(V\dagger

)
\alpha2,(s3...sN)

=

r2
\sum
\alpha2
s2
A
\alpha1,\alpha2
\psi
\alpha2,(s3...sN)
.Above we replace

U

by a set of

d

matrices of dimension

(r1 x r2)

with entries
s2
A
\alpha1,\alpha2

=

U
(\alpha1s2),\alpha2
. The dimension of
\psi
\alpha2,(s3...sN)
is

(r2 x dN-2)

with

r2\leqr1d\leqd2

. Hence,

|\Psi\rangle=\sum\{s\

} \sum_ A^_ \psi_ |s_1 \ldots s_N\rangle = \sum_ \sum_ A^_ A^_ \psi_ |s_1 \ldots s_N\rangle.

Following the steps described above, the state

|\Psi\rangle

can be represented as a product of matrices

|\Psi\rangle=\sum\{s\

} \sum_ A^_ A^_\ldots A^_ A^_ |s_1 \ldots s_N\rangle.

The maximal dimensions of the

A

-matrices take place in the case of the exact decomposition, i.e., assuming for simplicity that

N

is even,

(1 x d),(d x d2),\ldots,(dN/2-1 x dN/2),(dN/2 x dN/2-1),\ldots,(d2 x d),(d x 1)

going from the first to the last site. However, due to the exponential growth of the matrix dimensions in most of the cases it is impossible to perform the exact decomposition.

The dual MPS is defined by replacing each matrix

A

with

A*

:

\langle\Psi| =\sum\limits\{s\

}\sum\limits_A^_A^_...A^_A^_\langle s'_1...s'_N|.

Note that each matrix

U

in the SVD is a semi-unitary matrix with property

U\daggerU=I

. This leads to
\delta
\alphai,\alphaj

=

\sum
\alphai-1si
\dagger)
(U
\alphai,(\alphai-1si)
U
(\alphai-1si),\alphaj

=

\sum
\alphai-1si
si\dagger
(A
)
\alphai,\alphai-1
si
A
\alphai-1,\alphaj

=

\sum
si
si\dagger
(A
si
A
)
\alphai,\alphaj
.

To be more precise,

\sum
si
si\dagger
A
si
A

=I

. Since matrices are left-normalized, we call the composition left-canonical.

Right-Canonical Decomposition

Similarly, the decomposition can be started from the very right site. After the separation of the first index, the tensor

\psi
s1...sN
transforms as follows:
\psi
s1...sN

=

\psi
(s1...sN-1),sN

=

\sum
\alphaN-1
U
(s1...sN-1),\alphaN-1
D
\alphaN-1,\alphaN-1
\dagger)
(V
\alphaN-1,sN

=

\sum
\alphaN-1
\psi
(s1...sN-1),\alphaN-1
sN
B
\alphaN-1
.

The matrix

\psi
(s1...sN-1),\alphaN-1
was obtained by multiplying matrices

U

and

D

, and the reshaping of
\dagger)
(V
\alphaN-1,sN
into

d

column vectors forms
sN
B
\alphaN-1
. Performing the series of reshaping and SVD, the state vector takes the form

|\Psi\rangle=\sum\{s\

} \sum_ B^_ B^_\ldots B^_ B^_ |s_1 \ldots s_N\rangle.

Since each matrix

V

in the SVD is a semi-unitary matrix with property

V\daggerV=I

, the

B

-matrices are right-normalized and obey
\sum
si
si
B
si\dagger
B

=I

. Hence, the decomposition is called right-canonical.

Mixed-Canonical Decomposition

The decomposition performs from both the right and from the left. Assuming that the left-canonical decomposition was performed for the first n sites,

\psi
s1...sN
can be rewritten as
\psi
s1...sN

=

\sum
\alpha1,\ldots,\alphan
s1
A
\alpha1
s2
A
\alpha1,\alpha2

\ldots

sn
A
\alphan-1,\alphan
D
\alphan,\alphan
\dagger)
(V
\alphan,(sn+1...sN)
.In the next step, we reshape
\dagger)
(V
\alphan,(sn+1...sN)

as
\psi
(\alphansn+1...sn-1),sN
and proceed with the series of reshaping and SVD from the right up to site

sn+1

:
\begin{align} \psi
(\alphansn+1...sn-1),sN

&=&

\sum
\alphan+1...\alphaN
U
(\alphansn+1),\alphan+1
D
\alphan+1,\alphan+1
sn+2
B
\alphan+1,\alphan+2

\ldots

sN-1
B
\alphaN-2,\alphaN-1
sN
B
\alphaN-1

\\ &=&

\sum
\alphan+1...\alphaN
sn+1
B
\alphan,\alphan+1
sn+2
B
\alphan+1,\alphan+2

\ldots

sN-1
B
\alphaN-2,\alphaN-1
sN
B
\alphaN-1

\end{align}

.

As the result,

\psi
s1...sN

=

\sum
\alpha1,\ldots,\alphaN
s1
A
\alpha1
s2
A
\alpha1,\alpha2

\ldots

sn
A
\alphan-1,\alphan
D
\alphan,\alphan
sn+1
B
\alphan,\alphan+1
sn+2
B
\alphan+1,\alphan+2

\ldots

sN-1
B
\alphaN-2,\alphaN-1
sN
B
\alphaN-1

.

Examples

Greenberger–Horne–Zeilinger state

Greenberger–Horne–Zeilinger state, which for particles can be written as superposition of zeros and ones

|GHZ\rangle=

|0\rangle+|1\rangle
\sqrt{2
}can be expressed as a Matrix Product State, up to normalization, with

A(0)= \begin{bmatrix} 1&0\\ 0&0 \end{bmatrix} A(1)= \begin{bmatrix} 0&0\\ 0&1 \end{bmatrix},

or equivalently, using notation from:

A= \begin{bmatrix} |0\rangle&0\\ 0&|1\rangle \end{bmatrix}.

This notation uses matrices with entries being state vectors (instead of complex numbers), and when multiplying matrices using tensor product for its entries (instead of product of two complex numbers). Such matrix is constructed as

A\equiv|0\rangleA(0)+|1\rangleA(1)+\ldots+|d-1\rangleA(d-1).

Note that tensor product is not commutative.

In this particular example, a product of two A matrices is:

AA= \begin{bmatrix} |00\rangle&0\\ 0&|11\rangle \end{bmatrix}.

W state

W state, i.e., the superposition of all the computational basis states of Hamming weight one.

|W\rangle=

1
\sqrt{3
}(|001\rangle + |010\rangle + |100\rangle)

Even though the state is permutation-symmetric, its simplest MPS representation is not. For example:

A1= \begin{bmatrix} |0\rangle&0\\ |0\rangle&|1\rangle \end{bmatrix} A2= \begin{bmatrix} |0\rangle&|1\rangle\\ 0&|0\rangle \end{bmatrix} A3= \begin{bmatrix} |1\rangle&0\\ 0&|0\rangle \end{bmatrix}.

AKLT model

See main article: AKLT model.

The AKLT ground state wavefunction, which is the historical example of MPS approach, corresponds to the choice

A+=\sqrt{

2
3
}\ \sigma^=\begin0 & \sqrt\\0 & 0\end

A0=

-1
\sqrt{3
}\ \sigma^=\begin-1/\sqrt & 0\\0 & 1/\sqrt\end

A-=-\sqrt{

2
3
}\ \sigma^=\begin0 & 0\\-\sqrt & 0\end

where the

\sigma's

are Pauli matrices, or

A=

1
\sqrt{3
}\begin- | 0 \rangle & \sqrt | + \rangle\\- \sqrt | - \rangle & | 0 \rangle\end.

Majumdar–Ghosh model

See main article: Majumdar–Ghosh model.

Majumdar–Ghosh ground state can be written as MPS with

A= \begin{bmatrix} 0&\left|\uparrow\right\rangle&\left|\downarrow\right\rangle\\

-1
\sqrt{2
} \left| \downarrow \right\rangle & 0 & 0 \\\frac \left| \uparrow \right\rangle & 0 & 0\end.

See also

External links

Notes and References

  1. Bridgeman . Jacob C . Chubb . Christopher T . 2017-06-02 . Hand-waving and interpretive dance: an introductory course on tensor networks . Journal of Physics A: Mathematical and Theoretical . 50 . 22 . 223001 . 10.1088/1751-8121/aa6dc3 . 1603.03039 . 2017JPhA...50v3001B . 1751-8113.
  2. Crosswhite . Gregory M. . Bacon . Dave . 2008-07-29 . Finite automata for caching in matrix product algorithms . Physical Review A . en . 78 . 1 . 012356 . 10.1103/PhysRevA.78.012356 . 0708.1221 . 2008PhRvA..78a2356C . 1050-2947.
  3. Baker . Thomas E. . Desrosiers . Samuel . Tremblay . Maxime . Thompson . Martin P. . Méthodes de calcul avec réseaux de tenseurs en physique . Canadian Journal of Physics . 2021 . en . 99 . 4 . 207–221 . 10.1139/cjp-2019-0611 . 1911.11566 . 2021CaJPh..99..207B . 0008-4204.