Classical capacity explained
In quantum information theory, the classical capacity of a quantum channel is the maximum rate at which classical data can be sent over it error-free in the limit of many uses of the channel. Holevo, Schumacher, and Westmoreland proved the following least upper bound on the classical capacity of any quantum channel
:
\chi(l{N})=
I(X;B)l{N(\rho)}
where
is a classical-quantum state of the following form:
\rhoXA=\sumxpX(x)\vertx\rangle\langlex\vertX ⊗
,
is a probability distribution, and each
is a density operator that can be input to the channel
.
Achievability using sequential decoding
We briefly review the HSW coding theorem (thestatement of the achievability of the Holevo information rate
forcommunicating classical data over a quantum channel). We first review theminimal amount of quantum mechanics needed for the theorem. We then coverquantum typicality, and finally we prove the theorem using a recent sequentialdecoding technique.
Review of quantum mechanics
In order to prove the HSW coding theorem, we really just need a few basicthings from quantum mechanics. First, a quantum state is a unit trace,positive operator known as a density operator. Usually, we denote itby
,
,
, etc. The simplest model for a
quantum channelis known as a classical-quantum channel:
The meaning of the above notation is that inputting the classical letter
at the transmitting end leads to a quantum state
at the receivingend. It is the task of the receiver to perform a measurement to determine theinput of the sender. If it is true that the states
are perfectlydistinguishable from one another (i.e., if they have orthogonal supports suchthat
Tr\left\{\rhox
\right\}=0
for
), then the channel is a noiseless channel. We are interested in situationsfor which this is not the case. If it is true that the states
allcommute with one another, then this is effectively identical to the situationfor a classical channel, so we are also not interested in these situations.So, the situation in which we are interested is that in which the states
have overlapping support and are non-commutative.
The most general way to describe a quantum measurement is with a positive operator-valued measure (POVM). We usually denote the elements of a POVM as
. These operators should satisfypositivity and completeness in order to form a valid POVM:
The probabilistic interpretation of
quantum mechanics states that if someonemeasures a quantum state
using a measurement device corresponding tothe POVM
, then the probability
for obtaining outcome
is equal to
p\left(m\right)=Tr\left\{Λm\rho\right\},
and the post-measurement state is
}\rho\sqrt,if the person measuring obtains outcome
. These rules are sufficient for usto consider classical communication schemes over cq channels.
Quantum typicality
The reader can find a good review of this topic in the article about the typical subspace.
Gentle operator lemma
The following lemma is important for our proofs. Itdemonstrates that a measurement that succeeds with high probability on averagedoes not disturb the state too much on average:
Lemma: [Winter] Given anensemble
\left\{pX\left(x\right),\rhox\right\}
with expecteddensity operator
\rho\equiv\sumxpX\left(x\right)\rhox
, supposethat an operator
such that
succeeds with highprobability on the state
:
Tr\left\{Λ\rho\right\}\geq1-\epsilon.
Then the subnormalized state
is closein expected trace distance to the original state
:
EX\left\{\left\Vert\sqrt{Λ}\rhoX\sqrt{Λ}
-\rhoX\right\Vert1\right\}\leq2\sqrt{\epsilon}.
(Note that
is the nuclear norm of the operator
so that
\left\VertA\right\Vert1\equiv
Tr
\left\{\sqrt{A\daggerA}\right\}
.)
The following inequality is useful for us as well. It holds for any operators
,
,
such that
:
The quantum information-theoretic interpretation of the above inequality isthat the probability of obtaining outcome
from a quantum measurementacting on the state
is upper bounded by the probability of obtainingoutcome
on the state
summed with the distinguishability ofthe two states
and
.
Non-commutative union bound
Lemma: [Sen's bound] The following boundholds for a subnormalized state
such that
and
Tr\left\{\sigma\right\}\leq1
with
, ...,
beingprojectors:
Tr\left\{\sigma\right\}-Tr\left\{\PiN … \Pi
1 \sigma \Pi1 … \PiN\right\}
Tr\left\{\left(I-\Pii\right)\sigma\right\}},
We can think of Sen's bound as a "non-commutative unionbound" because it is analogous to the following union boundfrom probability theory:
\Pr\left\{\left(A1\cap … \capAN\right)c\right\}
=\Pr\left\{
\cup … \cup
\right\}
\Pr\left\{
\right\},
where
are events. The analogous bound for projectorlogic would be
Tr\left\{\left(I-\Pi1 … \PiN … \Pi1\right)
\rho\right\}
Tr\left\{\left(I-\Pii\right)
\rho\right\},
if we think of
as a projector onto the intersection ofsubspaces. Though, the above bound only holds if the projectors
,...,
are commuting (choosing
\Pi1=\left\vert+\right\rangle
\left\langle+\right\vert
,
\Pi2=\left\vert0\right\rangle\left\langle
0\right\vert
, and
\rho=\left\vert0\right\rangle\left\langle0\right\vert
gives a counterexample). If the projectors are non-commuting, then Sen'sbound is the next best thing and suffices for our purposes here.
HSW theorem with the non-commutative union bound
We now prove the HSW theorem with Sen's non-commutative union bound. Wedivide up the proof into a few parts: codebook generation, POVM construction,and error analysis.
Codebook Generation. We first describe how Alice and Bob agree on arandom choice of code. They have the channel
and adistribution
. They choose
classical sequences
according to the IID\ distribution
.After selecting them, they label them with indices as
\left\{xn\left(m\right)\right\}m\in\left[
. This leads to the followingquantum codewords:
⊗
… ⊗ \rho | |
| xn\left(m\right) |
.
The quantum codebook is then
. The average state of the codebook is then
where
\rho=\sumxpX\left(x\right)\rhox
.
POVM Construction . Sens' bound from the above lemma suggests a method for Bob to decode a state that Alice transmits. Bob shouldfirst ask "Is the received state in the average typicalsubspace?" He can do this operationally by performing atypical subspace measurement corresponding to
\left\{\Pi\rho,\deltan
\right\}
. Next, he asks in sequential order,"Is the received codeword in the
conditionally typical subspace?" This is in some senseequivalent to the question, "Is the received codeword the
transmitted codeword?" He can ask thesequestions operationally by performing the measurements corresponding to theconditionally typical projectors
\left\{
\Pi | |
| \rho | | ,\delta | | xn\left(m\right)
| |
|
,I-\Pi | |
| \rho | | ,\delta | | xn\left(m\right) | |
|
\right\}
.
Why should this sequential decoding scheme work well? The reason is that thetransmitted codeword lies in the typical subspace on average:
\left\{Tr\left\{\Pi\rho,\delta
\right\}\right\}=Tr\left\{\Pi\rho,\delta
\left\{
\right\}\right\}
=Tr\left\{\Pi\rho,\delta \rho ⊗ \right\}
where the inequality follows from (\ref). Also, theprojectors
\Pi | |
| \rho | | ,\delta | | xn\left(m\right) | |
|
are "good detectors" for the states
(on average) because the following condition holds from conditional quantumtypicality:
\left\{Tr\left\{
\right\}\right\}\geq1-\epsilon.
Error Analysis. The probability of detecting the
codeword correctly under our sequential decoding scheme is equal to
Tr\left\{
\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\hat{\Pi}
| |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
… \hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(1\right) | |
|
\hat{\Pi} | |
| \rho | | ,\delta
| | Xn\left(1\right) | |
|
… \hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
\Pi | |
| \rho
| | ,\delta | | Xn\left(m\right) | |
|
\right\},
where we make the abbreviation
. (Observe that weproject into the average typical subspace just once.) Thus, the probability ofan incorrect detection for the
codeword is given by
1-Tr\left\{
\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\hat{\Pi}
| |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
… \hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(1\right) | |
|
\hat{\Pi} | |
| \rho | | ,\delta
| | Xn\left(1\right) | |
|
… \hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
\Pi | |
| \rho
| | ,\delta | | Xn\left(m\right) | |
|
\right\},
and the average error probability of this scheme is equal to
\summTr\left\{
\Pi | |
| \rho | | ,\delta | | Xn\left(m\right)
| |
|
\hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
… \hat{\Pi
} | |
| \rho | | ,\delta | | Xn\left(1\right) | |
|
\hat{\Pi} | |
| \rho
| | ,\delta | | Xn\left(1\right) | |
|
… \hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\right\}.
Instead of analyzing the average error probability, we analyze the expectationof the average error probability, where the expectation is with respect to therandom choice of code:
Our first step is to apply Sen's bound to the above quantity. But before doingso, we should rewrite the above expression just slightly, by observing that
1
\left\{
\summ
Tr\left\{
\rho | |
| Xn\left(m\right) |
\right\}\right\}
\left\{
\summTr\left\{
\right\}
| n |
+Tr\left\{
\hat{\Pi} | |
| \rho,\delta |
\right\}\right\}
\left\{
\summTr\left\{
\right\}
\right\}+
\summTr\left\{\hat{\Pi}\rho,\deltan
\left\{
\right\}
\right\}
\left\{
\summTr\left\{
\right\}
\right\}+Tr\left\{
\rho ⊗
n\right\}
\left\{
\summ
| n |
Tr\left\{
\Pi | |
| \rho,\delta |
\Pi\rho,\deltan\right\}\right\}+\epsilon
Substituting into (and forgetting about the small
term for now) gives an upper bound of
\left\{
\summTr\left\{
\right\}
\right\}
\left\{
\summTr\left\{
\Pi
| |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
… \hat{\Pi} | |
| \rho | | ,\delta
| | Xn\left(1\right) | |
|
\hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(1\right) | |
|
… \hat{\Pi}
| |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
\Pi | |
| \rho | | ,\delta | | Xn\left(m\right)
| |
|
\right\}\right\}.
We then apply Sen's bound to this expression with
and the sequentialprojectors as
\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
,
\hat{\Pi}
| |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
, ...,
\hat{\Pi} | |
| \rho
| | ,\delta | | Xn\left(1\right) | |
|
. This gives the upper bound
\left\{
\summ2\left[Tr\left\{
\left(
I-\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\right)
Tr\left\{
\Pi | |
| \rho | | ,\delta
| | Xn\left(i\right) | |
|
\Pi\rho,\deltan\right\}\right]1/2\right\}.
Due to concavity of the square root, we can bound this expression from aboveby
2\left[
\left\{
\summTr\left\{
\left(
I-\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\right)
Tr\left\{
\Pi | |
| \rho | | ,\delta
| | Xn\left(i\right) | |
|
\Pi\rho,\deltan\right\}\right\}\right]1/2
\leq2\left[
\left\{
\summTr\left\{
\left(
I-\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\right)
\right\}
+\sumi ≠ Tr\left\{
\Pi | |
| \rho | | ,\delta
| | Xn\left(i\right) | |
|
\Pi\rho,\deltan\right\}\right\}\right]1/2,
where the second bound follows by summing over all of the codewords not equalto the
codeword (this sum can only be larger).
We now focus exclusively on showing that the term inside the square root canbe made small. Consider the first term:
\left\{
\summTr\left\{
\left(I-\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\right)\Pi\rho,\deltan
\right\}\right\}
\left\{
\summTr\left\{
\left(I-\Pi | |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\right)
\right\}+\left\Vert
\Pi\rho,\deltan\right\Vert1\right\}
\leq\epsilon+2\sqrt{\epsilon}.
where the first inequality follows from and thesecond inequality follows from the gentle operator lemma and theproperties of unconditional and conditional typicality. Consider now thesecond term and the following chain of inequalities:
\sumi ≠
\left\{Tr\left\{
\Pi | |
| \rho
| | ,\delta | | Xn\left(i\right) | |
|
\right\}\right\}
=\sumi ≠ Tr\left\{
\left\{
\Pi
| |
| \rho | | ,\delta | | Xn\left(i\right) | |
|
\right\} \Pi\rho,\deltan
\left\{
\right\}
=\sumi ≠ Tr\left\{
\left\{
\Pi
| |
| \rho | | ,\delta | | Xn\left(i\right) | |
|
\right\} \Pi\rho,\deltan \rho ⊗
\right\}
\leq\sumi ≠ 2-n\left[ Tr\left\{
\left\{
\Pi | |
| \rho | | ,\delta | | Xn\left(i\right) | |
|
\right\}
\right\}
The first equality follows because the codewords
and
are independent since they are different. The secondequality follows from . The first inequality follows from(\ref). Continuing, we have
\leq\sumi ≠ 2-n\left[
\left\{Tr\left\{
\Pi | |
| \rho | | ,\delta | | Xn\left(i\right) | |
|
\right\}\right\}
\leq\sumi ≠ 2-n\left[ 2n\left[
The first inequality follows from
and exchangingthe trace with the expectation. The second inequality follows from(\ref). The next two are straightforward.
Putting everything together, we get our final bound on the expectation of theaverage error probability:
\left\{
\summTr\left\{
\Pi
| |
| \rho | | ,\delta | | Xn\left(m\right) | |
|
\hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
… \hat{\Pi} | |
| \rho | | ,\delta
| | Xn\left(1\right) | |
|
\hat{\Pi} | |
| \rho | | ,\delta | | Xn\left(1\right) | |
|
… \hat{\Pi}
| |
| \rho | | ,\delta | | Xn\left(m-1\right) | |
|
\Pi | |
| \rho | | ,\delta | | Xn\left(m\right)
| |
|
\right\}\right\}
\leq\epsilon+2\left[\left(\epsilon+2\sqrt{\epsilon}\right)
+M 2-n\left[\right]1/2.
Thus, as long as we choose
, there exists a code with vanishing error probability.
See also
References