Z-channel (information theory) explained

In coding theory and information theory, a Z-channel or binary asymmetric channel is a communications channel used to model the behaviour of some data storage systems.

Definition

A Z-channel is a channel with binary input and binary output, where each 0 bit is transmitted correctly, but each 1 bit has probability p of being transmitted incorrectly as a 0, and probability 1–p of being transmitted correctly as a 1. In other words, if X and Y are the random variables describing the probability distributions of the input and the output of the channel, respectively, then the crossovers of the channel are characterized by the conditional probabilities:

\begin{align} \operatorname{Pr}[Y=0|X=0]&=1\\ \operatorname{Pr}[Y=0|X=1]&=p\\ \operatorname{Pr}[Y=1|X=0]&=0\\ \operatorname{Pr}[Y=1|X=1]&=1-p \end{align}

Capacity

cap(Z)

of the Z-channel

Z

with the crossover 1 → 0 probability p, when the input random variable X is distributed according to the Bernoulli distribution with probability

\alpha

for the occurrence of 0, is given by the following equation:

cap(Z)=H\left(

1
1+2s(p)

\right)-

s(p)
1+2s(p)

=

-s(p)
log
2(1{+}2

)=log2\left(1+(1-p)pp/(1-p)\right)

where

s(p)=

H(p)
1-p
for the binary entropy function

H()

.

This capacity is obtained when the input variable X has Bernoulli distribution with probability

\alpha

of having value 0 and

1-\alpha

of value 1, where:

\alpha=1-

1
(1-p)(1+2H(p)/(1-p))

,

For small p, the capacity is approximated by

cap(Z)1-0.5H(p)

as compared to the capacity

1{-}H(p)

of the binary symmetric channel with crossover probability p.

For any p,

\alpha<0.5

(i.e. more 0s should be transmitted than 1s) because transmitting a 1 introduces noise. As

p1

, the limiting value of

\alpha

is
1
e
.

Bounds on the size of an asymmetric-error-correcting code

Define the following distance function

dA(x,y)

on the words

x,y\in\{0,1\}n

of length n transmitted via a Z-channel

dA(x,y)\stackrel{\vartriangle}{=}max\left\{|\{i\midxi=0,yi=1\}|,|\{i\midxi=1,yi=0\}|\right\}.

Define the sphere

Vt(x)

of radius t around a word

x\in\{0,1\}n

of length n as the set of all the words at distance t or less from

x

, in other words,

Vt(x)=\{y\in\{0,1\}n\middA(x,y)\leqt\}.

A code

l{C}

of length n is said to be t-asymmetric-error-correcting if for any two codewords

c\nec'\in\{0,1\}n

, one has

Vt(c)\capVt(c')=\emptyset

. Denote by

M(n,t)

the maximum number of codewords in a t-asymmetric-error-correcting code of length n.

The Varshamov bound.For n≥1 and t≥1,

M(n,t)\leq

2n+1
t{\left(
\sum\binom{\lfloorn/2\rfloor
j=0

{j}+\binom{\lceiln/2\rceil}{j}\right)}}.

The constant-weight code bound.For n > 2t ≥ 2, let the sequence B0, B1, ..., Bn-2t-1 be defined as

B0=2,Bi=min0\{Bj+A(n{+}t{+}i{-}j{-}1,2t{+}2,t{+}i)\}

for

i>0

.Then

M(n,t)\leqBn-2t-1.

References