Expander code explained
Expander codes |
Block Length: |
|
Message Length: |
|
Rate: |
|
Distance: |
|
Alphabet Size: |
|
Notation: | [n,n-m,2(1-\epsilon)\gamma ⋅ n]2
-code |
In coding theory, expander codes form a class of error-correcting codes that are constructed from bipartite expander graphs.Along with Justesen codes, expander codes are of particular interest since they have a constant positive rate, a constant positive relative distance, and a constant alphabet size.In fact, the alphabet contains only two elements, so expander codes belong to the class of binary codes.Furthermore, expander codes can be both encoded and decoded in time proportional to the block length of the code.
Expander codes
In coding theory, an expander code is a
linear block code whose parity check matrix is the adjacency matrix of a bipartite
expander graph. These codes have good relative distance
, where
and
are properties of the expander graph as defined later, rate
\left(1-\tfrac{m}{n}\right)
, and decodability (algorithms of running time
exist).
Definition
Let
be a
-
biregular graph between a set of
nodes
, called
variables, and a set of
nodes
, called
constraints.
Let
be a function designed so that, for each constraint
, the variables neighboring
are
.
Let
be an error-correcting code of block length
. The
expander code
is the code of block length
whose codewords are the words
such that, for
,
is a codeword of
.
[1] It has been shown that nontrivial lossless expander graphs exist. Moreover, we can explicitly construct them.[2]
Rate
The rate of
is its dimension divided by its block length. In this case, the parity check matrix has size
, and hence
has rate at least
.
Distance
Suppose
. Then the distance of a
(n,m,d,\gamma,1-\varepsilon)
expander code
is at least
.
Proof
Note that we can consider every codeword
in
as a subset of vertices
, by saying that vertex
if and only if the
th index of the codeword is a 1. Then
is a codeword iff every vertex
is adjacent to an even number of vertices in
. (In order to be a codeword,
, where
is the parity check matrix. Then, each vertex in
corresponds to each column of
. Matrix multiplication over
then gives the desired result.) So, if a vertex
is adjacent to a single vertex in
, we know immediately that
is not a codeword. Let
denote the neighbors in
of
, and
denote those neighbors of
which are unique, i.e., adjacent to a single vertex of
.
Lemma 1
For every
of size
,
d|S|\geq|N(S)|\geq|U(S)|\geqd(1-2\varepsilon)|S|
.
Proof
Trivially,
, since
implies
.
follows since the degree of every vertex in
is
. By the expansion property of the graph, there must be a set of
edges which go to distinct vertices. The remaining
edges make at most
neighbors not unique, so
U(S)\geqd(1-\varepsilon)|S|-d\varepsilon|S|=d(1-2\varepsilon)|S|
.
Corollary
Every sufficiently small
has a unique neighbor. This follows since
.
Lemma 2
Every subset
with
|T|<2(1-\varepsilon)\gamman
has a unique neighbor.
Proof
Lemma 1 proves the case
, so suppose
2(1-\varepsilon)\gamman>|T|>\gamman
. Let
such that
. By Lemma 1, we know that
|U(S)|\geqd(1-2\varepsilon)|S|
. Then a vertex
is in
iff
, and we know that
|T\setminusS|\leq2(1-\varepsilon)\gamman-\gamman=(1-2\varepsilon)\gamman
, so by the first part of Lemma 1, we know
|N(T\setminusS)|\leqd(1-2\varepsilon)\gamman
. Since
,
|U(T)|\geq|U(S)\setminusN(T\setminusS)|\geq|U(S)|-|N(T\setminusS)|>0
, and hence
is not empty.
Corollary
Note that if a
has at least 1 unique neighbor, i.e.
, then the corresponding word
corresponding to
cannot be a codeword, as it will not multiply to the all zeros vector by the parity check matrix. By the previous argument,
c\inC\implieswt(c)\geq2(1-\varepsilon)\gamman
. Since
is linear, we conclude that
has distance at least
.
Encoding
The encoding time for an expander code is upper bounded by that of a general linear code -
by matrix multiplication. A result due to Spielman shows that encoding is possible in
time.
[3] Decoding
Decoding of expander codes is possible in
time when
using the following algorithm.
Let
be the vertex of
that corresponds to the
th index in the codewords of
. Let
be a received word, and
V(y)=\{vi\midtheithpositionofyisa1\}
. Let
be
|\{v\inR\midvi\inN(v)andN(v)\capV(y)iseven\}|
, and
be
|\{v\inR\midvi\inN(v)andN(v)\capV(y)isodd\}|
. Then consider the greedy algorithm:----
Input: received word
. initialize y' to y while there is a v in R adjacent to an odd number of vertices in V(y') if there is an i such that o(i) > e(i) flip entry i in y' else fail
Output: fail, or modified codeword
.----
Proof
We show first the correctness of the algorithm, and then examine its running time.
Correctness
We must show that the algorithm terminates with the correct codeword when the received codeword is within half the code's distance of the original codeword. Let the set of corrupt variables be
,
, and the set of unsatisfied (adjacent to an odd number of vertices) vertices in
be
. The following lemma will prove useful.
Lemma 3
If
, then there is a
with
.
Proof
By Lemma 1, we know that
U(S)\geqd(1-2\varepsilon)s
. So an average vertex has at least
unique neighbors (recall unique neighbors are unsatisfied and hence contribute to
), since
, and thus there is a vertex
with
.
So, if we have not yet reached a codeword, then there will always be some vertex to flip. Next, we show that the number of errors can never increase beyond
.
Lemma 4
If we start with
s<\gamma(1-2\varepsilon)n
, then we never reach
at any point in the algorithm.
Proof
When we flip a vertex
,
and
are interchanged, and since we had
, this means the number of unsatisfied vertices on the right decreases by at least one after each flip. Since
s<\gamma(1-2\varepsilon)n
, the initial number of unsatisfied vertices is at most
, by the graph's
-regularity. If we reached a string with
errors, then by Lemma 1, there would be at least
unique neighbors, which means there would be at least
unsatisfied vertices, a contradiction.
Lemmas 3 and 4 show us that if we start with
s<\gamma(1-2\varepsilon)n
(half the distance of
), then we will always find a vertex
to flip. Each flip reduces the number of unsatisfied vertices in
by at least 1, and hence the algorithm terminates in at most
steps, and it terminates at some codeword, by Lemma 3. (Were it not at a codeword, there would be some vertex to flip). Lemma 4 shows us that we can never be farther than
away from the correct codeword. Since the code has distance
2(1-\varepsilon)\gamman>\gamman
(since
), the codeword it terminates on must be the correct codeword, since the number of bit flips is less than half the distance (so we couldn't have traveled far enough to reach any other codeword).
Complexity
We now show that the algorithm can achieve linear time decoding. Let
be constant, and
be the maximum degree of any vertex in
. Note that
is also constant for known constructions.
- Pre-processing: It takes
time to compute whether each vertex in
has an odd or even number of neighbors.
- Pre-processing 2: We take
time to compute a list of vertices
in
which have
.
- Each Iteration: We simply remove the first list element. To update the list of odd / even vertices in
, we need only update
entries, inserting / removing as necessary. We then update
entries in the list of vertices in
with more odd than even neighbors, inserting / removing as necessary. Thus each iteration takes
time.
- As argued above, the total number of iterations is at most
.
This gives a total runtime of
time, where
and
are constants.
See also
Notes
This article is based on Dr. Venkatesan Guruswami's course notes.[4]
References
- 10.1109/18.556667. Expander codes . 1996 . Sipser . M. . Spielman . D.A. . IEEE Transactions on Information Theory . 42 . 6 . 1710–1722 .
- Book: M. . Capalbo . O. . Reingold . S. . Vadhan . A. . Wigderson . Randomness conductors and constant-degree lossless expanders . http://dl.acm.org/citation.cfm?id=510003 . STOC '02 Proceedings of the thirty-fourth annual ACM symposium on Theory of computing . ACM . 2002 . 978-1-58113-495-7 . 659–668 . 10.1145/509907.510003. 1918841 .
- D. . Spielman . Linear-time encodable and decodable error-correcting codes . IEEE Transactions on Information Theory . 42 . 6 . 1723–31 . 1996 . 10.1109/18.556668 . 10.1.1.47.2736 .
- Web site: V. . Guruswami . Lecture 13: Expander Codes . 15 November 2006 . CSE 533: Error-Correcting . University of Washington .
Web site: V. . Guruswami . Notes 8: Expander Codes and their decoding . March 2010 . Introduction to Coding Theory . Carnegie Mellon University .
V. . Guruswami . Guest column: error-correcting codes and expander graphs . ACM SIGACT News . 35 . 3 . 25–41 . September 2004 . 10.1145/1027914.1027924 . 17550280 . subscription .