In coding theory, block codes are a large and important family of error-correcting codes that encode data in blocks.There is a vast number of examples for block codes, many of which have a wide range of practical applications. The abstract definition of block codes is conceptually useful because it allows coding theorists, mathematicians, and computer scientists to study the limitations of all block codes in a unified way.Such limitations often take the form of bounds that relate different parameters of the block code to each other, such as its rate and its ability to detect and correct errors.
Examples of block codes are Reed–Solomon codes, Hamming codes, Hadamard codes, Expander codes, Golay codes, Reed–Muller codes and Polar codes. These examples also belong to the class of linear codes, and hence they are called linear block codes. More particularly, these codes are known as algebraic block codes, or cyclic block codes, because they can be generated using boolean polynomials.
Algebraic block codes are typically hard-decoded using algebraic decoders.
The term block code may also refer to any error-correcting code that acts on a block of
k
n
(n,k)
This article deals with "algebraic block codes".
Error-correcting codes are used to reliably transmit digital data over unreliable communication channels subject to channel noise.When a sender wants to transmit a possibly very long data stream using a block code, the sender breaks the stream up into pieces of some fixed size. Each such piece is called message and the procedure given by the block code encodes each message individually into a codeword, also called a block in the context of block codes. The sender then transmits all blocks to the receiver, who can in turn use some decoding mechanism to (hopefully) recover the original messages from the possibly corrupted received blocks.The performance and success of the overall transmission depends on the parameters of the channel and the block code.
Formally, a block code is an injective mapping
C:\Sigmak\to\Sigman
\Sigma
k
n
The data stream to be encoded is modeled as a string over some alphabet
\Sigma
|\Sigma|
q
q=2
q
\Sigma
Fq
Messages are elements
m
\Sigmak
k
k
The block length
n
c
\Sigman
n
c=C(m)
m
c
m
The rate of a block code is defined as the ratio between its message length and its block length:
R=k/n
1-R
1
C
The distance or minimum distance of a block code is the minimum number of positions in which any two distinct codewords differ, and the relative distance
\delta
d/n
c1,c
n | |
2\in\Sigma |
\Delta(c1,c2)
c1
c2
c1
c2
d
C
d:=
min | |||||||||||||
|
\Delta[C(m1),C(m2)]
1
min | |||||||||||||
|
\Delta[C(m1),C(m2)]=
min | |||||||||||||
|
\Delta[0,C(m1)+C(m2)]=
min | |
m\in\Sigmak\atopm ≠ 0 |
w[C(m)]=wmin
A larger distance allows for more error correction and detection.For example, if we only consider errors that may change symbols of the sent codeword but never erase or add them, then the number of errors is the number of positions in which the sent codeword and the received word differ.A code with distance allows the receiver to detect up to
d-1
d-1
(d-1)/2
(d-1)/2
(d-1)/2
The notation
(n,k,d)q
\Sigma
q
n
k
d
[n,k,d]q
q=2
d=n-k+1
d
Sometimes, especially for non-block codes, the notation
(n,M,d)q
M
n
k
q
M=qk
As mentioned above, there are a vast number of error-correcting codes that are actually block codes.The first error-correcting code was the Hamming(7,4) code, developed by Richard W. Hamming in 1950. This code transforms a message consisting of 4 bits into a codeword of 7 bits by adding 3 parity bits. Hence this code is a block code. It turns out that it is also a linear code and that it has distance 3. In the shorthand notation above, this means that the Hamming(7,4) code is a
[7,4,3]2
Reed–Solomon codes are a family of
[n,k,d]q
d=n-k+1
q
[n,k,d]q
d\leqn-k+1
[n,k,d]2
n=2k-1
d=2k-2
A codeword
c\in\Sigman
n
\Sigman
l{C}
\Sigman
l{C}
d
\forallc\inl{C}
c
d-1
n
c
d-1
l{C}
d
l{C}
d-1
c
d-1
d-1
l{C}
l{C}
style\left\lfloor{{d-1}\over2}\right\rfloor
c
d-1
style\left\lfloor{{d-1}\over2}\right\rfloor
y
style\left\lfloor{{d-1}\over2}\right\rfloor
y
style\left\lfloor{{d-1}\over2}\right\rfloor
(d-1)/2
l{C}
d-1
q
ith
ith
style\left\lfloor{{d-1}\over2}\right\rfloor
C=\{Ci\}i\ge1
Ci
(ni,ki,di)q
ni
Rate of family of codes is defined as
R(C)=\limi\toinfty{ki\overni}
Relative distance of family of codes is defined as
\delta(C)=\limi\toinfty{di\overni}
To explore the relationship between
R(C)
\delta(C)
See main article: article and Hamming bound.
R\le1-{1\overn} ⋅ logq ⋅
\left\lfloor{{\delta ⋅ n-1 | |
\left[\sum | |
i=0 |
\over2}\right\rfloor}\binom{n}{i}(q-1)i\right]
See main article: article and Singleton bound. The Singleton bound is that the sum of the rate and the relative distance of a block code cannot be much larger than 1:
R+\delta\le1+
1 | |
n |
k+d\len+1
See main article: article and Plotkin bound. For
q=2
R+2\delta\le1
k+2d\len
For the general case, the following Plotkin bounds holds for any
C\subseteq
n | |
F | |
q |
d=\left(1-{1\overq}\right)n,|C|\le2qn
d>\left(1-{1\overq}\right)n,|C|\le{qd\over{qd-\left(q-1\right)n}}
For any -ary code with distance
\delta
R\le1-\left({q\over{q-1}}\right)\delta+o\left(1\right)
See main article: article and Gilbert–Varshamov bound.
R\ge1-Hq\left(\delta\right)-\epsilon
0\le\delta\le1-{1\overq},0\le\epsilon\le1-Hq\left(\delta\right)
Hq\left(x\right)~\overset{\underset{def
See main article: article and Johnson bound. Define
Jq\left(\delta\right)~\overset{\underset{def
Jq\left(n,d,e\right)
C\subseteq
n | |
F | |
q |
Then we have the Johnson Bound :
Jq\left(n,d,e\right)\leqnd
{e\overn}\le{{q-1}\overq}\left({1-\sqrt{1-{q\over{q-1}} ⋅ {d\overn}}}\right)=Jq\left({d\overn}\right)
See main article: article and Elias Bassalygo bound.
R={logq{|C|}\overn}\le1-Hq\left(Jq\left(\delta\right)\right)+o\left(1\right)
Block codes are tied to the sphere packing problem which has received some attention over the years. In two dimensions, it is easy to visualize. Take a bunch of pennies flat on the table and push them together. The result is a hexagon pattern like a bee's nest. But block codes rely on more dimensions which cannot easily be visualized. The powerful Golay code used in deep space communications uses 24 dimensions. If used as a binary code (which it usually is), the dimensions refer to the length of the codeword as defined above.
The theory of coding uses the N-dimensional sphere model. For example, how many pennies can be packed into a circle on a tabletop or in 3 dimensions, how many marbles can be packed into a globe. Other considerations enter the choice of a code. For example, hexagon packing into the constraint of a rectangular box will leave empty space at the corners. As the dimensions get larger, the percentage of empty space grows smaller. But at certain dimensions, the packing uses all the space and these codes are the so-called perfect codes. There are very few of these codes.
Another property is the number of neighbors a single codeword may have.[1] Again, consider pennies as an example. First we pack the pennies in a rectangular grid. Each penny will have 4 near neighbors (and 4 at the corners which are farther away). In a hexagon, each penny will have 6 near neighbors. Respectively, in three and four dimensions, the maximum packing is given by the 12-face and 24-cell with 12 and 24 neighbors, respectively. When we increase the dimensions, the number of near neighbors increases very rapidly. In general, the value is given by the kissing numbers.
The result is that the number of ways for noise to make the receiver choosea neighbor (hence an error) grows as well. This is a fundamental limitationof block codes, and indeed all codes. It may be harder to cause an error toa single neighbor, but the number of neighbors can be large enough so thetotal error probability actually suffers.[1]
. J.H. van Lint . Jack van Lint . Introduction to Coding Theory . 2nd . Springer-Verlag . . 86 . 1992 . 3-540-54894-7 . 31 .
. F.J. MacWilliams . Jessie MacWilliams . N.J.A. Sloane . Neil Sloane . The Theory of Error-Correcting Codes . registration . North-Holland . 1977 . 0-444-85193-3 . 35.