Johnson bound
In applied mathematics, the Johnson bound (named after Selmer Martin Johnson) is a limit on the size of error-correcting codes, as used in coding theory for data transmission or communications.
Definition
Let be a q-ary code of length
, i.e. a subset of
. Let
be the minimum distance of
, i.e.
where is the Hamming distance between
and
.
Let be the set of all q-ary codes with length
and minimum distance
and let
denote the set of codes in
such that every element has exactly
nonzero entries.
Denote by the number of elements in
. Then, we define
to be the largest size of a code with length
and minimum distance
:
Similarly, we define to be the largest size of a code in
:
Theorem 1 (Johnson bound for ):
If ,
If ,
Theorem 2 (Johnson bound for ):
(i) If
(ii) If , then define the variable
as follows. If
is even, then define
through the relation
; if
is odd, define
through the relation
. Let
. Then,
where is the floor function.
Remark: Plugging the bound of Theorem 2 into the bound of Theorem 1 produces a numerical upper bound on .
See also
- Singleton bound
- Hamming bound
- Plotkin bound
- Elias Bassalygo bound
- Gilbert–Varshamov bound
- Griesmer bound
References
- Johnson, Selmer Martin (April 1962). "A new upper bound for error-correcting codes". IRE Transactions on Information Theory: 203–207.
- Huffman, William Cary; Pless, Vera S. (2003). Fundamentals of Error-Correcting Codes. Cambridge University Press. ISBN 978-0-521-78280-7.