Skip to main content
Top
Published in:

Open Access 16-10-2024

Decoding error probability of random parity-check matrix ensemble over the erasure channel

Authors: Chin Hei Chan, Fang-Wei Fu, Maosheng Xiong

Published in: Designs, Codes and Cryptography | Issue 1/2025

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

The article delves into the decoding error probability of random parity-check matrix ensembles over the erasure channel, a crucial area in digital communication. It introduces the concept of error-correcting codes and their importance in mitigating channel noise. The focus is on the erasure channel, where symbols are either received correctly or totally erased. The study presents explicit formulas for the average decoding error probability under three decoding principles: unambiguous, maximum likelihood, and list decoding. Additionally, it computes error exponents for these decoding principles, providing a comprehensive analysis of the ensemble's performance. The paper also includes a strong concentration result for the unsuccessful decoding probability under unambiguous decoding. This research builds on previous work, offering stronger results and leveraging the fine algebraic structure of the ensemble. The explicit formulas and error exponents enable meaningful guidance in selecting good codes for the erasure channel, making this article a valuable resource for specialists in coding theory and error-correcting codes.
Notes
Communicated by L. Mérai.
The research of Fang-Wei Fu was supported in part by the National Key Research and Development Program of China under Grants 2022YFA1005000 and 2018YFA0704703, in part by the National Natural Science Foundation of China under Grants 12141108, 62371259, 12226336, in part by the Fundamental Research Funds for the Central Universities of China (Nankai University), and in part by the Nankai Zhide Foundation. The research of M. Xiong was supported by RGC grant number 16306520 from Hong Kong.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

1 Introduction

1.1 Background

In digital communication, it is common that messages transmitted through a public channel may be distorted by the channel noise. The theory of error-correcting codes is the study of mechanisms to cope with this problem. This is an important research area with many applications in modern life. For example, error-correcting codes are widely employed in cell phones to correct errors arising from fading noise during high frequency radio transmission. One of the major challenges in coding theory remains to construct new error-correcting codes with good properties and to study their decoding and encoding algorithms.
In a binary erasure channel (BEC), a binary symbol is either received correctly or totally erased with probability \(\varepsilon \). The concept of BEC was first introduced by Elias in 1955 [3]. Together with the binary symmetric channel (BSC), they are frequently used in coding theory and information theory because they are among the simplest channel models, and many problems in communication theory can be reduced to problems in a BEC. Here we consider more generally a q-ary erasure channel in which a q-ary symbol is either received correctly, or totally erased with probability \(\varepsilon \).
The problem of decoding linear codes over the erasure channel has received renewed attention in recent years due to their wide application in the internet and the distributed storage system in analyzing random packet losses [2, 11, 12]. Three important decoding principles, namely unambiguous decoding, maximum likelihood decoding and list decoding, were studied in recent years for linear codes over the erasure channel, the corresponding decoding error probabilities under these principles were also investigated (see [5, 9, 14, 17] and reference therein).
In particular in [14], upon improving previous results, the authors provided a detailed study on the decoding error probabilities of a general q-ary linear code over the erasure channel under the three decoding principles. Via the notion of \(q^\ell \)-incorrigible sets for linear codes, they showed that all these decoding error probabilities can be expressed explicitly by the r-th support weight distribution of the linear codes. As applications they obtained explicit formulas of the decoding error probabilities for some of the most interesting linear codes such as MDS codes, the binary Golay code, the simplex codes and the first-order Reed-Muller codes etc. where the r-support weight distributions were known. They also computed the average decoding error probabilities of a random \([n,k]_q\) code over the erasure channel and obtained the error exponent of a random \([n,nR]_q\) code (\(0<R<1\)) under unambiguous decoding. The error exponents of a random [nnR] code under list and maximum likelihood decoding were obtained in [18].

1.2 Statement of the main results

In this paper we consider a different code ensemble, namely the random parity-check matrix ensemble \(\mathcal {R}_{m,n}\), that is, the set of all \(m \times n\) matrices over \(\mathbb {F}_q\) endowed with uniform probability, each of which is associated with a parity-check code as follows: for each \(H \in \mathcal {R}_{m,n}\), the corresponding parity-check code \(C_H\) is given by
$$\begin{aligned} C_H= \left\{ \textbf{x} \in \mathbb {F}_q^n: H\textbf{x}^T=\textbf{0}\right\} . \end{aligned}$$
(1)
Here boldface letters such as \(\textbf{x}\) denote row vectors.
The ensemble \(\mathcal {R}_{m,n}\) has been studied for a long time and many strong results have been obtained. For example, in the classical work of Gallager [7], an upper bound of the average number of codewords of a given weight in \(\mathcal {R}_{m,n}\) was obtained, from which information about the minimum-distance distribution of \(\mathcal {R}_{m,n}\) can be derived (see [7, Theorems 2.1\(-\)2.2]); in [1] (see also [4]) union bounds on the block erasure probabilities of the ensemble \(\mathcal {R}_{m,n}\) was obtained under the maximum likelihood decoding (see [1, 4]). More recently, the undetected error probability of the ensemble \(\mathcal {R}_{m,n}\) was studied in the binary symmetric channel by Wadayama [16] (i.e. \(q=2\)), and some bounds on the error probability under the maximum likelihood decoding principle were obtained in the q-ary erasure channel [6, 10]. It is easy to see that \(\mathcal {R}_{m,n}\) contains all linear codes in the random \([n,n-m]_q\) code ensemble considered in [14], but these two ensembles are different for two reasons: first, in the random \([n,k]_q\) code ensemble considered in [14], each \([n,k]_q\) code is counted exactly once, while in \(\mathcal {R}_{m,n}\) each code is counted with some multiplicity as different choices for the matrix H may give rise to the same code; second, some codes in \(\mathcal {R}_{m,n}\) may have rates strictly larger than \(1-\frac{m}{n}\) as the rows of H may not be linearly independent.
It is conceivable that most of the codes in \(\mathcal {R}_{m,n}\) have rate \(1-\frac{m}{n}\), and the average behavior of codes in \(\mathcal {R}_{m,n}\) should be similar to that of the random \([n,n-m]_q\) code ensemble considered in [14]. We will show that this is indeed the case. Actually we will obtain stronger results, taking advantage of the fine algebraic structure of the ensemble \(\mathcal {R}_{m,n}\), which may not be readily available in the random \([n,k]_q\) code ensemble. Such structure has been exploited in [16].
We first obtain explicit formulas for the average decoding error probability of the ensemble \(\mathcal {R}_{m,n}\) over the erasure channel under the three different decoding principles. This is comparable to [14, Theorem 2] for the random \([n,k]_q\) code ensemble. Such formulas are useful as they allow explicit evaluations of the average decoding error probabilities for any given m and n, hence giving us a meaningful guidance as to what to expect for a good \([n,n-m]_q\) code over the erasure channel.
Theorem 1
Let \(\mathcal {R}_{m,n}\) be the random matrix ensemble described above. Denote by https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_IEq37_HTML.gif the Gaussian q-binomial coefficient and denote
$$\begin{aligned} \psi _m(i):=\prod _{k=0}^{i-1}(1-q^{k-m}), \quad 1 \le i \le m. \end{aligned}$$
(2)
1.
The average unsuccessful decoding probability of \(\mathcal {R}_{m,n}\) under list decoding with list size \(q^\ell \), where \(\ell \) is a non-negative integer, is given by
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ3_HTML.png
(3)
 
2.
The average unsuccessful decoding probability of \(\mathcal {R}_{m,n}\) under unambiguous decoding is given by
$$\begin{aligned} P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )=\sum _{i=1}^n\left( 1-\psi _m(i)\right) \left( {\begin{array}{c}n\\ i\end{array}}\right) \varepsilon ^i(1-\varepsilon )^{n-i}; \end{aligned}$$
(4)
 
3.
The average decoding error probability of \(\mathcal {R}_{m,n}\) under maximum likelihood decoding is given by
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ5_HTML.png
(5)
 
Remark 1
By using the elementary identity
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ126_HTML.png
we see that \(P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )=P_{\textrm{ld}}(\mathcal {R}_{m,n},0,\varepsilon )\). We list the formula for \(P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )\) here to emphasis the unambiguous decoding, which we will study further later.
Remark 2
In Table 1, we use Mathematica to compute the numerical values of \(P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon ), P_{\textrm{mld}}(\mathcal {R}_{m,n},\varepsilon )\) and \(P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon )\) (\(\ell =1,2\)) for \(m=50,n=100,q=2\) and \(\varepsilon =i*0.05\) with \(1 \le i \le 10\). It can be seen that when \(\varepsilon \) is much smaller than the rate of the code which is \(\frac{1}{2}\), the decoding error probabilities are all very small, in particular \(P_{\textrm{mld}}(\mathcal {R}_{m,n},\varepsilon )\) and \(P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )\) are of the same magnitude, and \(P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon )\) decreases even much further as \(\ell \) increases. The fact that these decoding error probabilities are small will be captured by the error exponents which we will study in Theorem 2.
Table 1
Decoding error probability for \(\mathcal {R}_{m,n}\) where \(m=50,n=100,q=2,\epsilon =i*0.05\) with \(1 \le i \le 10\)
\(\epsilon \)
\({\scriptstyle P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )}\)
\({\scriptstyle P_{\textrm{ld}}(\mathcal {R}_{m,n},1,\varepsilon )}\)
\({\scriptstyle P_{\textrm{ld}}(\mathcal {R}_{m,n},2,\varepsilon )}\)
\({\scriptstyle P_{\textrm{mld}}(\mathcal {R}_{m,n},\varepsilon )}\)
0.05
\({\scriptstyle 1.16 \times 10^{-13}}\)
\({\scriptstyle 1.54\times 10^{-25}}\)
\({\scriptstyle 4.50\times 10^{-35}}\)
\({\scriptstyle 5.80\times 10^{-14}}\)
0.10
\({\scriptstyle 1.22\times 10^{-11}}\)
\({\scriptstyle 3.26\times 10^{-20}}\)
\({\scriptstyle 3.83\times 10^{-25}}\)
\({\scriptstyle 6.12\times 10^{-12}}\)
0.15
\({\scriptstyle 1.04\times 10^{-9}}\)
\({\scriptstyle 1.74\times 10^{-15}}\)
\({\scriptstyle 6.59\times 10^{-18}}\)
\({\scriptstyle 5.21\times 10^{-10}}\)
0.20
\({\scriptstyle 7.35\times 10^{-8}}\)
\({\scriptstyle 2.02\times 10^{-11}}\)
\({\scriptstyle 8.03\times 10^{-13}}\)
\({\scriptstyle 3.67\times 10^{-8}}\)
0.25
\({\scriptstyle 4.26\times 10^{-6}}\)
\({\scriptstyle 3.43\times 10^{-8}}\)
\({\scriptstyle 3.95\times 10^{-9}}\)
\({\scriptstyle 2.14\times 10^{-6}}\)
0.30
\({\scriptstyle 1.78\times 10^{-4}}\)
\({\scriptstyle 1.01\times 10^{-5}}\)
\({\scriptstyle 2.16\times 10^{-6}}\)
\({\scriptstyle 9.20\times 10^{-5}}\)
0.35
\({\scriptstyle 4.12\times 10^{-3}}\)
\({\scriptstyle 7.03\times 10^{-4}}\)
\({\scriptstyle 2.33\times 10^{-4}}\)
\({\scriptstyle 2.27\times 10^{-3}}\)
0.40
\({\scriptstyle 0.044797}\)
\({\scriptstyle 0.015130}\)
\({\scriptstyle 7.06\times 10^{-3}}\)
\({\scriptstyle 0.027361}\)
0.45
\({\scriptstyle 0.226807}\)
\({\scriptstyle 0.121768}\)
\({\scriptstyle 0.074765}\)
\({\scriptstyle 0.15745}\)
0.50
\({\scriptstyle 0.580678}\)
\({\scriptstyle 0.429632}\)
\({\scriptstyle 0.328953}\)
\({\scriptstyle 0.463703}\)
Remark 3
To make a comparison of Theorem 1 with corresponding results for the random \([n,k]_q\) code ensemble, we take \(k=n-m\) and rewrite the explicit formulas obtained in [14, 18] as follows (see [14, Theorem 2] and [18, Section 6]): Denote by \(\mathcal {C}_{m,n}\) the random \([n,n-m]_q\) code ensemble.
1.
The average unsuccessful decoding probability of \(\mathcal {C}_{m,n}\) under list decoding with list size \(q^\ell \), where \(\ell \) is a non-negative integer, is given by
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ6_HTML.png
(6)
 
2.
The average unsuccessful decoding probability of \(\mathcal {C}_{m,n}\) under unambiguous decoding is given by
$$\begin{aligned} P_{\textrm{ud}}(\mathcal {C}_{m,n},\varepsilon )=\sum _{i=1}^n\left( 1-\frac{\psi _m(i)}{\psi _n(i)}\right) \left( {\begin{array}{c}n\\ i\end{array}}\right) \varepsilon ^i(1-\varepsilon )^{n-i};\end{aligned}$$
(7)
 
3.
The average decoding error probability of \(\mathcal {C}_{m,n}\) under maximum likelihood decoding is given by
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ8_HTML.png
(8)
 
It is easy to check that for all indices ij appearing in (3) and (6) we have \(0<\frac{\psi _{n-m}(j)}{\psi _n(i)}<1\), this shows that
$$\begin{aligned}P_{\textrm{ld}}(\mathcal {C}_{m,n},\ell ,\varepsilon )<P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon ).\end{aligned}$$
Similarly we also have
$$\begin{aligned}P_{\textrm{ud}}(\mathcal {C}_{m,n},\varepsilon )<P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon ),\quad P_{\textrm{mld}}(\mathcal {C}_{m,n},\varepsilon )<P_{\textrm{mld}}(\mathcal {R}_{m,n},\varepsilon ).\end{aligned}$$
With this respect, the random code ensemble \(\mathcal {C}_{m,n}\) behaves slightly better than \(\mathcal {R}_{m,n}\) in terms of the average decoding error probabilities. On the other hand, the difference between these two ensembles seems to be rather small: we use Mathematica to compute the numerical values of the corresponding decoding error probabilities for \(\mathcal {C}_{m,n}\) but these values appear exactly the same as those in Table 1. As it turns out, the difference of the decoding error probabilities between these two ensembles are much smaller in magnitude, as can be seen in Table 2.
Table 2
Difference of decoding error probability between \(\mathcal {R}_{m,n}\) and \(\mathcal {C}_{m,n}\) where \(m=50,n=100,q=2,\epsilon =i*0.05\) with \(1 \le i \le 10\)
\(\epsilon \)
\({\scriptstyle P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )}\)
\({\scriptstyle P_{\textrm{ld}}(\mathcal {R}_{m,n},1,\varepsilon )}\)
\({\scriptstyle P_{\textrm{ld}}(\mathcal {R}_{m,n},2,\varepsilon )}\)
\({\scriptstyle P_{\textrm{mld}}(\mathcal {R}_{m,n},\varepsilon )}\)
\({\scriptstyle -P_{\textrm{ud}}(\mathcal {C}_{m,n},\varepsilon )}\)
\({\scriptstyle -P_{\textrm{ld}}(\mathcal {C}_{m,n},1,\varepsilon )}\)
\({\scriptstyle -P_{\textrm{ld}}(\mathcal {C}_{m,n},2,\varepsilon )}\)
\({\scriptstyle -P_{\textrm{mld}}(\mathcal {C}_{m,n},\varepsilon )}\)
0.05
\({\scriptstyle 1.03\times 10^{-28}}\)
\({\scriptstyle 4.13\times 10^{-40}}\)
\({\scriptstyle 2.89\times 10^{-49}}\)
\({\scriptstyle 5.14\times 10^{-29}}\)
0.10
\({\scriptstyle 1.09\times 10^{-26}}\)
\({\scriptstyle 9.63\times 10^{-35}}\)
\({\scriptstyle 2.16\times 10^{-39}}\)
\({\scriptstyle 5.44\times 10^{-27}}\)
0.15
\({\scriptstyle 9.26\times 10^{-25}}\)
\({\scriptstyle 4.14\times 10^{-30}}\)
\({\scriptstyle 2.70\times 10^{-32}}\)
\({\scriptstyle 4.63\times 10^{-25}}\)
0.20
\({\scriptstyle 6.52\times 10^{-23}}\)
\({\scriptstyle 4.52\times 10^{-26}}\)
\({\scriptstyle 2.12\times 10^{-27}}\)
\({\scriptstyle 3.26\times 10^{-23}}\)
0.25
\({\scriptstyle 3.70\times 10^{-21}}\)
\({\scriptstyle 5.96\times 10^{-23}}\)
\({\scriptstyle 8.27\times 10^{-24}}\)
\({\scriptstyle 1.87\times 10^{-21}}\)
0.30
\({\scriptstyle 1.37\times 10^{-19}}\)
\({\scriptstyle 1.36\times 10^{-20}}\)
\({\scriptstyle 3.39\times 10^{-21}}\)
\({\scriptstyle 7.21\times 10^{-20}}\)
0.35
\({\scriptstyle 2.45\times 10^{-18}}\)
\({\scriptstyle 5.42\times 10^{-19}}\)
\({\scriptstyle 1.63\times 10^{-19}}\)
\({\scriptstyle 1.41\times 10^{-18}}\)
0.40
\({\scriptstyle 1.81\times 10^{-17}}\)
\({\scriptstyle 5.20\times 10^{-18}}\)
\({\scriptstyle 3.47\times 10^{-18}}\)
\({\scriptstyle 1.19\times 10^{-17}}\)
0.45
\({\scriptstyle 5.39\times 10^{-17}}\)
\({\scriptstyle 2.78\times 10^{-17}}\)
\({\scriptstyle 2.77\times 10^{-17}}\)
\({\scriptstyle 4.23\times 10^{-17}}\)
0.50
\({\scriptstyle 6.47\times 10^{-17}}\)
\({\scriptstyle 5.55\times 10^{-17}}\)
\({\scriptstyle 5.55\times 10^{-17}}\)
\({\scriptstyle 6.45\times 10^{-17}}\)
Next, letting \(m=(1-R)n\) for \(0< R < 1\), we compute the error exponents of the average decoding error probability of the ensemble series \(\{\mathcal {R}_{(1-R)n,n}\}\) as \(n \rightarrow \infty \) under these decoding principles.
Theorem 2
Let the rate \(0< R < 1\) be fixed and \(n \rightarrow \infty \).
1.
For any fixed integer \(\ell \ge 0\), the error exponent \(T_{\textrm{ld}}(\ell ,\varepsilon )\) for average unsuccessful decoding probability of \(\{\mathcal {R}_{(1-R)n,n}\}\) under list decoding with list size \(q^\ell \) is given by
$$\begin{aligned} T_{\textrm{ld}}(\ell ,\varepsilon )={\left\{ \begin{array}{ll} 0 & (1-\varepsilon \le R< 1)\\ (1-R)\log _q\left( \frac{1-R}{\varepsilon }\right) +R\log _q\left( \frac{R}{1-\varepsilon }\right) & \left( \frac{1-\varepsilon }{1-\varepsilon +q^{\ell +1}\varepsilon }< R< 1-\varepsilon \right) \\ (\ell +1)(1-R)-\log _q\left( 1-\varepsilon +q^{\ell +1}\varepsilon \right) & \left( 0 < R \le \frac{1-\varepsilon }{1-\varepsilon +q^{\ell +1}\varepsilon }\right) . \end{array}\right. } \end{aligned}$$
(9)
 
2.
The error exponents for average unsuccessful decoding probability of \(\{\mathcal {R}_{(1-R)n,n}\}\) under unambiguous decoding and maximum likelihood decoding (respectively) are both given by
$$\begin{aligned} T_{\textrm{ud}}(\varepsilon )=T_{\textrm{mld}}(\varepsilon )={\left\{ \begin{array}{ll} 0 & (1-\varepsilon \le R< 1)\\ (1-R)\log _q\left( \frac{1-R}{\varepsilon }\right) +R\log _q\left( \frac{R}{1-\varepsilon }\right) & \left( \frac{1-\varepsilon }{1-\varepsilon +q \varepsilon }< R< 1-\varepsilon \right) \\ 1-R-\log _q \left( 1-\varepsilon +q \varepsilon \right) & \left( 0 < R \le \frac{1-\varepsilon }{1-\varepsilon +q \varepsilon }\right) . \end{array}\right. } \end{aligned}$$
(10)
 
A plot of the function \(T_{\textrm{ld}}(\ell ,\varepsilon )\) for \(q=2,\varepsilon =0.25,\ell =0,1,2\) in the range \(0< R < 1\) is given by Fig. 1.
Fig. 1
The error exponent \(T_{\textrm{ld}}(\ell ,\varepsilon )\) for \(0<R<1\) where \(\ell =0,1,2\) and \(q=2, \varepsilon =0.25\)
Full size image
It turns out that the error exponents obtained here under these decoding principles are identical with those for the random \([n,nR]_q\) code ensemble obtained in [14, Theorem 3] and [18, Theorems 1.3 and 1.4].
We establish a strong concentration result for the unsuccessful decoding probability of a random code in the ensemble \(\mathcal {R}_{(1-R)n,n}\) towards the mean under unambiguous decoding.
Theorem 3
Let the rate \(0< R < 1\) be fixed and \(n \rightarrow \infty \). Then as \(H_n\) runs over the ensemble \(\mathcal {R}_{(1-R)n,n}\), we have
$$\begin{aligned} \frac{P_\textrm{ud}(H_n,\varepsilon )}{P_\textrm{ud}(\mathcal {R}_{(1-R)n,n},\varepsilon )} \rightarrow 1 \quad {\textbf{WHP}}, \end{aligned}$$
(11)
under either of the following conditions:
(1).
if \(\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2} \le R<1-\varepsilon \) for any \(q \ge 2\), or
 
(2).
if \(\frac{1-\varepsilon }{1+(q-1)\varepsilon } \le R<1-\varepsilon \) for \(q=2,3,4\).
 
Here the notion WHP in (11) refers to “with high probability”, that is, for any \(\delta >0\), there is \(c>0\) and \(n_0>0\) such that
$$\begin{aligned}\mathbb {P}\left( \left| \frac{P_\textrm{ud}(H_n,\varepsilon )}{P_\textrm{ud}(\mathcal {R}_{(1-R)n,n},\varepsilon )}-1\right| <\delta \right)>1-q^{-nc} \quad \forall n >n_0.\end{aligned}$$
Noting that in the range \(0<R<1-\varepsilon \), it was known that \(T_\textrm{ud}(\varepsilon ) > 0\) (see Theorem 2), hence
$$\begin{aligned} P_\textrm{ud}(\mathcal {R}_{(1-R)n,n},\varepsilon )=q^{-n(T_\textrm{ud}(\varepsilon )+o(1))} \rightarrow 0, \quad \text{ as } n \rightarrow \infty , \end{aligned}$$
so (11) shows that \(P_\textrm{ud}(H_n,\varepsilon )\) also tends to zero exponentially fast with high probability for the ensemble \(\mathcal {R}_{(1-R)n,n}\) under either Condition (1) or (2) of Theorem 3.
The paper is now organized as follows. In Sect. 2, we introduce the three decoding principles and the Gaussian q-binomial coefficient in more details. Then in Sect. 3, we provide three counting results regarding \(m \times n\) matrices of certain rank over \(\mathbb {F}_q\). Afterwards in Sects. 4, 5 and 6, we give the proofs of Theorems 1, 2 and 3 respectively. The proofs of Theorem 3 involves some technical calculus computations on the error exponent of the variance. In order to streamline the proofs, we put some of the arguments in Sect. 7 in Appendix. Finally we conclude this paper in Sect. 8.

2 Preliminaries

2.1 Three decoding principles, the average decoding error probability, and the error exponent

The three decoding principles have been well studied in the literature. For the sake of readers, we explain these terms here, following the presentation outlined in [14].
In a q-ary erasure channel, the channel input alphabet is a finite field \(\mathbb {F}_q\) of order q, and during transmission, each symbol \(x \in \mathbb {F}_q\) is either received correctly with probability \(1-\varepsilon \) or erased with probability \(\varepsilon \) \((0<\varepsilon <1)\).
Let C be a code of length n over \(\mathbb {F}_q\). For a codeword \(\textbf{c}=(c_1,c_2,\ldots ,c_n)\in C\) that was transmitted through the channel, suppose the word \(\textbf{r}=(r_1,r_2,\ldots ,r_n)\) is received. Denote by E the set of i’s such that \(r_i=\square \), i.e., the i-th symbol was erased during the transmission. In this case, we say that an erasure set E occurs. The probability P(E) that this erasure set E occurs for \(\textbf{c}\) is clearly \(\varepsilon ^{\#E}(1-\varepsilon )^{n-\#E}\). Here \(\#E\) denote the cardinality of the set E.
Let \(C(E,\textbf{r})\) be the set of all codewords of C that match the received word \(\textbf{r}\), that is,
$$\begin{aligned} C(E,\textbf{r})=\{\textbf{c} \in C: c_i=r_i \quad \forall i \in [n]\setminus E\}. \end{aligned}$$
Here \([n]=\{1,2,\ldots ,n\}\). The decoding problem is about how to choose one or more possible codewords in the set \(C(E,\textbf{r})\) when a word \(\textbf{r}\) is received. Now we consider three decoding principles:
  • In unambiguous decoding, the decoder outputs the only one codeword in \(C(E,\textbf{r})\) if \(\#C(E,\textbf{r})=1\) and declares “failure” otherwise. The unsuccessful decoding probability \(P_{\textrm{ud}}(C,\varepsilon )\) of C under unambiguous decoding is defined to be the probability that the decoder declares “failure”.
  • In list decoding, the decoder with list size \(q^{\ell }\) outputs all the codewords in \(C(E,\textbf{r})\) if \(\#C(E,\textbf{r})\le q^{\ell }\) and declares “failure” otherwise. The unsuccessful decoding probability \(P_{\textrm{ld}}(C,\ell ,\varepsilon )\) of C under list decoding with list size \(q^{\ell }\) is defined to be the probability that the decoder declares “failure”.
  • In maximum likelihood decoding, the decoder randomly choose a codeword in \(C(E,\textbf{r})\) uniformly and outputs this codeword. The decoding error probability \(P_{\textrm{mld}}(C,\varepsilon )\) of C under maximum likelihood decoding is defined to be the probability that the codeword outputted by the decoder is not the sent codeword.
The computation of \(P_{\textrm{ud}}(C,\varepsilon ), P_{\textrm{ld}}(C,\ell ,\varepsilon )\) and \(P_{\textrm{mld}}(C,\varepsilon )\) can be made much easier if C is a linear code.
Now assume that C is an \([n,k]_q\) linear code, that is, C is a k-dimensional subspace of \(\mathbb {F}_q^n\). For any \(E \subset [n]\), define
$$\begin{aligned}C(E):=\left\{ \textbf{c}=(c_1,\ldots ,c_n) \in C: c_i=0 \,\, \forall i \in [n] \setminus E\right\} .\end{aligned}$$
It is easy to see that if the set \(C(E,\textbf{r})\) is not empty, then the cardinality of \(C(E,\textbf{r})\) is the same as that of C(E), which is a vector space over \(\mathbb {F}_q\). Denote by \(\{I_i^{(\ell )}(C)\}_{i=1}^{n}\) the \(q^{\ell }\)-incorrigible set distribution of C, and \(\{I_i(C)\}_{i=1}^{n}\) the incorrigible set distribution of C, which are defined respectively as follows:
$$\begin{aligned} I_i^{(\ell )}(C)= & \#\left\{ E \subset [n]: \#E=i, \dim C(E) > \ell \right\} ,\nonumber \\ I_i(C)= & \#\left\{ E \subset [n]: \#E=i, C(E) \ne \{0\}\right\} . \end{aligned}$$
(12)
Here \(\dim C(E)\) denotes the dimension of C(E) as a vector space over \(\mathbb {F}_q\). We see that \(\dim C(E) \le \min \{k, \# E\}\), so if \(\ell \ge \min \{k, \# E\}\), then \(I_i^{(\ell )}(C)=\emptyset \). We also define
$$\begin{aligned} \lambda _i^{(\ell )}(C)=\#\left\{ E \subset [n]: \#E=i, \dim C(E) =\ell \right\} . \end{aligned}$$
(13)
It is easy to see that \(I_i^{(0)}(C)=I_i(C)\), \(\lambda _i^{(\ell )}(C)=0\) if \(\ell >\min \{i,k\}\) and
$$\begin{aligned} \sum _{\ell =0}^{i}\lambda _i^{(\ell )}(C)=\left( {\begin{array}{c}n\\ i\end{array}}\right) . \end{aligned}$$
(14)
We also have the identity
$$\begin{aligned} I^{(\ell )}_i(C)=\sum _{j=\ell +1}^{i}\lambda _i^{(j)}(C), \quad \forall i, \ell . \end{aligned}$$
(15)
Recall from [14] that the values \(P_{\textrm{ud}}(C,\varepsilon ), P_{\textrm{mld}}(C,\varepsilon )\) and \(P_{\textrm{ld}}(C,\ell ,\varepsilon )\) can all be expressed in terms of \(I_i^{(\ell )}(C)\), \(I_i(C)\) and \(\lambda _i^{(\ell )}(C)\) as follows:
$$\begin{aligned} P_{\textrm{ud}}(C,\varepsilon )= & \sum _{i=1}^n I_i(C) \varepsilon ^i (1-\varepsilon )^{n-i}, \\ P_{\textrm{ld}}(C,\ell ,\varepsilon )= & \sum _{i=1}^n I_i^{(\ell )}(C) \varepsilon ^i (1-\varepsilon )^{n-i}, \\ P_{\textrm{mld}}(C,\varepsilon )= & \sum _{i=1}^n \sum _{\ell =1}^i \lambda _i^{(\ell )}(C) (1-q^{-\ell })\varepsilon ^i (1-\varepsilon )^{n-i}. \end{aligned}$$
For \(H \in \mathcal {R}_{m,n}\), we write \(P_\textrm{ld}(H,\ell ,\varepsilon ):=P_\textrm{ld}(C_H,\ell ,\varepsilon )\) and \(P_*(H,\varepsilon ):=P_*(C_H,\varepsilon )\) for \(* \in \{\textrm{ud}, \textrm{mld}\}\), where \(C_H\) is the parity-check code defined by (1). The average decoding error probabilities over the ensemble \(\mathcal {R}_{m,n}\) are given by
$$\begin{aligned} P_\textrm{ld}(\mathcal {R}_{m,n},\ell ,\varepsilon ):=\mathbb {E}[P_\textrm{ld}(H,\ell ,\varepsilon )], \end{aligned}$$
and
$$\begin{aligned} P_*(\mathcal {R}_{m,n},\varepsilon ):=\mathbb {E}[P_*(H,\varepsilon )], \quad * \in \{\textrm{ud}, \textrm{mld}\}. \end{aligned}$$
Here the expectation \(\mathbb {E}\) is taken over the ensemble \(\mathcal {R}_{m,n}\).
Let \(0<R, \varepsilon <1\) be fixed constants. The error exponent \(T_{\textrm{ud}}(\varepsilon )\) for average unsuccessful decoding probability of the family of ensembles \(\{\mathcal {R}_{(1-R)n,n}\}\) under unambiguous decoding is defined as
$$\begin{aligned} T_{\textrm{ud}}(\varepsilon ):=-\lim _{n \rightarrow \infty } \frac{1}{n} \log _q P_{\textrm{ud}}(\mathcal {R}_{(1-R)n,n},\varepsilon ), \end{aligned}$$
provided that the limit exists [8, 14, 15]. The error exponents of \(\{\mathcal {R}_{(1-R)n,n}\}\) for the other two decoding principles are defined similarly.

2.2 Gaussian binomial coefficients

For integers \(n \ge k \ge 0\), the Gaussian binomial coefficients https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_IEq272_HTML.gif is defined as
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ127_HTML.png
where \((q)_n:=\prod _{i=1}^n \left( 1-q^i\right) \). By convention \((q)_0=1\), https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_IEq275_HTML.gif for any \(n \ge 0\) and https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_IEq277_HTML.gif if \(k<0\) or \(k>n\). The function \(\psi _m(i)\) defined in (2) can be written as
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ128_HTML.png
We may define \(\psi _m(0)=1\) for \(m \ge 0\) and \(\psi _m(i)=0\) if \(i<0\) or \(i>m\). Next, recall the well-known combinatorial interpretation of https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_IEq286_HTML.gif :
Lemma 1
([13]) The number of k-dimensional subspaces of an n-dimensional vector space over \(\mathbb {F}_q\) is https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_IEq288_HTML.gif .
The Gaussian binomial coefficient https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_IEq289_HTML.gif satisfies the property
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ129_HTML.png
and the identity
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ16_HTML.png
(16)

3 Three counting results for the ensemble \(\mathcal {R}_{m,n}\)

In this section we provide three counting results about matrices of certain rank in the ensemble \(\mathcal {R}_{m,n}\). Such results may not be new, but since it is not easy to locate them in the literature, we prove them here. These results will be used repeatedly in the proofs later on.
For \(H \in \mathcal {R}_{m,n}\), denote by \(\textrm{rk}(H)\) the rank of the matrix H over \(\mathbb {F}_q\).
Lemma 2
Let H be a random matrix in the ensemble \(\mathcal {R}_{m,n}\). Then for any integer j, we have
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ17_HTML.png
(17)
Proof
We may assume that j satisfies \(0 \le j \le n\), because if j is not in the range, then both sides of Equation (17) are obviously zero.
Denote by \(\textrm{Hom}(m,n)\) the set of \(\mathbb {F}_q\)-linear transformations from \(\mathbb {F}_q^m\) to \(\mathbb {F}_q^n\). Writing vectors in \(\mathbb {F}_q^m\) and \(\mathbb {F}_q^n\) as row vectors, we see that the random matrix ensemble \(\mathcal {R}_{m,n}\) can be identified with the set \(\textrm{Hom}(m,n)\) via the relation
$$\begin{aligned} H \,\,\, \leftrightarrow \,\,\, G : \begin{array}{c}\mathbb {F}_q^m \rightarrow \mathbb {F}_q^n\\ \textbf{x} \mapsto \textbf{x}H \end{array}. \end{aligned}$$
(18)
Since \(\textrm{rk}(H)=j\) if and only if \(\dim (\textrm{Im}G)=j\), and \(\#\mathcal {R}_{m,n}=q^{mn}\), we have
$$\begin{aligned} \mathbb {P}(\textrm{rk}(H)=j)&=q^{-mn}\sum _{\begin{array}{c} G \in \textrm{Hom}(m,n) \\ \dim (\textrm{Im}G)=j \end{array}} 1\\&=q^{-mn}\sum _{\begin{array}{c} V \le \mathbb {F}_q^n \\ \dim V=j \end{array}}\sum _{\begin{array}{c} G \in \textrm{Hom}(m,n) \\ \textrm{Im}G=V \end{array}} 1\, . \end{aligned}$$
The inner sum \(\sum _{G \in \textrm{Hom}(m,n),\textrm{Im}G=V} 1\) counts the number of surjective linear transformations from \(\mathbb {F}_q^m\) to V, a j-dimensional subspace of \(\mathbb {F}_q^n\). Since \(V \cong \mathbb {F}_q^j\), this is also the number of surjective linear transformations from \(\mathbb {F}_q^m\) to \(\mathbb {F}_q^j\), or, equivalently, the number of \(m \times j\) matrices K over \(\mathbb {F}_q\) such that the columns of K are linearly independent. The number of such matrices K can be counted as follows: the first column of K can be any nonzero vector over \(\mathbb {F}_q\), there are \(q^m-1\) choices; given the first column, the second column can be any vector lying outside the space of scalar multiples of the first column, so there are \(q^m-q\) choices; inductively, given the first k columns, the \((k+1)\)-th column lies outside a k-dimensional subspace, so the number of choices for the \((k+1)\)-th column is \(q^m-q^k\). Thus we have
$$\begin{aligned} \sum _{\begin{array}{c} G \in \textrm{Hom}_q(m,n) \\ \textrm{Im}G=V \end{array}} 1 =\prod _{k=0}^{j-1} (q^m-q^k) =q^{mj}\psi _m(j), \quad \dim V=j. \end{aligned}$$
(19)
Together with Lemma 1, we obtain
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ130_HTML.png
which is the desired result. \(\square \)
Lemma 3
Let H be a random matrix in the ensemble \(\mathcal {R}_{m,n}\). Let \(A \subset [n]:=\{1,2,\ldots ,n\}\) be a subset with cardinality s. Denote by \(H_A\) the \(m \times s\) submatrix formed by columns of H indexed from A. Then for any integers j and r, we have
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ20_HTML.png
(20)
Proof
We may assume that \(0 \le j \le n\) and \(\max \{0,s-n+j\} \le r \le \min \{j,s\}\), because if j or r does not satisfy this condition, then both sides of Equation (20) are zero.
Using the relations (18) and (19), we can expand the term \(\mathbb {P}(\textrm{rk}(H)=j \cap \textrm{rk}(H_A)=r)\) as
$$\begin{aligned} \mathbb {P}(\textrm{rk}(H)=j \cap \textrm{rk}(H_A)=r) =q^{-mn}\sum _{\begin{array}{c} V \le \mathbb {F}_q^n\\ \dim V=j \\ \dim V_A=r \end{array}} \sum _{\begin{array}{c} G \in \textrm{Hom}(m,n)\\ \textrm{Im} G=V \end{array}} 1=q^{-m(n-j)}\psi _m(j)\sum _{\begin{array}{c} V \le \mathbb {F}_q^n\\ \dim V=j \\ \dim V_A=r \end{array}} 1. \end{aligned}$$
(21)
Here \(V_A\) is the subspace of \(\mathbb {F}_q^A\) formed by restricting V to coordinates with indices from A. We may consider the projection given by
$$\begin{aligned} \pi _A:&V \rightarrow V_A\\&(v_k)_{k=1}^n \mapsto (v_k)_{k \in A}\,. \end{aligned}$$
The kernel of \(\pi _A\) has dimension \(j-r\) and is of the form \(W \times \{(0)_A\}\) for some subspace \(W \le \mathbb {F}_q^{[n]-A}\). So we can further decompose the sum on the right hand side of (21) as
$$\begin{aligned} \sum _{\begin{array}{c} V \le \mathbb {F}_q^n\\ \dim V=j \\ \dim V_A=r \end{array}} 1=\sum _{\begin{array}{c} W \le \mathbb {F}_q^{[n]-A}\\ \dim W=j-r \end{array}} \sum _{\begin{array}{c} V \le \mathbb {F}_q^n \\ \dim V=j \\ \ker (\pi _A)=W \times \{(0)_A\} \end{array}} 1. \end{aligned}$$
(22)
Now we compute the inner sum on the right hand side of (22). Suppose we are given an ordered basis of the \((j-r)\)-dimensional subspace \(W \times \{(0)_A\}\) of \(\mathbb {F}_q^n\). We extend it to an ordered basis of some j-dimensional subspace V as follows: first we need r other basis vectors \(\textbf{v}_1,\textbf{v}_2,\ldots ,\textbf{v}_r\) to be linearly independent. At the same time, they have to be linearly independent with any nonzero vector in \(\mathbb {F}_q^{[n]-A} \times \{(0)_A\}\) due to the kernel condition. This requires the set \(\{\pi _A(\textbf{v}_1), \pi _A(\textbf{v}_2), \ldots , \pi _A(\textbf{v}_r)\}\) to be linearly independent in \(\mathbb {F}_q^A\). On the other hand, if this condition is satisfied, then the r vectors \(\textbf{v}_1,\textbf{v}_2,\ldots ,\textbf{v}_r\) are also linearly independent with one another as well as with any nonzero vector in \(W \times \{(0)_A\}\). Therefore it reduces to counting the number of ordered linearly independent sets of r vectors in \(\mathbb {F}_q^A\). This number is clearly given by \(\prod _{k=0}^{r-1}(q^s-q^k)\), so the total number of different ordered bases is given by \(q^{r(n-s)}\prod _{k=0}^{r-1}(q^s-q^k)\).
On the other hand, given a fixed j-dimensional subspace V with \(\ker (\pi _A)=W \times \{(0)_A\}\), we count the number of ordered bases of V of the form stated in previous paragraph as follows: we choose \(\textbf{v}_1\) to be any vector in V but not in \(W \times \{(0)_A\}\), which gives \(q^j-q^{j-r}\) many choices for \(\textbf{v}_1\); similarly \(\textbf{v}_2\) is any vector in V but not in the span of \(W \times \{(0)_A\}\) and \(\textbf{v}_1\), this gives us \(q^j-q^{j-r+1}\) many choices for \(\textbf{v}_2\); using this argument, we see that the number of such ordered bases is given by \(\prod _{k=0}^{r-1} (q^j-q^{j-r+k})\).
We conclude from the above arguments that
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ131_HTML.png
Putting this into the right hand side of (22) and using Lemma 1 again, we get
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ132_HTML.png
The desired result is obtained immediately by plugging this into the right hand side of (21). This completes the proof of Lemma 3. \(\square \)
Lemma 4
Let H be a random matrix in the ensemble \(\mathcal {R}_{m,n}\). Let \(E,E'\) be subsets of [n] such that
$$\begin{aligned}i=\#E, \quad i'=\#E', \quad s=\#(E \cap E').\end{aligned}$$
Then
$$\begin{aligned} \mathbb {P}(\textrm{rk}(H_E)=i \cap \textrm{rk}(H_{E'})=i')= \frac{\psi _m(i)\psi _m(i')}{\psi _m(s)}. \end{aligned}$$
(23)
Proof
It is clear that if a matrix H has full rank, then so is the submatrix \(H_A\) for any index subset A. Hence we have
$$\begin{aligned} \mathbb {P}(\textrm{rk}(H_E)&=i \cap \textrm{rk}(H_{E'})=i')\\&=\mathbb {P}(\textrm{rk}(H_E)=i \cap \textrm{rk}(H_{E'})=i' | \textrm{rk}(H_{E \cap E'})=s)\mathbb {P}(\textrm{rk}(H_{E \cap E'})=s). \end{aligned}$$
It is easy to see that the two events \(\textrm{rk}(H_E)=i\) and \(\textrm{rk}(H_{E'})=i'\) are conditionally independent given \(\textrm{rk}(H_{E \cap E'})=s\), since columns of \(H_E\) and \(H_{E'}\) are independent as random vectors over \(\mathbb {F}_q\). Hence we get
$$\begin{aligned}&\mathbb {P}(\textrm{rk}(H_E)=i \cap \textrm{rk}(H_{E'})=i')\\&\quad =\mathbb {P}(\textrm{rk}(H_E)=i | \textrm{rk}(H_{E \cap E'})=s)\mathbb {P}(\textrm{rk}(H_E)=i' | \textrm{rk}(H_{E \cap E'})=s)\mathbb {P}(\textrm{rk}(H_{E \cap E'})=s)\\&\quad =\frac{\mathbb {P}(\textrm{rk}(H_E)=i \cap \textrm{rk}(H_{E \cap E'})=s)\mathbb {P}(\textrm{rk}(H_E)=i' \cap \textrm{rk}(H_{E \cap E'})=s)}{\mathbb {P}(\textrm{rk}(H_{E \cap E'})=s)}\\&\quad =\frac{\psi _m(i)\psi _m(i')}{\psi _m(s)}. \end{aligned}$$
Here we have applied Lemmas 2 and 3 in the last equality with \(j=i, j'=i'\) and \(r=s, H=H_E\) and \(A=E \cap E'\). \(\square \)

4 Proof of Theorem 1

Let C be an \([n,k]_q\) linear code. Recall from Sect. 2 that the values \(P_{\textrm{ud}}(C,\varepsilon ), P_{\textrm{mld}}(C,\varepsilon )\) and \(P_{\textrm{ld}}(C,\ell ,\varepsilon )\) can all be expressed explicitly as
$$\begin{aligned} P_{\textrm{ud}}(C,\varepsilon )= & \sum _{i=1}^n I_i(C) \varepsilon ^i (1-\varepsilon )^{n-i}, \end{aligned}$$
(24)
$$\begin{aligned} P_{\textrm{ld}}(C,\ell ,\varepsilon )= & \sum _{i=1}^n I_i^{(\ell )}(C) \varepsilon ^i (1-\varepsilon )^{n-i}, \end{aligned}$$
(25)
$$\begin{aligned} P_{\textrm{mld}}(C,\varepsilon )= & \sum _{i=1}^n \sum _{\ell =1}^i \lambda _i^{(\ell )}(C) (1-q^{-\ell })\varepsilon ^i (1-\varepsilon )^{n-i}, \end{aligned}$$
(26)
where \(I_i(C)\) and \(I_i^{(\ell )}(C)\) are the incorrigible set distribution of C and the \(q^{\ell }\)-incorrigible set distribution of C respectively, and \(\lambda _i^{(\ell )}(C)\) is defined in (13) also in Sect. 2.
Now we can start the proof of Theorem 1. For \(H \in \mathcal {R}_{m,n}\), we denote
$$\begin{aligned} I_i:=I_i(C_H),\,I_i^{(\ell )}:=I_i^{(\ell )}(C_H),\,\lambda _i^{(\ell )}:=\lambda _i^{(\ell )}(C_H). \end{aligned}$$
Taking expectations on both sides of Equations (24)-(26) over the ensemble \(\mathcal {R}_{m,n}\), we obtain
$$\begin{aligned} P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )= & \sum _{i=1}^n \mathbb {E}[I_i] \varepsilon ^i (1-\varepsilon )^{n-i}, \end{aligned}$$
(27)
$$\begin{aligned} P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon )= & \sum _{i=1}^n \mathbb {E}[I_i^{(\ell )}] \varepsilon ^i (1-\varepsilon )^{n-i}, \end{aligned}$$
(28)
$$\begin{aligned} P_{\textrm{mld}}(\mathcal {R}_{m,n},\varepsilon )= & \sum _{i=1}^n \sum _{\ell =1}^i \mathbb {E}[\lambda _i^{(\ell )}] (1-q^{-\ell })\varepsilon ^i (1-\varepsilon )^{n-i}. \end{aligned}$$
(29)
We now compute \(\mathbb {E}[\lambda _i^{(\ell )}]\). Noting that for \(H \in \mathcal {R}_{m,n}\) and \(E \subset [n]\), we have \(\textrm{rk}(H_E)+\dim C_H(E)=\#E\), thus
$$\begin{aligned} \mathbb {E}[\lambda _i^{(\ell )}]= & \frac{1}{\#\mathcal {R}_{m,n}} \sum _{\begin{array}{c} H \in \mathcal {R}_{m,n} \end{array}} \sum _{\begin{array}{c} E \subset [n]\\ \#E=i \end{array}} \mathbb {1}_{\dim C_H(E)=\ell }\\= & \frac{1}{\#\mathcal {R}_{m,n}} \sum _{\begin{array}{c} E \subset [n]\\ \#E=i \end{array}} \sum _{\begin{array}{c} H \in \mathcal {R}_{m,n}\\ \textrm{rk}(H_E)=i-\ell \end{array}} 1. \end{aligned}$$
By the symmetry of the ensemble \(\mathcal {R}_{m,n}\), the inner sum on the right hand side depends only on the cardinality of E, so we may assume \(E=[i]\) to obtain
$$\begin{aligned} \mathbb {E}[\lambda _i^{(\ell )}]=\frac{1}{\#\mathcal {R}_{m,i}} \left( {\begin{array}{c}n\\ i\end{array}}\right) \sum _{\begin{array}{c} H \in \mathcal {R}_{m,i}\\ \textrm{rk}(H)=i-\ell \end{array}}1. \end{aligned}$$
The right hand side is exactly \(\left( {\begin{array}{c}n\\ i\end{array}}\right) \mathbb {P}(\textrm{rk}(H)=i-\ell )\) where the probability is over the ensemble \(\mathcal {R}_{m,i}\). So from Lemma 2 we have
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ30_HTML.png
(30)
Using this and (15), we also obtain
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ133_HTML.png
Inserting the above values \(\mathbb {E}[I_i^{(\ell )}]\) and \(\mathbb {E}[\lambda _i^{(\ell )}]\) into (28) and (29) respectively, we obtain explicit expressions of \(P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon )\) and \(P_{\textrm{mld}}(\mathcal {R}_{m,n},\varepsilon )\), which agree with (3) and (5) of Theorem 1.
As for \(P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )\), noting by (14) that \(\sum _{\ell =0}^i \mathbb {E}[\lambda _i^{(\ell )}]=\left( {\begin{array}{c}n\\ i\end{array}}\right) \) and by (30) that \(\mathbb {E}[\lambda _i^{(0)}]=\psi _m(i)\left( {\begin{array}{c}n\\ i\end{array}}\right) \), therefore
$$\begin{aligned} \mathbb {E}[I_i]=\sum _{j=1}^i \mathbb {E}[\lambda _i^{(j)}]=\left( {\begin{array}{c}n\\ i\end{array}}\right) -\mathbb {E}[\lambda _i^{(0)}]=\left( 1-\psi _m(i)\right) \left( {\begin{array}{c}n\\ i\end{array}}\right) . \end{aligned}$$
(31)
Inserting this value into (27), we also obtain the desired expression of \(P_{\textrm{ud}}(\mathcal {R}_{m,n},\varepsilon )\). This completes the proof of Theorem 1.

5 Proof of Theorem 2

First recall that the error exponents of the average decoding error probability of the ensemble \(\mathcal {R}_{(1-R)n,n}\) over the erasure channel under the three decoding principles are defined by
$$\begin{aligned} T_{\textrm{ld}}(\ell ,\varepsilon ):=-\lim _{n \rightarrow \infty }\frac{1}{n}\log _q P_{\textrm{ld}}(\mathcal {R}_{(1-R)n,n},\ell ,\varepsilon ), \end{aligned}$$
(32)
and
$$\begin{aligned} T_*(\varepsilon ):=-\lim _{n \rightarrow \infty }\frac{1}{n}\log _q P_*(\mathcal {R}_{(1-R)n,n},\varepsilon ), \quad * \in \{\textrm{ud}, \textrm{mld}\}, \end{aligned}$$
provided that the three limits exist [8, 14, 15].
Unambiguous decoding corresponds to list decoding with \(\ell =0\), and it is also easy to see that
$$\begin{aligned}\frac{1}{2}P_{\textrm{ud}}(C,\varepsilon ) \le P_{\textrm{mld}}(C,\varepsilon ) \le P_{\textrm{ud}}(C,\varepsilon ), \end{aligned}$$
hence we have
$$\begin{aligned}T_{\textrm{ud}}(\varepsilon )=T_{\textrm{mld}}(\varepsilon )=T_{\textrm{ld}}(0,\varepsilon ),\end{aligned}$$
if the limit exists say in (32) for \(\ell =0\). So we only need to prove Part 1) of Theorem 2 for the case of list decoding.
Write \(m=(1-R)n\), and we define
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ134_HTML.png
It is easy to see that \(f_{i,j} \ge 0\) for all integers ij. In addition, \(f_{i,j} \ne 0\) if and only if \(1 \le i \le n, 1 \le j \le i\) and \(i-j \le m\).
We can rewrite (3) as
$$\begin{aligned} P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon )=\sum _{i,j} f_{i,j}, \end{aligned}$$
where ij are integers satisfying the conditions
$$\begin{aligned} \ell +1 \le i \le n, \quad \max \{i-m,\ell +1\} \le j \le i. \end{aligned}$$
(33)
The number of such integer pairs (ij) is at most mn. Noting that \(\lim _{n \rightarrow \infty } \frac{\log _q(mn)}{n}=0\), we have
$$\begin{aligned} \frac{1}{n}\log _q P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon )=o(1)+\frac{1}{n}\max _{i,j}\log _q f_{i,j}. \end{aligned}$$
(34)
Now we focus on the quantity \(f_{i,j}\). First, set \(T=q^{-1}\) so that \(0< T < 1\). It is easy to verify that for \(0 \le j \le i\),
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ35_HTML.png
(35)
Hence we have
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ135_HTML.png
The infinite product \((T)_\infty :=\prod _{r=1}^\infty (1-T^r)\) converges absolutely to some positive real number M which only depends on q, and \(M=(T)_\infty < (T)_m \le 1\) for any m. This implies that
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ136_HTML.png
and thus
https://static-content.springer.com/image/art%3A10.1007%2Fs10623-024-01516-5/MediaObjects/10623_2024_1516_Equ137_HTML.png
Therefore we have
$$\begin{aligned} \frac{1}{n}\log _q f_{i,j}=-\frac{j(m-i+j)}{n}+\frac{1}{n}\log _q \left( {\begin{array}{c}n\\ i\end{array}}\right) +\frac{i}{n}\log _q \varepsilon +\frac{n-i}{n}\log _q(1-\varepsilon )+o(1). \end{aligned}$$
We want to maximize this quantity over ij satisfying (33). Since the term \(-\frac{j(m-i+j)}{n}\) is always non-positive, it is easy to see that for any fixed i, to maximize the term \(\frac{1}{n}\log _q f_{i,j}\), we shall take
$$\begin{aligned}j=\left\{ \begin{array}{lcl} \ell +1 & :& \text{ if } \ell +1 > i-m, \text{ or } \text{ equivalently } \text{ if } i < m+\ell +1, \\ i-m & :& \text{ otherwise }. \end{array}\right. \end{aligned}$$
So we can simplify (34) as
$$\begin{aligned}&\frac{1}{n} \log _q P_{\textrm{ld}}(\mathcal {R}_{m,n},\ell ,\varepsilon )\\&\quad =o(1)+\max _{\ell +1 \le i \le n}\left[ -\frac{(\ell +1)(m-i+\ell +1)_+}{n}\right. \\&\qquad \left. +\frac{1}{n}\log _q \left( {\begin{array}{c}n\\ i\end{array}}\right) +\frac{i}{n}\log _q \varepsilon +\frac{n-i}{n}\log _q(1-\varepsilon )\right] , \end{aligned}$$
where \((x)_+:=\max \{0,x\}\).
Let \(i=tn\) with \(0 < t \le 1\). Using the result (see [13] and [14, Lemma 3] for example)
$$\begin{aligned} \frac{1}{n}\log _q \left( {\begin{array}{c}n\\ tn\end{array}}\right) =h(t)+o(1), \end{aligned}$$
(36)
where \(h(t)=-t\log _q t-(1-t)\log _q(1-t)\) is the binary entropy function (in q-its), and taking \(n \rightarrow \infty \), we obtain
$$\begin{aligned} T_{\textrm{ld}}(\ell ,\varepsilon )=-\sup _{0 < t \le 1} f(t), \end{aligned}$$
where
$$\begin{aligned} f(t)=-(\ell +1)(1-R-t)_+ +h(t)+t\log _q\varepsilon +(1-t)\log _q(1-\varepsilon ). \end{aligned}$$
(37)
Note that f(t) is continuous on the interval (0, 1).
Differentiating (37) with respect to t, we get
$$\begin{aligned} \frac{\textrm{d}f}{\textrm{d}t}={\left\{ \begin{array}{ll} \ell +1+\log _q\left( \frac{\varepsilon }{t}\right) -\log _q\left( \frac{1-\varepsilon }{1-t}\right) & (0< t< 1-R)\\ \log _q\left( \frac{\varepsilon }{t}\right) -\log _q\left( \frac{1-\varepsilon }{1-t}\right) & (1-R< t < 1). \end{array}\right. } \end{aligned}$$
It is easy to check from the right hand side that \(\frac{\textrm{d}f}{\textrm{d}t}=0\) at
$$\begin{aligned} t_1=\frac{q^{\ell +1}\varepsilon }{1+(q^{\ell +1}-1)\varepsilon }, \quad t_2=\varepsilon , \end{aligned}$$
and at these points the function f(t) has a local maximum. There are three cases to consider:
Case 1: \(0 < R \le \frac{1-\varepsilon }{1+(q^{\ell +1}-1)\varepsilon }\)
In this case we have \(\varepsilon < t_1 \le 1-R\), then f(t) is maximized at \(t=t_1\), so
$$\begin{aligned} T_{\textrm{ld}}(\ell ,\varepsilon )=-f(t_1)=(\ell +1)(1-R)-\log _q[1+(q^{\ell +1}-1)\varepsilon ]. \end{aligned}$$
Case 2: \(\frac{1-\varepsilon }{1+(q^{\ell +1}-1)\varepsilon }< R < 1-\varepsilon \)
In this case we have \(\varepsilon< 1-R < t_1\), then f(t) is maximized at \(t=1-R\), so
$$\begin{aligned} T_{\textrm{ld}}(\ell ,\varepsilon )=-f(1-R)=(1-R)\log _q\left( \frac{1-R}{\varepsilon }\right) +R\log _q\left( \frac{R}{1-\varepsilon }\right) . \end{aligned}$$
Case 3: \(1-\varepsilon \le R < 1\)
In this case we get \(1-R \le \varepsilon < t_1\), then f(t) is maximized at \(t=t_2=\varepsilon \), so
$$\begin{aligned} T_{\textrm{ld}}(\ell ,\varepsilon )=-f(\varepsilon )=0. \end{aligned}$$
Combining Cases 1–3 above gives Equation (9). This completes the proof of Theorem 2.

6 Proof of Theorem 3

The proof of Theorem 3 depend on the computation of the variance of the unsuccessful decoding probability under unambiguous decoding and its error exponent.

6.1 The variance of unsuccessful decoding probability and its error exponent

Note from (24) that the variance \(\sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )\) of the unsuccessful decoding probability under unambiguous decoding can be expressed as
$$\begin{aligned} \sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )&=\mathbb {E}[P_\textrm{ud}(H,\varepsilon )^2]-\left( \mathbb {E}[P_\textrm{ud}(H,\varepsilon )]\right) ^2\\&=\mathbb {E}\left[ \left( \sum _{i=1}^n I_i \varepsilon ^i(1-\varepsilon )^{n-i}\right) ^2\right] -\left( \sum _{i=1}^n \mathbb {E}[I_i]\varepsilon ^i(1-\varepsilon )^{n-i}\right) ^2\\&=\sum _{i,i'=1}^n\textrm{Cov}(I_i,I_{i'})\varepsilon ^{i+i'}(1-\varepsilon )^{2n-i-i'}, \end{aligned}$$
where the term \(\textrm{Cov}(I_i,I_{i'})\) is given by
$$\begin{aligned}\textrm{Cov}(I_i,I_{i'})=\mathbb {E}[I_iI_{i'}]-\mathbb {E}[I_i]\mathbb {E}[I_{i'}]. \end{aligned}$$
We first obtain:
Lemma 5
For \(1 \le i,i' \le n\), we have
$$\begin{aligned}&\textrm{Cov}(I_i,I_{i'})\nonumber \\&\quad =\psi _m(i)\psi _m(i')\sum _{s=1}^{\min \{i,i'\}} \left( \frac{1}{\psi _m(s)}-1\right) \left( {\begin{array}{c}n\\ s, i-s, i'-s, n-i-i'+s\end{array}}\right) . \end{aligned}$$
(38)
Here the multinomial coefficient \(\left( {\begin{array}{c}n\\ a,b,c,d\end{array}}\right) \) for any non-negative integers abcdn such that \(a+b+c+d=n\) is given by
$$\begin{aligned} \left( {\begin{array}{c}n\\ a,b,c,d\end{array}}\right) :=\frac{n!}{a!b!c!d!}. \end{aligned}$$
Proof of Lemma 5
Obviously if i or \(i'\) does not satisfy the relation \(1 \le i, i' \le m\), then both sides of (38) are zero. So we may assume that \(1 \le i, i' \le m\).
For any \(H\in \mathcal {R}_{m,n}\), from (12) we see that
$$\begin{aligned} I_i(C_H) =\sum _{\begin{array}{c} E \subset [n]\\ \#E=i \end{array}} \left( 1-\mathbb {1}_{\dim C_H(E)=0}\right) =\sum _{\begin{array}{c} E \subset [n]\\ \#E=i \end{array}} \left( 1-\mathbb {1}_{\textrm{rk}(H_E)=i}\right) .\end{aligned}$$
So we can expand the term \(\textrm{Cov}(I_i,I_{i'})\) as
$$\begin{aligned}&\textrm{Cov}(I_i,I_{i'})\nonumber \\&\quad =\mathbb {E}[I_iI_{i'}]-\mathbb {E}[I_i]\mathbb {E}[I_{i'}]\nonumber \\&\quad =\sum _{\begin{array}{c} E \subset [n]\\ \# E=i \end{array}} \sum _{\begin{array}{c} E' \subset [n]\\ \# E'=i' \end{array}} \mathbb {E}[\mathbb {1}_{\textrm{rk}(H_E)=i}\mathbb {1}_{\textrm{rk}(H_E')=i'}]-\mathbb {E}[\mathbb {1}_{\textrm{rk}(H_E)=i}]\mathbb {E}[\mathbb {1}_{\textrm{rk}(H_E')=i'}] \nonumber \\&\quad =\sum _{\begin{array}{c} E \subset [n]\\ \# E=i \end{array}} \sum _{\begin{array}{c} E' \subset [n]\\ \# E'=i' \end{array}} \Big [\mathbb {P}(\textrm{rk}(H_E) = i \cap \textrm{rk}(H_{E'}) = i')-\mathbb {P}(\textrm{rk}(H_E) = i)\mathbb {P}(\textrm{rk}(H_{E'}) = i')\Big ]. \end{aligned}$$
(39)
Applying Lemmas 2 and 4 to the terms in the inner sums on the right hand side of (39), we have
$$\begin{aligned} \textrm{Cov}(I_i,I_{i'})=\sum _{s}\sum _{\begin{array}{c} E,E' \subset [n]\\ \# E=i, \#E'=i'\\ \#(E \cap E')=s \end{array}} \left( \frac{\psi _m(i)\psi _m(i')}{\psi _m(s)}-\psi _m(i)\psi _m(i')\right) . \end{aligned}$$
(40)
Together with the combinatorial identity
$$\begin{aligned} \sum _{\begin{array}{c} E,E' \subset [n]\\ \# E=i,\# E'=i'\\ \#(E \cap E')=s \end{array}} 1=\left( {\begin{array}{c}n\\ i\end{array}}\right) \left( {\begin{array}{c}i\\ s\end{array}}\right) \left( {\begin{array}{c}n-i\\ i'-s\end{array}}\right) =\left( {\begin{array}{c}n\\ s, i-s, i'-s, n-i-i'+s\end{array}}\right) , \end{aligned}$$
(41)
we obtain the desired result as described by Lemma 5. \(\square \)
From Lemma 5, the variance \(\sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )\) can be obtained easily, which we summarize below:
Theorem 4
$$\begin{aligned}&\sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )\\&\quad = \sum _{i,i'=1}^m\sum _{s=1}^{\min \{i,i'\}}\psi _m(i)\psi _m(i') \left( \frac{1}{\psi _m(s)}-1\right) \left( {\begin{array}{c}n\\ s, i-s, i'-s, n-i-i'+s\end{array}}\right) \\&\qquad \times \varepsilon ^{i+i'}(1-\varepsilon )^{2n-i-i'}. \end{aligned}$$
For \(0<R<1\), the error exponent of the variance \(\sigma _{\textrm{ud}}^2(\mathcal {R}_{(1-R)n,n},\varepsilon )\) is defined by
$$\begin{aligned} S_{\textrm{ud}}(\varepsilon ):=-\lim _{n \rightarrow \infty } \frac{1}{n}\log _q \sigma _{\textrm{ud}}^2(\mathcal {R}_{(1-R)n,n},\varepsilon ), \end{aligned}$$
(42)
if the limit exists. We obtain:
Theorem 5
$$\begin{aligned} S_\textrm{ud}(\varepsilon )={\left\{ \begin{array}{ll}1-R-\log _q[1+(q-1)\varepsilon ^2] & \left( 0< R \le \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\right) \\ (1-R)\log _q\left( \frac{\kappa _0}{\varepsilon ^2}\right) +R\log _q\left( \frac{2R-1+\kappa _0}{(1-\varepsilon )^2}\right) & \left( \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}< R < 1\right) \end{array}\right. }, \end{aligned}$$
where \(\kappa _0:=\kappa _0(R)\) is given by
$$\begin{aligned} \kappa _0(R):=1-R-\frac{\sqrt{4R(1-R)(q-1)+1}-1}{2(q-1)}. \end{aligned}$$
(43)
A plot of the function \(S_{\textrm{ud}}(\varepsilon )\) for \(q=2,\varepsilon =0.25\) in the range \(0< R < 1\) is given by Fig. 2.
Fig. 2
The error exponent \(S_{\textrm{ud}}(\varepsilon )\) for \(0<R<1\), where \(q=2, \varepsilon =0.25\)
Full size image
We remark that the proof of Theorem 5 follows a similar argument as that of Theorem 2, though here the computation of \(S_{\textrm{ud}}(\varepsilon )\) is much more complex as it involves a lot more technical manipulations. In order to streamline the idea of the proof in this section, we first assume Theorem 5 and leave its proof to Sect. 7 in Appendix. Then Theorem 3 can be proved easily by using the standard Chebyshev’s inequality.

6.2 Proof of Theorem 3

We have, by definition of error exponents,
$$\begin{aligned} P_{\textrm{ud}}(\mathcal {R}_{(1-R)n,n},\varepsilon )=q^{-n(T_\textrm{ud}(\varepsilon )+o(1))}, \end{aligned}$$
and
$$\begin{aligned} \sigma _{\textrm{ud}}^2(\mathcal {R}_{(1-R)n,n},\varepsilon )=q^{-n(S_\textrm{ud}(\varepsilon )+o(1))}. \end{aligned}$$
These imply
$$\begin{aligned} \frac{\sigma _{\textrm{ud}}^2(\mathcal {R}_{(1-R)n,n},\varepsilon )}{P_{\textrm{ud}}^2(\mathcal {R}_{(1-R)n,n},\varepsilon )}= \frac{q^{-n(S_\textrm{ud}(\varepsilon )+o(1))}}{q^{-2n(T_{\textrm{ud}}(\varepsilon )+o(1))}} =q^{-n(S_\textrm{ud}(\varepsilon )-2T_{\textrm{ud}}(\varepsilon )+o(1))}. \end{aligned}$$
The right hand side of the above will tend to 0 as \(n \rightarrow \infty \) if
$$\begin{aligned} c(\varepsilon ,R):=S_\textrm{ud}(\varepsilon )-2T_\textrm{ud}(\varepsilon ) > 0. \end{aligned}$$
(44)
Thus under (44), by Chebyshev’s inequality, we have, for any fixed \(\delta > 0\),
$$\begin{aligned} \mathbb {P}_{\mathcal {R}_{(1-R)n,n}}\left[ \left| \frac{P_\textrm{ud}(H,\varepsilon )}{P_{\textrm{ud}}(\mathcal {R}_{(1-R)n,n},\varepsilon )} -1\right| \ge \delta \right] \le \frac{\sigma _{\textrm{ud}}^2(\mathcal {R}_{(1-R)n,n},\varepsilon )}{\delta ^2P_{\textrm{ud}}^2(\mathcal {R}_{(1-R)n,n},\varepsilon )} =q^{-n(c(\varepsilon ,R)+o(1))}, \end{aligned}$$
that is, \(\frac{P_\textrm{ud}(H,\varepsilon )}{P_{\textrm{ud}}(\mathcal {R}_{(1-R)n,n},\varepsilon )} \rightarrow 1\) WHP as \(n \rightarrow \infty \).
To prove Theorem 3, it remains to verify that (44) holds true under the assumptions of either (1) or (2) of Theorem 3.
Case 1. \(\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2} \le R < 1-\varepsilon \)
Since \(\frac{1-\varepsilon }{1+(q-1)\varepsilon }<\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\), by a simple calculation, we have
$$\begin{aligned} c(\varepsilon ,R)&=(1-R)\log _q\left( \frac{\kappa _0}{\varepsilon ^2}\right) -R\log _q\left( \frac{2R-1+\kappa _0}{(1-\varepsilon )^2}\right) \\&\quad + -2\left[ (1-R)\log _q\left( \frac{1-R}{\varepsilon }\right) +R\log _q\left( \frac{R}{1-\varepsilon }\right) \right] \\&=(1-R)\log _q\left( \frac{\kappa _0}{(1-R)^2}\right) +R\log _q\left( \frac{2R-1+\kappa _0}{R^2}\right) \\&>(1-R)\log _q\left( \frac{(1-R)^2}{(1-R)^2}\right) +R\log _q\left( \frac{2R-1+(1-R)^2}{R^2}\right) =0, \end{aligned}$$
so (44) always holds true in this case.
Case 2. \(\frac{1-\varepsilon }{1+(q-1)\varepsilon } \le R < \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\)
We can actually obtain a slightly more general result than (2) of Theorem 3 as follows:
Lemma 6
If (44) holds for \(R=R_0\) where \(R_0\) satisfies \(0<R_0 \le \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\), then (44) also holds true for any \(R \in \left[ R_0,\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\right] \).
Proof of Lemma 6
First, from Case 1 and the continuity of \(S_\textrm{ud}(\varepsilon )\) and \(T_\textrm{ud}(\varepsilon )\), we know that (44) holds for \(R=\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\). Hence such \(R_0\) always exists.
Now if \(R \in \left[ \frac{1-\varepsilon }{1+(q-1)\varepsilon },\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\right] \), then we have
$$\begin{aligned} c(\varepsilon ,R)=1-R-\log _q[1+(q-1)\varepsilon ^2]-2\left[ (1-R)\log _q\left( \frac{1-R}{\varepsilon }\right) +R\log _q\left( \frac{R}{1-\varepsilon }\right) \right] . \end{aligned}$$
Differentiating \(c(\varepsilon ,R)\) twice with respect to R, we have
$$\begin{aligned} \frac{\partial ^2 c(\varepsilon ,R)}{\partial R^2}=-\frac{2}{R(1-R)\ln q}, \end{aligned}$$
which is negative for \(R \in \left[ \frac{1-\varepsilon }{1+(q-1)\varepsilon },\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\right] \), and thus \(c(\varepsilon ,R)\) is convex \(\cap \) for R within that interval.
Next, for \(0< R < \frac{1-\varepsilon }{1+(q-1)\varepsilon }\), we note that
$$\begin{aligned} c(\varepsilon ,R)&=1-R-\log _q[1+(q-1)\varepsilon ^2]-2\{1-R-\log _q[1+(q-1)\varepsilon ]\}\\&=R-1-\log _q[1+(q-1)\varepsilon ^2]+2\log _q[1+(q-1)\varepsilon ] \end{aligned}$$
is a linear function in R with positive slope. Hence the function \(c(\varepsilon ,R)\) is convex \(\cap \) for R within the whole interval \(\left[ R_0,\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\right] \), and Lemma 6 follows. \(\square \)
Now we let \(q \in \{2,3,4\}\). Consider \(R=\frac{1-\varepsilon }{1+(q-1)\varepsilon }\). Under this special value,
$$\begin{aligned} Q(\varepsilon ):=c(\varepsilon ,R)=-\frac{q\varepsilon }{1+(q-1)\varepsilon }-\log _q[1+(q-1)\varepsilon ^2]+2\log _q[1+(q-1)\varepsilon ]. \end{aligned}$$
Differentiating \(Q(\varepsilon )\) with respect to \(\varepsilon \), we have
$$\begin{aligned} Q'(\varepsilon )&=-\frac{q}{[1+(q-1)\varepsilon ]^2}-\frac{2(q-1)\varepsilon }{[1+(q-1)\varepsilon ^2]\ln q}+\frac{2(q-1)}{[1+(q-1)\varepsilon ]\ln q}\\&=-\frac{(q-1)(q\ln q+2q-2)}{[1+(q-1)\varepsilon ]^2[1+(q-1)\varepsilon ^2]\ln q} \left( \varepsilon ^2-\frac{2(q-2)}{q\ln q+2q-2} \, \varepsilon -\frac{2-\frac{q \ln q}{q-1}}{q\ln q+2q-2}\right) . \end{aligned}$$
It is easy to check that when \(q \in \{2,3,4\}\), then we have
$$\begin{aligned}0 \le \frac{2(q-2)}{q\ln q+2q-2}< 1, \quad Q'(0)>0, Q'(1)<0.\end{aligned}$$
Therefore exactly one of the two roots for \(Q'(\varepsilon )\) lies in the desired range \(0< \varepsilon < 1\), and \(Q(\varepsilon )\) attains local maximum at that point. Since \(Q(0)=Q(1)=0\), we conclude that \(Q(\varepsilon )\) is positive for \(0< \varepsilon < 1\), and therefore (44) holds for \(R=\frac{1-\varepsilon }{1+(q-1)\varepsilon }\).
Now by Lemma 6, we see that Eq. (44) holds for \(R \in \left[ \frac{1-\varepsilon }{1+(q-1)\varepsilon },\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\right] \) when \(q=2,3,4\). This completes the proof of Theorem 3.

Acknowledgements

The authors would like to acknowledge the editor and anonymous reviewers for their valuable comments and suggestions to improve our manuscript.

Declarations

Conflict of interest

The authors declared that they hav no conflict of interest.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix

Appendix: Proof of Theorem 5

In order to prove Theorem 5, we first need to obtain
Lemma 7
$$\begin{aligned} S_{\textrm{ud}}(\varepsilon )=-\sup _{t,t',\kappa } g(t,t',\kappa ), \end{aligned}$$
where
$$\begin{aligned} g(t,t',\kappa )=(\kappa -1+R)+h(\kappa ,t-\kappa ,t'-\kappa ,1-t-t'+\kappa )+(t+t')\log _q\varepsilon +(2-t-t')\log _q(1-\varepsilon ), \end{aligned}$$
(45)
and the supremum is taken over positive real numbers \(t,t',\kappa \) satisfying
$$\begin{aligned} t,t' \le 1-R, \quad t+t'-1 \le \kappa \le \min \{t,t'\}. \end{aligned}$$
(46)
Here
$$\begin{aligned} h(t_1,t_2,\ldots ,t_r)=-\sum _{j=1}^r t_j\log _q t_j\quad \left( t_j \ge 0, \sum _{j=1}^r t_j=1\right) \end{aligned}$$
(47)
is the multi-entropy function (in q-its).
Proof of Lemma 7
Write \(m=(1-R)n\). We define
$$\begin{aligned} g_{i,i',s}:=\psi _m(i)\psi _m(i')\left( \frac{1}{\psi _m(s)}-1\right) \left( {\begin{array}{c}n\\ s, i-s, i'-s, n-i-i'+s\end{array}}\right) \varepsilon ^{i+i'}(1-\varepsilon )^{2n-i-i'}. \end{aligned}$$
We note that \(g_{i,i',s} \ge 0\) for all integers \(i,i',s\), and is nonzero if and only if \(1 \le i,i' \le m, 1 \le s \le \min \{i,i'\}\) and \(i'-s \le n-i\). Then we can rewrite the term \(\sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )\) (see Theorem 4) as
$$\begin{aligned} \sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )=\sum _{i,i',s} g_{i,i',s}, \end{aligned}$$
where summation is over all integers \(i,i',s\) satisfying the conditions
$$\begin{aligned} 1 \le i,i' \le m, \quad \max \{1,i+i'-n\} \le s \le \min \{i,i'\}. \end{aligned}$$
(48)
There are at most \(m^3\) such integer triples \((i,i',s)\). Since \(\lim _{n \rightarrow \infty } \frac{\log _q m^3}{n}=0\), we have
$$\begin{aligned} \frac{1}{n}\log _q\sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )=o(1)+\frac{1}{n}\max _{i,i',s} \log _q g_{i,i',s}. \end{aligned}$$
(49)
Now we need a careful analysis of the quantity \(g_{i,i',s}\). First, using \(T:=q^{-1}\) and \(M:=(T)_\infty \in (0,1)\), we observe that
$$\begin{aligned} M^2 \le \psi _m(i)\psi _m(i')= \frac{(T)_m^2}{(T)_{m-i}(T)_{m-i'}} \le \frac{1}{M^2}, \quad \forall 1 \le i, i' \le m. \end{aligned}$$
So
$$\begin{aligned} \frac{1}{n}\log _q[\psi _m(i)\psi _m(i')]=o(1). \end{aligned}$$
Next, for \(1 \le s \le m\), using (35), we have,
$$\begin{aligned} \frac{1}{\psi _m(s)}-1&=\frac{(T)_{m-s}}{(T)_m}-1 =\frac{(T)_{m-s}}{(T)_m}\left( 1-\prod _{j=m-s+1}^m(1-T^j)\right) . \end{aligned}$$
It is easy to see that
$$\begin{aligned}1-T^{m-s+1} \ge \prod _{j=m-s+1}^m(1-T^j) > 1-\sum _{j=m-s+1}^{\infty }T^j=1-\frac{q}{q-1}T^{m-s+1}, \end{aligned}$$
so we have
$$\begin{aligned}\frac{1}{n} \log _q \left( \frac{1}{\psi _m(s)}-1\right) =\frac{s-m-1}{n}+o(1).\end{aligned}$$
Therefore we can simplify (49) as
$$\begin{aligned}&\frac{1}{n}\log _q\sigma _{\textrm{ud}}^2(\mathcal {R}_{m,n},\varepsilon )\\&\quad =o(1)+\max _{i,i',s}\bigg [\frac{s-m-1}{n}+\frac{1}{n}\log _q\left( {\begin{array}{c}n\\ s, i-s, i'-s, n-i-i'+s\end{array}}\right) +\frac{i+i'}{n}\log _q \varepsilon \\&\qquad +\frac{2n-i-i'}{n}\log _q(1-\varepsilon )\bigg ]. \end{aligned}$$
Letting \(i=tn, j=t'n\) and \(s=\kappa n\), using the following generalization of (36) (which can be verified similarly)
$$\begin{aligned} \frac{1}{n}\log _q \left( {\begin{array}{c}n\\ t_1n,t_2n,\ldots ,t_rn\end{array}}\right) =h(t_1,t_2,\ldots ,t_r)+o(1), \end{aligned}$$
where \(h(t_1,t_2,\ldots ,t_r)\) (with \(\sum _{j=1}^r t_j=1\)) is defined as in (47), and then taking \(n \rightarrow \infty \), we obtain
$$\begin{aligned} S_{\textrm{ud}}(\varepsilon )=-\sup _{t,t',\kappa } g(t,t',\kappa ), \end{aligned}$$
where \(g(t,t',\kappa )\) is defined as in (45) and the supremum is taken over positive real numbers \(t,t',\kappa \) satisfying (46). This completes the proof of Lemma 7. \(\square \)
Using Lemma 7, now we provide a detailed proof of Theorem 5.
We start from the function \(g(t,t',\kappa )\) given in (45). It is best to first take the supremum over \(t'\) while fixing t and \(\kappa \). Therefore we differentiate g with respect to \(t'\) while keeping t and \(\kappa \) fixed, where the range of \(t'\) should be taken as \(\kappa \le t' \le \min \{\kappa -t+1,1-R\}\):
$$\begin{aligned} \frac{\partial g}{\partial t'}=\log _q\left( \frac{(1-t-t'+\kappa )\varepsilon }{(t'-\kappa )(1-\varepsilon )}\right) . \end{aligned}$$
Solving for \(\frac{\partial g}{\partial t'}=0\), we get \(t'=\varepsilon (1-t)+\kappa \), and we check that g attains local maximum at that point.
This leads us to consider two cases:
Case 1. \(t \ge 1-\frac{1-R-\kappa }{\varepsilon }\)
In this case the critical point \(\varepsilon (1-t)+\kappa \) lies within the required range of \(t'\). We see that the maximum of g is
$$\begin{aligned} G_1(t,\kappa )&:=g(t,\varepsilon (1-t)+\kappa ,\kappa )\\&=(\kappa -1+R)+h\left( \kappa ,t-\kappa ,\varepsilon (1-t),(1-\varepsilon )(1-t)\right) \\&\quad +[(\varepsilon +\kappa )+(1-\varepsilon )t]\log _q\varepsilon +[(2-\varepsilon -\kappa )-(1-\varepsilon )t]\log _q(1-\varepsilon ). \end{aligned}$$
Differentiating both sides with respect to t while keeping \(\kappa \) fixed, where the range of t is taken as \(\max \{\kappa ,1-\frac{1-R-\kappa }{\varepsilon }\} \le t \le 1-R\), we have
$$\begin{aligned} \frac{\partial G_1}{\partial t}=\log _q\left( \frac{\varepsilon (1-t)}{t-\kappa }\right) . \end{aligned}$$
Solving for \(\frac{\partial G_1}{\partial t}=0\), we get \(t=\frac{\varepsilon +\kappa }{1+\varepsilon }\). We also see that \(G_1\) attains local maximum at this point. We note that \(\kappa \le \frac{\varepsilon +\kappa }{1+\varepsilon }\) as \(\kappa < 1\). In order for the range of t to be nonempty, we also need \(1-\frac{1-R-\kappa }{\varepsilon } \le 1-R\), which is true if and only if \(0 < \kappa \le 1-R-R\varepsilon \) (and this range is nonempty if and only if \(0< R < \frac{1}{1+\varepsilon }\)). We then obtain
$$\begin{aligned} 1-\frac{1-R-\kappa }{\varepsilon } \le \frac{\varepsilon +\kappa }{1+\varepsilon } \le 1-R, \end{aligned}$$
so the maximum of g is
$$\begin{aligned} \mathcal {G}_1(\kappa )&:=G_1\left( \frac{\varepsilon +\kappa }{1+\varepsilon },\kappa \right) \\&=(\kappa -1+R)+h\left( \kappa ,\frac{\varepsilon (1-\kappa )}{1+\varepsilon },\frac{\varepsilon (1-\kappa )}{1+\varepsilon },\frac{(1-\varepsilon )(1-\kappa )}{1+\varepsilon }\right) \\&\quad +2\left( \frac{\varepsilon +\kappa }{1+\varepsilon }\right) \log _q\varepsilon +2\left( \frac{1-\kappa }{1+\varepsilon }\right) \log _q(1-\varepsilon ). \end{aligned}$$
Differentiating both sides with respect to \(\kappa \) and checking for \(\mathcal {G}_1'(\kappa )=0\), we have \(\kappa =\frac{q\varepsilon ^2}{1+(q-1)\varepsilon ^2}\). In addition, \(\mathcal {G}_1\) attains local maximum at this critical point.
Then we have the following two possible cases:
1.
\(0 < R \le \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\)
In this case \(\frac{q\varepsilon ^2}{1+(q-1)\varepsilon ^2} \le 1-R-R\varepsilon \), and so the maximum is
$$\begin{aligned} \mathcal {G}_1\left( \frac{q\varepsilon ^2}{1+(q-1)\varepsilon ^2}\right) =-1+R+\log _q[1+(q-1)\varepsilon ^2]. \end{aligned}$$
 
2.
\(\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}< R < \frac{1}{1+\varepsilon }\)
In this case \(0< 1-R-R\varepsilon < \frac{q\varepsilon ^2}{1+(q-1)\varepsilon ^2}\), and so the maximum is
$$\begin{aligned} \mathcal {G}_1(1-R-R\varepsilon )=-R\varepsilon -(1-R-R\varepsilon )\log _q\left( \frac{1-R-R\varepsilon }{\varepsilon ^2}\right) -R(1+\varepsilon )\log _q\left( \frac{R}{1-\varepsilon }\right) . \end{aligned}$$
 
Case 2. \(t < 1-\frac{1-R-\kappa }{\varepsilon }\)
In this case the critical point \(\varepsilon (1-t)+\kappa \) is larger than the upper bound \(1-R\) of \(t'\). Hence maximum is
$$\begin{aligned} G_2(t,\kappa )&:=g(t,1-R,\kappa )\\&=(\kappa -1+R)+h(\kappa ,t-\kappa ,1-R-\kappa ,R-t+\kappa )+(t+1-R)\log _q\varepsilon \\&\quad + (1+R-t)\log _q(1-\varepsilon ). \end{aligned}$$
Differentiating both sides with respect to t while keeping \(\kappa \) fixed, where the range of t is taken as \(\kappa \le t \le \min \{1-\frac{1-R-\kappa }{\varepsilon }, 1-R\}\), we get
$$\begin{aligned} \frac{\partial G_2}{\partial t}=\log _q\left( \frac{\varepsilon (R+\kappa -t)}{(t-\kappa )(1-\varepsilon )}\right) . \end{aligned}$$
Solving for \(\frac{\partial G_2}{\partial t}=0\), we have \(t=R\varepsilon +\kappa \). We also see that \(G_2\) attains local maximum at this point. We already see that \(\kappa \le R\varepsilon +\kappa \). Hence we need to compare the critical point with \(\min \{1-\frac{1-R-\kappa }{\varepsilon }, 1-R\}\). First in order that the range of t is nonempty, we must have \(\kappa \le \min \{1-\frac{1-R-\kappa }{\varepsilon },1-R\}\), which is true if and only if \(1-\frac{R}{1-\varepsilon } \le \kappa \le 1-R\). We can further divide into two subcases:
Case 2a. \(1-\frac{R}{1-\varepsilon } \le \kappa \le 1-R-R\varepsilon \)
In this case we have \(1-\frac{1-R-\kappa }{\varepsilon } \le R\varepsilon +\kappa \le 1-R\). Hence the maximum occurs at \(t=1-\frac{1-R-\kappa }{\varepsilon }\). Note that this value of t is precisely the one in which \(t'=1-R=\varepsilon (1-t)+\kappa \). Therefore it is covered in Case 1 already, and the maximum so obtained cannot be larger than the value calculated in that case. Note that this case can only happen when \(0< R < \frac{1}{1+\varepsilon }\).
Case 2b. \(\max \{0, 1-R-R\varepsilon \} < \kappa \le 1-R\)
In this case we have \(1-R< R\varepsilon +\kappa < 1-\frac{1-R-\kappa }{\varepsilon }\). Hence the maximum occurs at \(t=1-R\).
Then we get
$$\begin{aligned} \mathcal {G}_2(\kappa )&:=G_2(1-R,\kappa )\\&=(\kappa -1+R)+h(\kappa ,1-R-\kappa ,1-R-\kappa ,2R-1+\kappa )\\&\qquad +2(1-R)\log _q\varepsilon +2R\log _q(1-\varepsilon ). \end{aligned}$$
Differentiating both sides with respect to \(\kappa \), we have
$$\begin{aligned} \mathcal {G}_2'(\kappa )=\log _q\left( \frac{q(1-R-\kappa )^2}{\kappa (2R-1+\kappa )}\right) . \end{aligned}$$
(50)
Solving for \(\mathcal {G}_2'(\kappa )=0\), we obtain two roots
$$\begin{aligned} \kappa _{\pm }=1-R+\frac{1 \pm \sqrt{4R(1-R)(q-1)+1}}{2(q-1)}. \end{aligned}$$
However under our assumption we require \(\kappa \le 1-R\). It is easy to see that we should then take \(\kappa _-\). This is precisely \(\kappa _0\) given by (43).
Note that the number
$$\begin{aligned} N(\kappa )=\frac{q(1-R-\kappa )^2}{\kappa (2R-1+\kappa )} \end{aligned}$$
inside the logarithm in right hand side of (50) is a strictly decreasing function within our range of \(\kappa \). Hence it suffices to check the value of \(N(\kappa )\) at the two bounds to see whether \(\kappa _0\) is within our range (in particular \(N(\kappa _0)=1\)). It is clear that
$$\begin{aligned} N(1-R)=0,\quad N(1-R-R\varepsilon )=\frac{qR\varepsilon ^2}{(1-\varepsilon )(1-R-R\varepsilon )} \end{aligned}$$
and
$$\begin{aligned} \lim _{\kappa \rightarrow 0^+} N(\kappa )=+\infty . \end{aligned}$$
Then we have two cases again:
1.
\(0 < R \le \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\)
In this case we have \(N(1-R-R\varepsilon ) \le 1\), and so \(\mathcal {G}_2'(\kappa ) \le 0\) within the range of \(\kappa \). This shows that the maximum occurs at \(\kappa =1-R-R\varepsilon \). Note that this also implies \(t=1-R=1-\frac{1-R-\kappa }{\varepsilon }\). Since this value is already covered in Case 2a, the maximum cannot be greater than in that case and thereby in Case 1 too.
 
2.
\(\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}< R < 1\)
In this case we have \(N(1-R-R\varepsilon ) > 1\), so one of the critical points (in fact the smaller one) is within our range. That number is exactly \(\kappa _0\) defined in (43). Since \(\mathcal {G}_2'(\kappa )\) is decreasing in our range, this implies \(\mathcal {G}_2\) attains maximum at \(\kappa _0\). The maximum will then be
$$\begin{aligned} \mathcal {G}_2(\kappa _0)&=(\kappa _0-1+R)+h(\kappa _0,1-R-\kappa _0,1-R-\kappa _0,2R-1+\kappa _0)\\&\quad +2(1-R)\log _q\varepsilon +2R\log _q(1-\varepsilon )\\&=-(1-R)\log _q\left( \frac{\kappa _0}{\varepsilon ^2}\right) -R\log _q\left( \frac{2R-1+\kappa _0}{(1-\varepsilon )^2}\right) \end{aligned}$$
after simplification and applying the relation \(\mathcal {G}_2'(\kappa _0)=0\).
 
Note that in particular when \(\frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}< R < \frac{1}{1+\varepsilon }\), this value is larger than \(\mathcal {G}_2(1-R-R\varepsilon )=-R\varepsilon -(1-R-R\varepsilon )\log _q\left( \frac{1-R-R\varepsilon }{\varepsilon ^2}\right) -R(1+\varepsilon )\log _q\left( \frac{R}{1-\varepsilon }\right) =\mathcal {G}_1(1-R-R\varepsilon )\).
Combining all above cases, we finally obtain
$$\begin{aligned} S_\textrm{ud}(\varepsilon )=-\sup _{t,t',\kappa }g(t,t',\kappa )={\left\{ \begin{array}{ll}1-R-\log _q[1+(q-1)\varepsilon ^2] & \left( 0< R \le \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}\right) \\ (1-R)\log _q\left( \frac{\kappa _0}{\varepsilon ^2}\right) +R\log _q\left( \frac{2R-1+\kappa _0}{(1-\varepsilon )^2}\right) & \left( \frac{1-\varepsilon }{1+(q-1)\varepsilon ^2}< R < 1\right) . \end{array}\right. } \end{aligned}$$
This completes the proof of Theorem 5.

Conclusion

In this paper we carried out an in-depth study on the average decoding error probabilities of the random parity-check matrix ensemble \(\mathcal {R}_{m,n}\) over the erasure channel under three decoding principles, namely unambiguous decoding, maximum likelihood decoding and list decoding.
(1)
We obtained explicit formulas for the average decoding error probabilities of the ensemble under these three decoding principles and computed the error exponents. We also compare the results with the random \([n,k]_q\) code ensemble studied before.
 
(2)
For unambiguous decoding, we computed the variance of the decoding error probability of the ensemble and the error exponent of the variance, from which we derived a strong concentration result, that is, under general conditions, the ratio of the decoding error probability of a random code in the ensemble and the average decoding error probability of the ensemble converges to 1 with high probability when the code length goes to infinity.
 
It might be interesting to extend the results of (2) to general list decoding and maximum likelihood decoding for the ensemble \(\mathcal {R}_{m,n}\). As it turns out, the variance of decoding error probability in these two cases can still be computed, but the expressions are much more complicated and it is difficult to obtain explicit formulas for their error exponents, and hence a concentration result from them. We leave this as an open question for future research.
It may be also interesting to carry out such computation for the random \([n,k]_q\) code ensemble.
Literature
1.
go back to reference Berlekamp E.R.: The technology of error-correcting codes. Proc. IEEE 68(5), 564–593 (1980).CrossRefMATH Berlekamp E.R.: The technology of error-correcting codes. Proc. IEEE 68(5), 564–593 (1980).CrossRefMATH
2.
go back to reference Byers, J.W., Luby, M., Mitzenmacher, M., Rege, A.: A digital fountain approach to reliable distribution of bulk data. In: Proc. ACM SIGCOMM Conf. Appl., Technol., pp. 56–67. Architectures, Protocols Comput. Commun., Vancouver (1998). Byers, J.W., Luby, M., Mitzenmacher, M., Rege, A.: A digital fountain approach to reliable distribution of bulk data. In: Proc. ACM SIGCOMM Conf. Appl., Technol., pp. 56–67. Architectures, Protocols Comput. Commun., Vancouver (1998).
3.
go back to reference Cover T.M., Thomas J.A.: Elements of Information Theory, 1st edn Wiley-Interscience, New York (1991).MATH Cover T.M., Thomas J.A.: Elements of Information Theory, 1st edn Wiley-Interscience, New York (1991).MATH
4.
go back to reference Di C., Proietti D., Telatar I.E., Richardson T.J., Urbanke R.L.: Finite-length analysis of low-density parity-check codes on the binary erasure channel. Special issue on Shannon theory: perspective, trends, and applications. IEEE Trans. Inform. Theory 48(6), 1570–1579 (2002).MathSciNetCrossRefMATH Di C., Proietti D., Telatar I.E., Richardson T.J., Urbanke R.L.: Finite-length analysis of low-density parity-check codes on the binary erasure channel. Special issue on Shannon theory: perspective, trends, and applications. IEEE Trans. Inform. Theory 48(6), 1570–1579 (2002).MathSciNetCrossRefMATH
5.
go back to reference Didier F.: A new upper bound on the block error probability after decoding over the erasure channel. IEEE Trans. Inform. Theory 52(10), 4496–4503 (2006).MathSciNetCrossRefMATH Didier F.: A new upper bound on the block error probability after decoding over the erasure channel. IEEE Trans. Inform. Theory 52(10), 4496–4503 (2006).MathSciNetCrossRefMATH
6.
go back to reference Fashandi, S., Gharan, S.O., Khandani, A.K.: Coding over an erasure channel with a large alphabet size. In: Proc. 2008 IEEE Int. Symp. Inform. Theory, pp. 1053–1057 (2008). Fashandi, S., Gharan, S.O., Khandani, A.K.: Coding over an erasure channel with a large alphabet size. In: Proc. 2008 IEEE Int. Symp. Inform. Theory, pp. 1053–1057 (2008).
7.
go back to reference Gallager, R.G.: Low Density Parity Check Codes. MIT Press (1963); IRE Trans. IT-8, pp. 21–28. Cambridge (1962). Gallager, R.G.: Low Density Parity Check Codes. MIT Press (1963); IRE Trans. IT-8, pp. 21–28. Cambridge (1962).
8.
go back to reference Gallager R.G.: Information Theory and Reliable Communication. Wiley, New York (1968).MATH Gallager R.G.: Information Theory and Reliable Communication. Wiley, New York (1968).MATH
9.
go back to reference Lemes, L.C., Firer, M.: Generalized weights and bounds for error probability over erasure channels. In: Proc. Inf. Theory Appl. Workshop (ITA), p. 108. San Diego (2014). Lemes, L.C., Firer, M.: Generalized weights and bounds for error probability over erasure channels. In: Proc. Inf. Theory Appl. Workshop (ITA), p. 108. San Diego (2014).
10.
go back to reference Liva G., Paolini E., Chiani M.: Bounds on the error probability of block codes over the \(q\)-ary erasure channel. IEEE Trans. Inform. Theory 61(6), 2156–2165 (2013).MATH Liva G., Paolini E., Chiani M.: Bounds on the error probability of block codes over the \(q\)-ary erasure channel. IEEE Trans. Inform. Theory 61(6), 2156–2165 (2013).MATH
11.
go back to reference Luby, M., Mitzenmacher, M., Shokrollahi, A., Sipelman, D.A., Stemann, V.: Practical loss-recilient codes. In: Proc. 29th Annual ACM Symp. Theory Comput., pp. 150–159 (1997). Luby, M., Mitzenmacher, M., Shokrollahi, A., Sipelman, D.A., Stemann, V.: Practical loss-recilient codes. In: Proc. 29th Annual ACM Symp. Theory Comput., pp. 150–159 (1997).
12.
go back to reference Lun D.S., Médard M., Koetter R., Effros M.: On coding for reliable communication over packet networks. Phys. Commun. 1(1), 3–20 (2008).CrossRefMATH Lun D.S., Médard M., Koetter R., Effros M.: On coding for reliable communication over packet networks. Phys. Commun. 1(1), 3–20 (2008).CrossRefMATH
13.
go back to reference MacWilliams, F.J., Sloane,N.J.A.: The Theory of Error-Correcting Codes. North-Holland Mathematical Library, vol. 16. Amsterdam (1981). MacWilliams, F.J., Sloane,N.J.A.: The Theory of Error-Correcting Codes. North-Holland Mathematical Library, vol. 16. Amsterdam (1981).
14.
go back to reference Shen L., Fu F.-W.: The decoding error probability of linear codes over the erasure channel. IEEE Trans. Inform. Theory 65(10), 6194–6203 (2019).MathSciNetCrossRefMATH Shen L., Fu F.-W.: The decoding error probability of linear codes over the erasure channel. IEEE Trans. Inform. Theory 65(10), 6194–6203 (2019).MathSciNetCrossRefMATH
15.
go back to reference Viterbi A.J., Omura J.K.: Principles of Digital Communication and Coding. McGraw-Hill, New York (1979).MATH Viterbi A.J., Omura J.K.: Principles of Digital Communication and Coding. McGraw-Hill, New York (1979).MATH
16.
go back to reference Wadayama T.: On the undetected error probability of binary matrix ensemblesl. IEEE Trans. Inform. Theory 56(5), 2168–2176 (2010).MathSciNetCrossRefMATH Wadayama T.: On the undetected error probability of binary matrix ensemblesl. IEEE Trans. Inform. Theory 56(5), 2168–2176 (2010).MathSciNetCrossRefMATH
17.
go back to reference Weber J.H., Abdel-Ghaffar K.A.S.: Results on parity-check matrices with optimal stopping and/or dead-end set enumerators. IEEE Trans. Inform. Theory 54(3), 1368–1374 (2008).MathSciNetCrossRefMATH Weber J.H., Abdel-Ghaffar K.A.S.: Results on parity-check matrices with optimal stopping and/or dead-end set enumerators. IEEE Trans. Inform. Theory 54(3), 1368–1374 (2008).MathSciNetCrossRefMATH
Metadata
Title
Decoding error probability of random parity-check matrix ensemble over the erasure channel
Authors
Chin Hei Chan
Fang-Wei Fu
Maosheng Xiong
Publication date
16-10-2024
Publisher
Springer US
Published in
Designs, Codes and Cryptography / Issue 1/2025
Print ISSN: 0925-1022
Electronic ISSN: 1573-7586
DOI
https://doi.org/10.1007/s10623-024-01516-5

Premium Partner