1. Introduction
The zero-error capacity of discrete memoryless channels (DMCs) was introduced by Shannon in 1956 [
1]. Since then, numerous works have examined this capacity across various channel classes. From the outset, determining
for DMCs has been recognized as highly challenging. In response, Shannon posed a key question: can the zero-error capacity of a DMC be expressed via the other error capacities of suitably chosen channels? A major breakthrough came from Ahlswede, who proved that the
value for a DMC equals the maximum-error capacity of a related 0–1 arbitrarily varying channel (AVC) [
2].
This paper investigates the algorithmic computability of the zero-error capacity of DMCs and explores the broader computational implications of the Shannon and Ahlswede characterizations. We adopt Turing machine theory as our model of computability, which accurately reflects the capabilities of real-world digital computers.
Shannon’s original theory also provided a graph-theoretic interpretation: each channel corresponds to a simple graph whose Shannon capacity coincides with the channel’s zero-error capacity. In practice, however, channel descriptions are usually given directly by a transition mapping
where
and
are finite alphabets, and
denotes the set of probability distributions over
. A notable application of this formulation is remote state estimation and stabilization [
3].
The zero-error capacity is also central in quantum channels and entanglement-assisted classical channels. Research has focused on superactivation effects [
4,
5], entanglement-assisted gains [
6], and connections to noncommutative graph theory [
7]. Further studies have explored nonlocal correlations [
8], no-signaling assistance [
9], and noiseless feedback [
10]. Surveys of quantum channel capacities provide broader context [
11,
12], while recent advances in quantum graph theory offer fresh insights into zero-error communication [
13].
In general, one seeks the numerical value of , typically irrational, and strives for reliable approximation algorithms that compute it to any specified precision.
Shannon’s use of graph theory involved defining the confusability graph
of a DMC and using its Shannon capacity [
14,
15,
16,
17,
18,
19]. Since then, information theory has vastly expanded to cover multi-user systems, feedback channels, and advanced coding theory. Significant progress has been made in the zero-error capacity within relay, multiple-access, broadcast, and interference channels [
20] and in specific models like binary adder and duplication channels [
21,
22,
23]. Further studies have addressed list decoding [
24,
25], variable-length coding [
26], and adversarial multiple-access channels [
27].
Recent work [
28] has determined the Shannon capacity for two infinite subclasses of strongly regular graphs and analyzed novel graph-join types, strengthening earlier results.
Today, two main algorithmic strategies exist for approximating the zero-error capacity: Shannon’s graph-theoretic method and Ahlswede’s 0–1 AVC–based method. We show that both approaches are non-recursive: there is no Turing machine that, given W, produces the confusability graph , nor one that constructs the corresponding 0–1 AVC.
Moreover, the zero-error capacity plays a significant and important role in analyzing the
-capacity of compound channels under the average decoding error, even when the compound set has only
elements [
29].
This paper is structured as follows:
Section 2 introduces computability concepts and the zero-error capacity of noisy channels and clarifies its links to the Shannon graph capacity and Ahlswede’s AVC framework;
Section 3 presents our main results: the non-computability of the zero-error capacity and the unresolved computability status of the Shannon graph capacity and the maximum-error AVC capacity;
Section 4 analyzes 0–1 AVCs under average error constraints, establishes the computability of their capacity, and shows that the Shannon capacity
is Borel–Turing-computable if and only if the corresponding 0–1 AVC capacity is;
Section 6 summarizes our conclusions and discusses future directions.
Some findings were previously presented at the IEEE Information Theory Workshop 2021 in Kanazawa [
30], and related results from ISIT 2020 [
31] are revisited in
Section 5.
2. Basic Definitions and Results
We apply the theory of Turing machines [
32] and recursive functions [
33] to investigating the computability of the zero-error capacity. For brevity, we restrict ourselves to an informal description and refer to [
34,
35,
36,
37] for detailed treatment.
Table 1 gives an overview of the main definitions and notations.
Turing machines provide a mathematical idealization of real-world computational machines. Any algorithm that can be executed by a real-world computer can, in theory, be simulated by a Turing machine, and vice versa. However, unlike real-world computers, Turing machines are not constrained by factors such as energy consumption, computation time, or memory size. Furthermore, all computation steps on a Turing machine are assumed to be executed without error.
Recursive functions form a special subset of the set , where the symbol “↪” denotes a partial mapping. Turing machines and recursive functions are equivalent in the following sense: a function is computable by a Turing machine if and only if it is a recursive function.
Definition 1. A sequence of rational numbers is said to be computable
if recursive functions exist such thatholds true for all . Likewise, a double sequence of rational numbers is said to be computable
if recursive functions exist such thatholds true for all . Definition 2. A sequence of real numbers is said to converge effectively towards a number if a recursive function exists such that holds true for all that satisfy .
Definition 3. A real number x is said to be computable if a computable sequence of rational numbers exists that converges effectively towards x.
We denote the set of computable real numbers as .
Definition 4. A sequence of computable numbers is called computable
if a computable double sequence of rational numbers, as well as a recursive function , exists such thatholds true for all that satisfy . Definition 5. A sequence of functions with is computable if the mapping is computable.
Definition 6. A computable sequence of computable functions is called computably convergent to F if a partial recursive function exists such thatholds true for all , all , and all . In the following, we consider Turing machines with only one output state. We interpret this output status as the stopping of the Turing machine. This means that for an input , the Turing machine ends its calculation after an unknown but finite number of arithmetic steps, or it computes forever.
Definition 7. We call a set semi-decidable if there is a Turing machine that stops for the input , if and only if applies.
In [
38], Specker constructed a monotonically increasing computable sequence
of rational numbers that is bounded by 1, but the limit
, which naturally exists, is not a computable number. For all
,
exists such that for all
,
always holds, but the function
is not partial recursive. This means there are computable monotonically increasing sequences of rational numbers, which each converge to a finite limit value, but for which the limit values are not computable numbers and therefore the convergence is not effective. Of course, the set of computable numbers is countable.
We will later examine the zero-error capacity as a function of computable DMCs. To do this, we need to define computable functions generally.
Definition 8. A function is called Banach–Mazur-computable if f maps any given computable sequence of computable numbers into a computable sequence of real numbers.
Definition 9. A function is called Borel–Turing-computable if there is an algorithm that transforms each given computable sequence of a computable real x into a corresponding representation for .
We note that Turing’s original definition of computability conforms to the definition of Borel–Turing computability above. Banach–Mazur computability (see Definition 8) is the weakest form of computability. For an overview of the logical relations between different notions of computability, we again refer to [
39].
Now, we want to define the zero-error capacity. Therefore, we need the definition of a discrete memoryless channel. In the theory of transmission, the receiver must be in a position to successfully decode all of the messages transmitted by the sender.
Let be a finite alphabet. We denote the set of probability distributions as . We define the set of computable probability distributions as the set of all probability distributions such that for all . Furthermore, for finite alphabets and , let be the set of all conditional probability distributions (or channels) . denotes the set of all computable conditional probability distributions, i.e., for every .
Let . We call M semi-decidable (see Definition 7) if and only if there is a Turing machine that either stops or computes forever, depending on whether is true. That means accepts exactly the elements of M and calculates forever for an input .
Definition 10. A discrete memoryless channel (DMC) is a triple , where is the finite input alphabet, is the finite output alphabet, and with , . The probability of a sequence being received if was sent is defined by Definition 11. A block code with the rate R and the block length n consists of
A message set with ;
An encoding function ;
A decoding function .
We call such a code an -code.
Definition 12. - 1.
The individual message probability of error is defined by the conditional probability of error given that the message m is transmitted: - 2.
We define the maximal probability of the error as .
- 3.
A rate R is said to be achievable if a sequence of -codes exists with a probability of error as .
Two sequences and of the size n of input variables are distinguishable by a receiver if the vectors and are orthogonal. That means if , then , and if then . We denote as the maximum cardinality of a set of mutually orthogonal vectors among with .
There are different ways to define the capacity of a channel. The so-called pessimistic capacity is defined as
, and the optimistic capacity is defined as
. A discussion of these quantities can be found in [
40]. We define the zero-error capacity of
W as follows:
For the zero-error capacity, the pessimistic capacity and the optimistic capacity are equal.
First, we want to introduce the representation of the zero-error capacity of Ahlswede. Therefore, we need to introduce the arbitrarily varying channel (AVC). This was introduced under a different name by Blackwell, Breiman, and Thomasian [
41], and considerable progress has been made in the study of these channels.
Definition 13. Let and be finite sets. A (discrete) arbitrarily varying channel (AVC) is determined by a family of channels with a common input alphabet and output alphabet The index s is called the state, and the set is called the state set. Now, an AVC is defined by a family of sequences of channelsfor all . Definition 14. An code is a system with , , and for .
Definition 15. - 1.
The maximal probability of error of the code for an AVC is
.
- 2.
The average probability of error of the code for an AVC is
.
Definition 16. - 1.
The capacity of an AVC with the maximal probability of error is the maximal number such that for all , an code of the AVC exists for all large n with a maximal probability lower than λ and ;
- 2.
The capacity of an AVC with an average probability of error is the maximal number such that for all ε, , an code of the AVC exists for all large n with an average probability lower than and .
In the following, we denote to be the set of AVCs that satisfies for all , all , and all .
Theorem 1 (Ahlswede [
2])
. Let and be finite alphabets with and .- (i)
For all DMCs , exists such that for the zero-error capacity of - (ii)
Conversely, for each , a DMC exists such that (6) holds.
The construction is interesting. Therefore, we cite it from [
42]:
- (i)
For a given
, we let
be the set of stochastic matrices with the index (state) set
such that for all
,
, and
, it holds that
implies
. Then, for all
n,
, and
,
if and only if
such that
Notice that for all
, a code for
with the maximal probability of error
is a zero-error code for
. Thus, it follows from (
7) that a code is a zero-error code for
if and only if it is a code for
with the maximal probability of error
.
- (ii)
For a given 0-1 type AVC
(with the state set
) and any probability
with
for all
s, let
. Then, (
7) holds.
The zero-error capacity can be characterized in graph-theoretic terms as well. Let
be given and
. Shannon [
1] introduced the confusability graph
with
. In this graph, two letters/vertices
x and
are connected, if they can be confused with one another due to the channel noise (i.e.,
y exists such that
and
). Therefore, the maximum independent set is the maximum number of single-letter messages which can be sent without danger of confusion. In other words, the receiver knows whether the received message is correct or not. It follows that
is the maximum number of messages which can be sent without danger of confusion. Furthermore, the definition is extended to words of a length
n by
. Therefore, we can give the following graph-theoretic definition of the Shannon capacity.
Definition 17. The Shannon capacity of a graph is defined by Shannon discovered the following.
Theorem 2 (Shannon [
1])
. Let be a DMC. Then, This limit exists and equals the supremum
according to Fekete’s lemma.
Observe that Theorem 2 yields no further information on whether and are computable real numbers.
3. The Algorithmic Computability of the Zero-Error Capacity
In this section, we investigate the algorithmic computability of the zero-error capacity for discrete memoryless channels (DMCs), since no closed-form expression for is known to date. Furthermore, we analyze the algorithmic relationship between Shannon’s and Ahlswede’s characterizations of the zero-error capacity.
We show that the function and the cardinality of a maximum-size zero-error code of the blocklength n are not Banach–Mazur-computable. Alon and Lubetzky raised the question of whether the set is semi-decidable. We provide three equivalent conditions under which the answer is affirmative.
Moreover, we demonstrate that the set of channels with a 0 zero-error capacity—which are channels that are useless in this context—is not computable, though they are semi-decidable. To prove this result, we rely on the following auxiliary lemmas.
Lemma 1. There is a Turing machine that stops for , if and only if applies. Hence, the set is semi-decidable.
There is a Turing machine that stops for , if and only if applies. Hence, the set is semi-decidable.
There is no Turing machine that stops for , if and only if applies.
Proof. Let
be given by the quadruple
, with
Then,
with
is a computable monotonically increasing sequence and converges to
x. The Turing machine
sequentially computes the sequence
. Obviously,
with
if and only if
. Since
is always a rational number,
can directly check algorithmically whether
applies. We set
Then,
applies if and only if
applies.
The construction of
is analogous to the computable sequence
where
converges monotonically to
x. Consequently,
We now want to prove the last statement of this lemma. We provide the proof indirectly. Assume that the corresponding Turing machine
exists. Let
be arbitrary. We consider an arbitrary Turing machine
and the computable sequence
of computable numbers:
Obviously, for all
, it holds that
and
, where
if and only if
for the input
n does not stop in a finite number of steps. For all
with
,
holds, as we will show by considering the following cases:
Assume that stops for the input n after steps. For all , then applies, and thus .
Assume that does not stop for the input n after steps. For all , then , and thus .
Hence, we can use the pair
with the estimate
as a computable real number, which we can pass to a potential Turing machine
as input. The partial recursive function
is a representation of the computable number
x, that is,
. Consequently,
stops for the input
x if and only if
for the input
n does not stop in a finite number of steps. Thus, every Turing machine
solves for every input
n the
halting problem. The halting problem cannot be solved by a Turing machine ([
32]). This proves the lemma. □
In the following lemma, we give an example of a function that is not Banach–Mazur-computable.
Lemma 2. Let be arbitrary. We consider the following function:The function is not Banach–Mazur-computable. Proof. For all holds . We assume that is Banach–Mazur-computable. Let be an arbitrary computable sequence of computable numbers with .
The sequence
is a computable sequence of computable numbers. We take a set
that is recursively enumerable but not recursive. Then, let
be a Turing machine that stops for the input
n if and only if
holds.
accepts exactly the elements from
A. Let
be arbitrary. We now define
Then,
is a computable (double) sequence of computable numbers. For
,
and
implies
This means that there is effective convergence for every
. Consequently, according to Lemma 1, for every
,
with
, and the sequence
is a computable sequence of computable numbers. This means that
is a computable sequence of computable numbers, where
applies. The following Turing machine
exists:
computes the value
for the input
n. If
, then the set
, i.e.,
. If
, then set
, i.e.,
. This applies to every
and therefore
A is recursive, which contradicts the assumption. This means that
is not Banach–Mazur-computable. □
Theorem 3. Let be finite alphabets with and . Then, is not Banach–Mazur-computable.
Proof. Let
; then, we will show that
is not Banach–Mazur-computable. For
, we choose
and
. Then, we have
We consider the function
with
. It follows from Lemma 1 that
is not Banach–Mazur-computable. □
Therefore, the zero-error capacity cannot be computed algorithmically.
Remark 1. There are still some questions that we would like to discuss.
- 1.
It is not clear whether applies to all channels .
- 2.
In addition, it is not clear whether Θ is Borel–Turing-computable. Theorem 3 shows that this does not apply to the zero-error capacity for DMCs. We show that even is not Banach–Mazur-computable.
In the following, we want to investigate the semi-decidability of the set .
Theorem 4. Let be finite alphabets with and . For all with , the sets are not semi-decidable.
Proof. Let
and
be arbitrary finite alphabets with
and
, and let
. First, we consider the case
. Let us consider the channel
It holds that
. For
, we define the channel
It holds that
for
. Let us now assume that
with
such that the set
is semi-decidable. Then, we consider the Turing machine
which accepts this set. Furthermore, we consider for
the following Turing machine
:
simulates two Turing machines and ;
In parallel, receives the input and tests if ;
stops if and only if .
It is shown in Lemma 1 that such a Turing machine exists. For the input , computes forever. The second Turing machine is defined by
;
For , it holds that ;
Therefore, stops for if and only if .
We now let stop for the input if and only if one of the two Turing machines or stops. Exactly one Turing machine has to stop for every .
If the Turing machine
stops at the input
, we set
. If the Turing machine
stops at the input
, we set
. Therefore, we have
We have shown in Lemma 1 that such a Turing machine cannot exist. This proves the theorem for
. The proof for
is very similar. □
For
and
, let
be the cardinality of a maximum code with a decoding error of 0. This maximum code always exists because we only have a finite set of possible codes for the blocklength
n. Of course, a well-defined function
exists for every
. Because of Fekete’s lemma, we have
We now have the following theorem regarding the Banach–Mazur computability of the function
.
Theorem 5. Let be finite alphabets with and . The functionis not Banach–Mazur-computable for all . Proof. Let
and
be finite alphabets with
and
, and let
be arbitrary. Consider the “ideal channel”
with
. Furthermore, consider any channel
with
for all
and all
. Then,
for all
, and consequently, because of (
14),
. Now, we can directly apply the proof of Theorem 3 to the function
.
is therefore not Banach–Mazur-computable. □
We now want to examine the question of whether a computable sequence of Banach–Mazur-computable lower bounds can be found for
. We set
For all
and for all
, we have
and
. However, this cannot be expressed algorithmically because due to Theorem 5, the functions
are not Banach–Mazur-computable. We next want to show that this is a general phenomenon for
.
Theorem 6. Let be finite alphabets with and . No computable sequence of Banach–Mazur-computable functions exists that simultaneously satisfies the following:
- 1.
For all , it holds that for all ;
- 2.
For all , it holds that .
Proof. Assume to the contrary that finite alphabets and exist with and , as well as a computable sequence of Banach–Mazur-computable functions, such that the following holds true:
For all and all , we have ;
For all , we have .
We consider for
the function
The function
is Banach–Mazur-computable (see [
37]). The sequence
is a computable sequence of Banach–Mazur-computable functions. For all
, it holds that
and
for
. Since the sequence
can be computed, we can find a Turing machine
, so that for all
and for all
,
applies (according the
Theorem [
43]). Let
be given arbitrarily with
. In addition, we also use the Turing machine
, which stops for the input
x if and only if
(see the proof of Theorem 4). Just as in the proof of Theorem 4, we use the two Turing machines
and
to build a Turing machine
. The Turing machine
stops exactly when
exists, so that
holds.
exists for
if and only if
. The set
will then be semi-decidable, which is a contradiction to the assumption. □
We make the following observation: For all with , the sets are semi-decidable. It holds that .
Theorem 7. The following three statements A, B, and C are equivalent:
- A
For all , where and are finite alphabets with and and for all with , the sets are semi-decidable.
- B
For all with , the sets are semi-decidable.
- C
For all , where and are finite alphabets with and and for all with , the sets are semi-decidable.
Proof. First, we show
. Let
with
. Then,
with
. Let
be finite sets with
and
. Then, the set
is semi-decidable by assumption. Let
be the associated Turing machine. Let
be chosen arbitrarily. From
, we algorithmically construct a channel
with
as follows. We consider the set
and an arbitrary output alphabet
with the bijection
. Therefore, it is obvious that
. For
, we set
. We define
Of course,
applies to the confusability graph of
. It holds that
. Therefore,
. Therefore,
stops. Conversely, if
stops for
, then
. Therefore,
. Thus, we have shown
.
Now, we show . Let and be finite alphabets with and . We construct a sequence of confusability graphs as follows.
For all pairs with , we start the first computation step of the Turing machine for in parallel for the number . That means for the input , we compute the first step of the calculation of . If the Turing machine stops at the input in the first step, then has the edge . If for with the Turing machine does not stop after the first step, then .
For
, we construct
as follows. For all
that have no edge in
, we let
calculate the second computation step at the input
. If
stops, then we set
to have the edge
and also receive all edges of
. If for
with
the Turing machine
does not stop after the second step, then
. We continue this process iteratively, generating a sequence of graphs
, all sharing the same vertex set, with the edges satisfying
. The Turing machine
stops for the input
if and only if
. We have a number of tests in each step that fall monotonically depending on
N (generally not strictly). It holds that
exists such that
is the confusability graph of
W. Note that we do not have a computable upper bound for
. However, the latter is not required for the proof. Therefore,
Let
be the Turing machine which accepts the set
. We have already shown that
holds if and only if
, so that the sequence
satisfies
. These are all graphs with the same set of nodes, and
with
exists. Furthermore, the sequence is computable. We only have to test for the sequence
, which is generated algorithmically from
, whether
applies. This means that we have to test whether
stops for a certain
n. We compute the first step for
. If the Turing machine stops, then
. Otherwise, we compute the second step for
and the first step for
. We continue recursively like this, and it is clear that the computation stops if and only if
. Otherwise, the Turing machine computes forever.
Now, we show . Let and let be arbitrarily chosen. From , we can effectively construct a DMC according to the Ahlswede approach (Theorem 1), so that . This means that if and only if . By assumption, the set is semi-decidable. We have used it to construct a Turing machine that stops when applies; otherwise , computes for ever. Therefore, C holds.
Now, we show . The idea of this part of the proof is similar to that of part . Let be arbitrary. Similar to case , we construct a suitable sequence of computable sequences of AVCs on and , such that the following assertions are satisfied:
For all
, we have
, as well as
for all
, all
, and all
.
exists such that
for all
and
for all
and all
.
The AVC then satisfies the requirements of Theorem 1.
In general, cannot be computed effectively depending on , but this is not a problem for the semi-decidability for all finite with and .
So, we have for
and it holds that
We can use this property and the semi-decidability requirement in C just like in the proof of
to construct a Turing machine
, which stops for
exactly then, if
applies or computes forever.
This proves the theorem. □
Remark 2 (See also
Section 1)
. Alon and Lubetzky have asked whether the set is semi-decidable (see [44]). We see that the answer to Alon and Lubetzky’s question is positive if and only if Assertion A from Theorem 7 holds true. This is interesting for the following reason: on the one hand, the set is semi-decidable for with , but on the other hand, even for and with , the set is not semi-decidable. So, there is no equivalence regarding the semi-decidability of these sets. In the next theorem, we look at useless channels in terms of the zero-error capacity. The set of useless channels is defined by
where
and
are finite alphabets with
and
. It is clear from our previous theorem that
is not semi-decidable.
Theorem 8. Let and be finite alphabets with and . Then, the set is semi-decidable.
Proof. For the proof of this theorem, we use the proof of Theorem 7. We have to construct a Turing machine as follows. is defined on and stops for an input W if and only if ; otherwise, it calculates forever. For the input W, we start the Turing machine in parallel for all with and test . We let all Turing machines compute one step of the computation in parallel. As soon as a Turing machine stops, it will not be continued. The Turing machine stops if and only if y exists for every with such that stops. Then, the confusability graph G is a complete graph, and consequently, and . □
In [
45], on the other hand, it was shown that the zero-error capacity for a fixed input and output alphabets can be calculated on a Blum–Shub–Smale machine.
5. The Computability of the Zero-Error Capacity with the Kolmogorov Oracle
We have shown that the zero-error capacity
is not Banach–Mazur-computable as a function of the channels. The question now arises as to whether a Turing machine with additional input can be found so that, for example, the upper bounds for the zero-error capacity can be calculated. This question will be briefly discussed in this section. In [
31], we showed that the zero-error capacity is semicomputable if we allow for the Kolmogorov oracle.
To define the Kolmogorov oracle, we need a special enumeration for
The problem is that the natural listing of the set of natural numbers is inappropriate because many numbers in
are too large for the natural enumerations. We start with the set of partially recursive functions from
to
. A listing
of the partial recursive functions is called an optimal listing if for any other recursive listing
of the set of recursive functions, there is a constant
such that for all
, the following holds:
exists with
and
. This means that all partial recursive functions
have a small Gödel number with respect to the system
. Schnorr [
34] has shown that such an optimal recursive listing of the set of partial recursive functions exists. The same holds true for the sets of natural numbers
.
For , let be an optimal listing and be a numbering of graphs.
For the set , we define This is the Kolmogorov complexity generated by and .
Definition 18. The Kolmogorov oracle is a function of in the power set of the set of graphs that produces a listfor each , where the graphs G are listed by size. Let be a Turing machine. We say that can use the oracle if for every , for the input n, the Turing machine acquires the list . With , we denote a Turing machine that has access to the oracle . We now consider for , the set , i.e., the -level set of the zero-error capacity. We have the following theorem:
Theorem 11 ([
31])
. Let . Then, the set is decidable with a Turing machine . This means a Turing machine exists such that the set is computable with this Turing machine with the oracle. Corollary 1 ([
31])
. Let , . Then, the set is semi-decidable for Turning machines with the oracle . Alon and Lubetzky have asked whether the set
is semi-decidable. We gave in [
31] a positive answer to this question on whether we can include the oracle. We do not know if
is computable concerning
.
Let be a number with . We set for . We have the following theorem:
Theorem 12. A Turing machine exists with such that for all , it holds that Thus, this approach does not directly provide the computability of through with the oracle . However, we can compute with any given accuracy.
We have seen that in order to prove the computability of
or
, we need computable converses. In this sense, the recent characterization by Zuiddam [
47] using the functions from the asymptotic spectrum of graphs is interesting.
6. Conclusions and Discussion
This paper revisited Ahlswede’s foundational approach in [
2] to characterizing the zero-error capacity using arbitrarily varying channels (AVCs). Although the theoretical connection remains intriguing, it has not yet yielded practical methods for calculating
for discrete memoryless channels (DMCs). Obstacles include the absence of explicit formulas for the maximum-error AVC capacities and the impossibility of algorithmically transforming any DMC into a finite 0–1 AVC, as shown by Theorems 3, 4, and 6. These results prove that no Turing machine can realize the map
Table 2 gives an overview of the main results of this paper.
Our focus has been on the computability of the zero-error capacity as a function of a DMC W; we did not address the computability questions arising from a graph-based perspective via the confusability graph . Whether the Shannon capacity is computable in that representation remains open.
This paper shows that the confusability graph derived from a channel’s transition matrix is not computable in general. This means that you cannot simply calculate the graph from the matrix data. Consequently, knowing the capacity of this confusability graph does not provide a concrete tool for evaluating the performance of the channel.
Furthermore, the capacity of the confusability graph is defined as a regularized limit, which makes it intrinsically difficult to evaluate in practice. In other words, this capacity is not given by a single, computable expression; instead, it emerges only in the limit of increasingly long codes or repeated graph operations. This regularization step complicates any attempt to actually compute or approximate the capacity, rendering it of limited utility in practical scenarios.
Nevertheless, because the descriptions of the DMC are standard in practical settings, our negative computability results are of broad significance.
Beyond coding theory, the zero-error capacity is relevant in areas like remote state estimation and quantum communication (see [
3,
48])
. Our findings are part of a broader narrative in information theory: many core problems—calculating the finite-state channel capacity [
49,
50,
51], optimizing mutual information, and even constructing capacity-achieving codes [
52,
53]—have been proven to be non–Turing-computable in general.
A compelling open question is whether computable channels
W exist for which
itself is a non-computable real. If so, this would establish that exact capacity statements require more than an algorithmic effort—they confront the fundamental limits of computability. Similar phenomena have been observed in compound channels [
54], colored Gaussian channels [
55], and Wiener prediction problems [
56], suggesting that rich, non-computable structures may also appear in zero-error contexts.
Moving forward, research should probe the algorithmic frontiers of zero-error information theory, especially in connection with automated systems and software-defined communications (see [
57,
58]). Though Turing computability hits a wall, other computational frameworks—such as Blum—Shub—Smale machines—may offer new possibilities [
45]. Understanding these alternative models may be key to effectively navigating the computability landscape of the zero-error capacity.
In summary, while the zero-error capacity remains a cornerstone of information theory, our results clarify that its algorithmic determination is blocked by deep, non-computable obstructions. Characterizing or circumventing these obstructions should be a key priority in future studies.