Next Article in Journal
Denoising Vanilla Autoencoder for RGB and GS Images with Gaussian Noise
Next Article in Special Issue
A Note on the Mixing Factor of Polar Codes
Previous Article in Journal
Quantifying Soil Complexity Using Fisher Shannon Method on 3D X-ray Computed Tomography Scans
Previous Article in Special Issue
Lossless Transformations and Excess Risk Bounds in Statistical Inference
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Variable-Length Resolvability for General Sources and Channels †

1
Department of Computer and Network Engineering, The University of Electro-Communications, Tokyo 182-8585, Japan
2
The National Institute of Information and Communications Technology (NICT), Tokyo 184-8795, Japan
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in the proceedings of 2017 IEEE International Symposium on Information Theory (ISIT2017), Aachen, Germany, titled “Variable-length resolvability for general sources”.
Entropy 2023, 25(10), 1466; https://doi.org/10.3390/e25101466
Submission received: 7 September 2023 / Revised: 30 September 2023 / Accepted: 3 October 2023 / Published: 19 October 2023
(This article belongs to the Special Issue Advances in Information and Coding Theory II)

Abstract

:
We introduce the problem of variable-length (VL) source resolvability, in which a given target probability distribution is approximated by encoding a VL uniform random number, and the asymptotically minimum average length rate of the uniform random number, called the VL resolvability, is investigated. We first analyze the VL resolvability with the variational distance as an approximation measure. Next, we investigate the case under the divergence as an approximation measure. When the asymptotically exact approximation is required, it is shown that the resolvability under two kinds of approximation measures coincides. We then extend the analysis to the case of channel resolvability, where the target distribution is the output distribution via a general channel due to a fixed general source as an input. The obtained characterization of channel resolvability is fully general in the sense that, when the channel is just an identity mapping, it reduces to general formulas for source resolvability. We also analyze the second-order VL resolvability.

1. Introduction

Generating a random number subject to a given probability distribution has a number of applications, such as in information security, statistical machine learning, and computer science. From the viewpoint of information theory, random number generation may be considered to be a transformation (encoding) of sequences emitted from a given source with coin distribution into other sequences with target distribution via a deterministic mapping [1,2,3]. Among others, there have been two major types of problems of random number generation: intrinsic randomness [4,5] and (source) resolvability [6,7]. In the former case, a fixed-length (FL) uniform random number is extracted from an arbitrary coin distribution, and we want to find the maximum achievable rate of such uniform random numbers. In the latter case, in contrast, an FL uniform random number used as a coin distribution is encoded to approximate a given target distribution, and we want to find the minimum achievable rate of such uniform random numbers. Thus, there is a duality between these two problems.
The problem of intrinsic randomness has been extended to the case of variable-length (VL) uniform random numbers, for which the length of random numbers may vary. This problem, referred to as the VL intrinsic randomness, was first introduced by Vembu and Verdú [5] for a finite source alphabet and later extended by Han [4] to a countably infinite alphabet. This problem was actually motivated because, in many practical situations, it is indispensable to consider cases where FL uniform random numbers are not available, and, instead, VL uniform random numbers are available; typically, this is in cases where we work with Elias’ universal random numbers [1]. The use of such uniform random numbers is expected generally to increase the achievable average length rate for intrinsic randomness. Then, the following natural question may be raised: Can we indeed lower the average length rate needed in the “resolvability” problem by using VL random numbers? The answer is “yes”. Despite the duality between these two kinds of problems for random number generation, the VL counterpart in the resolvability problem has not been discussed, where we focus on this problem.
We introduce the problem of VL source/channel resolvability, where a given target probability distribution is to be approximated by encoding a VL uniform random number. Distance measures between the target distribution and an approximated distribution are used to measure the fineness of the approximation. We first analyze the fundamental limit on VL source resolvabilities with the variational distance as an approximation measure in Section 3. We use the smooth Shannon entropy, which is a version of smooth Rényi entropy [8], to characterize the δ -source resolvability, which is defined as the minimum achievable length rate of uniform random numbers with an asymptotic distance of less than or equal to δ [ 0 , 1 ) . In the proof of the direct part, we will develop a simple version of information spectrum slicing [2], in which each “sliced” information density quantized to an integer is approximated by an FL uniform random number. Due to the simplicity of the method, the analysis with variational distance is first facilitated. As an important implication of general formulas for the δ -source resolvability, it is shown that the minimum resolvability rate of VL resolvability is equal to ( 1 δ ) times that of FL resolvability when the source is stationary and memoryless or is even with one-point spectrum (cf. Corollary 1). This result indicates an advantage of the use of a VL uniform random number when δ > 0 because we can make the VL resolvability rate strictly smaller than an FV one. We then extend these analyses to the case under the (unnormalized) divergence as an approximation measure in Section 4. When δ = 0 , that is, when the asymptotically exact approximation is required, it is shown that the 0-source resolvabilities under two kinds of approximation measures coincide with each other.
In Section 5, we then consider the problem of channel resolvability [6,9,10], in which not only a source but also a channel is fixed, and the output distribution via the channel is now the target of approximation. This problem, also referred to as the problem of output approximation, provides a powerful tool to analyze the fundamental limits of various problems in information theory and information security. Some such examples include identification codes [11,12,13], distributed hypothesis testing [14], message authentication [15], secret key generation [16], and coding for secure communication [17,18,19]. We consider two types of problems in which either a general source (mean-channel resolvability) or a VL uniform random number (VL channel resolvability) is used as a coin distribution. It is shown that the formulas established are equal for both coin distributions. In the special case that the channel is the identity mapping, the formulas established reduce to those in source resolvability as established in Section 3 and Section 4.
From Section 3, Section 4 and Section 5, the so-called first-order resolvability rates are analyzed, and the next important step may be the second-order analysis. Second-order analyses for various coding problems were initiated by Strassen [20] and have been studied in the past decade or so (cf. [21,22,23,24,25,26,27,28,29,30,31]). We also analyze the second-order fundamental limits of the VL channel/source resolvability in Section 6. In this paper, it is shown that the VL δ -source resolvability under the variational distance is equal to the minimum achievable rate of fixed-to-variable-length source codes with an error probability of less than or equal to δ . It is demonstrated that this close relationship provides a single-letter characterization for the first- and second-order source resolvability under the variational distance when the source is stationary and memoryless. It is worth noting that second-order analyses for the VL setting are relatively few, compared to those in the FL setting. The second-order formulas established in this paper are of importance from this perspective, too.
The remainder of the paper is organized as follows: Section 2 reviews the problem of FL source resolvability and the relations between the minimum resolvability rate and the minimum coding rate of FL source codes. Section 3 formally introduces the problem of VL source resolvability with variational distance as an approximation measure. Then, Section 4 discusses VL source resolvability with divergence as an approximation measure, and Section 5 generalizes the settings to channel resolvability. Section 6 investigates the second-order fundamental limits of the VL channel/source resolvability. Section 7 concludes the paper with a discussion of possible extensions.

2. FL Resolvability: Review

Let U = { 1 , 2 , , K } be a finite alphabet of size K, and let X be a finite or countably infinite alphabet. Let X = { X n } n = 1 be a general source [2], where P X n is a probability distribution on X n . We do not impose any assumptions, such as stationarity or ergodicity. In this paper, we identify X n with its probability distribution P X n , and these symbols are used interchangeably.
We first review the problem of FL (source) resolvability [2] using the variational distance as an approximation measure. Let U M n denote the uniform random number, which is a random variable uniformly distributed over U M n : = { 1 , , M n } . Consider the problem of approximating the target distribution  P X n by using U M n as a coin distribution via a deterministic mapping φ n : { 1 , , M n } X n . Denoting X ˜ n = φ n ( U M n ) , we want to make P X ˜ n approximate P X n (cf. Figure 1). A standard choice of the performance measure for approximation is
d ( P X n , P X ˜ n ) : = 1 2 x X n | P X n ( x ) P X ˜ n ( x ) | ,
which is referred to as the variational distance between P X n and P X ˜ n . It is easily seen that
0 d ( P X n , P X ˜ n ) 1 .
Let us now review the problem for source resolvability. Throughout this paper, logarithms are of the base K.
Definition 1
(FL resolvability). A resolution rate R 0 is said to be FL achievable or simply f-achievable (under the variational distance) if there exists a deterministic mapping φ n : { 1 , , M n } X n satisfying
lim sup n 1 n log M n R ,
lim n d ( P X n , P X ˜ n ) = 0 ,
where X ˜ n = φ n ( U M n ) and U M n is the uniform random number over U M n . The infimum of f-achievable rates, i.e.,
S f ( X ) : = inf { R : R i s   f - a c h i e v a b l e }
is called the FL resolvability or simply f-resolvability.
Then, we have the following theorem:
Theorem 1
(Han and Verdú [6]). For any general target source X ,
S f ( X ) = H ¯ ( X ) ,
where
H ¯ ( X ) : = inf a : lim n Pr 1 n log 1 P X n ( X n ) > a = 0 .
Remark 1.
As a dual counterpart of (7), we may define
H ̲ ( X ) : = sup b : lim n Pr 1 n log 1 P X n ( X n ) < b = 0 .
Sources such that H ¯ ( X ) = H ̲ ( X ) are called one-point spectrum sources (or equivalently, said to satisfy the strong converse property (cf. Han [2])), which includes stationary memoryless sources and stationary ergodic sources, etc. This class of sources is discussed later in Corollary 1.
The following problem is called the δ-resolvability problem [2,7], which relaxes the condition on the variational distance, compared to (4).
Definition 2
(FL δ -resolvability). For a fixed δ [ 0 , 1 ) , a resolution rate R 0 is said to be FL δ-achievable or simply f ( δ ) -achievable (under the variational distance) if there exists a deterministic mapping φ n : { 1 , , M n } X n satisfying
lim sup n 1 n log M n R ,
lim sup n d ( P X n , P X ˜ n ) δ ,
where X ˜ n = φ n ( U M n ) and U M n is the uniform random number over U M n . The infimum of all f ( δ ) -achievable rates, i.e.,
S f ( δ | X ) : = inf { R : R i s   f ( δ ) - a c h i e v a b l e }
is referred to as the FL δ-resolvability or simply f ( δ ) -resolvability.
Then, a characterization of S f ( δ | X ) is given by
Theorem 2
(Steinberg and Verdú [7]). For any general target source X ,
S f ( δ | X ) = H ¯ δ ( X ) ( δ [ 0 , 1 ) ) ,
where
H ¯ δ ( X ) : = inf a : lim sup n Pr 1 n log 1 P X n ( X n ) > a δ .
Remark 2.
The FL resolvability problem is deeply related to the FL source coding problem allowing a probability of a decoding error up to ε. Denoting by R f ( ε | X ) the minimum achievable rate for the source X , there is the relationship [7]:
R f ( ε | X ) = H ¯ ε ( X ) ( ε [ 0 , 1 ) )
and, hence, by Theorem 2,
S f ( δ | X ) = R f ( δ | X ) ( δ [ 0 , 1 ) ) .
Formula (14) can also be shown with a smooth Rényi entropy of order zero [32].

3. VL Resolvability: Variational Distance

In this section, we introduce the problem of variable-length (VL) resolvability, where the target probability distribution is approximated by encoding a VL uniform random number. As an initial step, we analyze the fundamental limit on the VL resolvability with the variational distance as an approximation measure.

3.1. Definitions

Let U * denote the set of all sequences u U m over m = 0 , 1 , 2 , , where U 0 = { λ } ( λ is the null string). Let L n denote a random variable which takes a value in { 0 , 1 , 2 , } . We define the VL uniform random number  U ( L n ) so that U ( m ) is uniformly distributed over U m given L n = m . In other words,
P U ( L n ) ( u , m ) : = Pr { U ( L n ) = u , L n = m } = Pr { L n = m } K m ( u U m ) ,
Pr { U ( L n ) = u | L n = m } = P U ( L n ) ( u , m ) Pr { L n = m } = 1 K m ( u U m ) ,
where K = | U | . It should be noticed that the VL sequence u U m is generated with the joint probability P U ( L n ) ( u , m ) .
We formally define the δ -resolvability problem under the variational distance using the VL random number, called the VL δ-resolvability or simply v ( δ ) -resolvability.
Definition 3
(VL δ -resolvability: variational distance). A resolution rate R 0 is said to be VL δ-achievable (under the variational distance) with δ [ 0 , 1 ) if there exists a VL uniform random number U ( L n ) and a deterministic mapping φ n : U * X n satisfying
lim sup n 1 n E [ L n ] R ,
lim sup n d ( P X n , P X ˜ n ) δ ,
where E [ · ] denotes the expected value and X ˜ n = φ n ( U ( L n ) ) . The infimum of all v ( δ ) -achievable rates, i.e.,
S v ( δ | X ) : = inf { R : R i s   v ( δ ) - a c h i e v a b l e }
is referred to as the VL δ-resolvability or simply v ( δ ) -resolvability.
If δ = 0 , v ( 0 ) -achievable is said to be VL achievable or simply v-achievable (under the variational distance). The infimum of all v-achievable rates, i.e.,
S v ( X ) : = inf { R : R i s   v - a c h i e v a b l e }
is called the VL resolvability or simply v-resolvability.
Remark 3.
One may think that condition (18) can be replaced with the condition on the sup-entropy rate:
lim sup n 1 n H ( U ( L n ) ) R
as in [6], where H ( · ) denotes the Shannon entropy. Indeed, both conditions yield the same resolvability result. To see this, let us denote by S ˜ v ( δ | X ) the infimum of v-achievable rates R under constraints (19) and (22). It is easily checked that
E [ L n ] = m = 1 u U m P U ( L n ) ( u , m ) log K m = m = 1 u U m P U ( L n ) ( u , m ) log Pr { L n = m } P U ( L n ) ( u , m ) = H ( U ( L n ) ) H ( L n ) H ( U ( L n ) ) .
This implies S v ( δ | X ) S ˜ v ( δ | X ) . On the other hand, by invoking the well-known relation (cf. ([33], Corollary 3.12)), it holds that
H ( L n ) log ( e · E [ L n ] ) .
Consider any resolution rate R > S v ( δ | X ) . Then, (18) holds for some U ( L n ) and φ n and, hence, (24) leads to
lim n 1 n H ( L n ) = lim n 1 n log ( e · E [ L n ] ) = 0 .
From this equation, (23) yields that
lim sup n 1 n H ( U ( L n ) ) = lim sup n 1 n E [ L n ] R
to obtain R S ˜ v ( δ | X ) , implying that S v ( δ | X ) S ˜ v ( δ | X ) . Thus, S v ( δ | X ) = S ˜ v ( δ | X ) .

3.2. Smooth Shannon Entropy

To establish a general formula for S v ( δ | X ) , we introduce the following quantity for a general source X . Let P ( X n ) denote the set of all probability distributions on X n . For δ [ 0 , 1 ) , by defining the δ-ball using the variational distance
B δ ( X n ) = P V n P ( X n ) : d ( P X n , P V n ) δ ,
we introduce the smooth Shannon entropy:
H [ δ ] ( X n ) : = inf P V n B δ ( X n ) x X n P V n ( x ) log 1 P V n ( x ) = inf P V n B δ ( X n ) H ( V n ) ,
where H ( V n ) denotes the Shannon entropy of P V n . The H [ δ ] ( X n ) is a nonincreasing function of δ . Based on this quantity for a general source X = { X n } n = 1 , we define
H [ δ ] ( X ) = lim sup n 1 n H [ δ ] ( X n ) .
Remark 4.
Renner and Wolf [8] have defined the smooth Rény entropy of order α ( 0 , 1 ) ( 1 , ) as
H [ δ ] α ( X n ) = inf P V n B δ ( X n ) 1 1 α log x X n P V n ( x ) α .
By letting α 1 , we have
lim α 1 H [ δ ] α ( X n ) = H [ δ ] ( X n ) .
As for the proof, see Appendix A.

3.3. General Formula for General δ [ 0 , 1 )

The following main theorem indicates that the v ( δ ) -resolvability S v ( δ | X ) can be characterized by the smooth Shannon entropy for X .
Theorem 3.
For any general target source X ,
S v ( δ | X ) = lim γ 0 H [ δ + γ ] ( X ) ( δ [ 0 , 1 ) ) .
Remark 5.
In Formula (32), the limit lim γ 0 of the offset term + γ appears in the characterization of S v ( δ | X ) . This is because the smooth entropy H [ δ ] ( X n ) for X n involves the infimum over the non-asymptotic δ-ball B δ ( X n ) for a given length n. Alternatively, we may consider the asymptotic δ-ball defined as
B δ ( X ) = V = { V n } n = 1 : lim sup n d ( P X n , P V n ) δ ,
and then we obtain the alternative formula
S v ( δ | X ) = inf V B δ ( X ) H ( V ) ( δ [ 0 , 1 ) )
without an offset term, where
H ( V ) : = lim sup n 1 n H ( V n )
is the sup-entropy rate for V with the Shannon entropy H ( V n ) . The proof of (34) is given in Appendix B.
The same remark also applies to general formulas to be established in the subsequent sections.
Remark 6.
Independently of this work, Tomita, Uyematsu, and Matsumoto [34] have investigated the following problem: the coin distribution is given by fair coin-tossing and the average number of coin tosses should be asymptotically minimized as in [3] while the variational distance between the target and approximated distributions should satisfy (19). In this case, the asymptotically minimum average number of coin tosses is also characterized by the right-hand side (r.h.s.) of (32) (cf. [34]). Since the coin distribution is restricted to that given by fair coin-tossing with a stopping algorithm, realizations of L n must satisfy the Kraft inequality (for prefix codes), whereas the problem addressed in this paper allows the probability distribution of L n to be an arbitrary discrete one, not necessarily implying prefix codes. In this sense, our problem is more relaxed, while the coin is constrained to be conditionally independent given L n . Theorem 3 indicates that the v ( δ ) -resolvability does not differ in both problems. Later, we shall show that, even in the case where the coin distribution may be any general source X , the δ-resolvability remains the same (cf. Theorem 5 and Remark 14).
On the other hand, we now define the following information quantity (to be used in Remark 7 below) to discuss the relationship with VL source coding: For δ [ 0 , 1 ) we define
G [ δ ] ( X n ) = inf A n X n : Pr { X n A n } 1 δ x A n P X n ( x ) log 1 P X n ( x ) .
The G [ δ ] ( X n ) is a nonincreasing function of δ . Based on this quantity, we define
G [ δ ] ( X ) = lim sup n 1 n G [ δ ] ( X n ) .
Then, we have
Remark 7.
There is a deep relation between the δ-resolvability problem and VL δ-source coding with the error probability asymptotically not exceeding δ. Koga and Yamamoto [35] (also, cf. Han [36]) showed that the minimum average length rate R v * ( δ | X ) of VL δ-source codes is given by
R v * ( δ | X ) = lim γ 0 G [ δ + γ ] ( X ) ( δ [ 0 , 1 ) ) .
Theorem 3 and Proposition 1 (to be shown just below) reveal that
S v ( δ | X ) = R v * ( δ | X ) ( δ [ 0 , 1 ) ) .
The following proposition shows a general relationship between G [ δ ] ( X ) and H [ δ ] ( X ) .
Proposition 1.
For any general source X ,
H [ δ ] ( X ) = G [ δ ] ( X ) ( 1 δ ) H ¯ δ γ ( X ) ( δ ( 0 , 1 ) , γ ( 0 , δ ] ) .
In particular,
lim γ 0 H [ δ + γ ] ( X ) = lim γ 0 G [ δ + γ ] ( X ) ( 1 δ ) H ¯ δ ( X ) ( δ [ 0 , 1 ) ) .
(Proof) See Appendix C.
By plugging γ = δ into (40), a looser but sometimes useful bound
H [ δ ] ( X ) = G [ δ ] ( X ) ( 1 δ ) H ¯ ( X )
can be obtained. Equation (40) has been derived by [21], which improves a bound established in [35,37]. In view of Theorems 2 and 3, (41) in Proposition 1 implies
S v ( δ | X ) ( 1 δ ) S f ( δ | X )
for all δ [ 0 , 1 ) , where S f ( δ | X ) denotes the f ( δ ) -resolvability. This general relationship elucidates the advantage of the use of VL uniform random numbers to make the average length rate lower. The proposition also claims that G [ δ ] ( X ) coincides with H [ δ ] ( X ) for all δ [ 0 , 1 ) for any general source X .
A consequence of Theorem 3 is the following corollary:
Corollary 1.
Let X = { X n } n = 1 be a one-point spectrum source ( H ¯ ( X ) = H ̲ ( X ) ) with X n = ( X 1 , X 2 , , X n ) , then we have
S f ( δ | X ) = H ¯ δ ( X ) = H * ( X ) ( δ [ 0 , 1 ) ) ,
where H * ( X ) : = H ¯ ( X ) = H ̲ ( X ) . Moreover, it holds that
S v ( δ | X ) = H [ δ ] ( X ) = G [ δ ] ( X ) = ( 1 δ ) H * ( X ) ( δ [ 0 , 1 ) ) ,
where we notice that H * ( X ) = H ( X 1 ) (entropy) for stationary memoryless sources.
(Proof) See Appendix D.
Now, we are ready to give the proof of Theorem 3:
Proof of Theorem 3.
(1)
Converse Part:
Let R be v ( δ ) -achievable. Then, there exists U ( L n ) and φ n satisfying (18) and
lim sup n δ n δ ,
where we define δ n = d ( P X n , P X ˜ n ) with X ˜ n = φ n ( U ( L n ) ) . Equation (46) implies that, for any given γ > 0 , it holds that δ n δ + γ for all n n 0 with some n 0 > 0 , and thus we have
H [ δ + γ ] ( X n ) H [ δ n ] ( X n ) ( n n 0 ) ,
because H [ δ ] ( X n ) is a nonincreasing function of δ . Since P X ˜ n B δ n ( X n ) , we have
H [ δ n ] ( X n ) H ( X ˜ n ) .
On the other hand, it follows from (23) that
H ( X ˜ n ) H ( U ( L n ) ) = E [ L n ] + H ( L n ) ,
where the inequality is due to the fact that φ n is a deterministic mapping and X ˜ n = φ n ( U ( L n ) ) .
Combining (47)–(49) yields
H [ δ + γ ] ( X ) = lim sup n 1 n H [ δ + γ ] ( X n ) lim sup n 1 n E [ L n ] + lim sup n 1 n H ( L n ) R ,
where we have used (18) and (25) for the last inequality. Since γ > 0 is arbitrary, we obtain
lim γ 0 H [ δ + γ ] ( X ) R .
(2)
Direct Part:
Without loss of generality, we assume that H + : = lim γ 0 H [ δ + γ ] ( X ) is finite ( H + < + ). Letting R = H + + 3 γ , where γ > 0 is an arbitrary constant, we shall show that R is v ( δ ) -achievable. In what follows, we use a simpler form of information spectrum slicing [2], where each piece of sliced information quantized to a positive integer is approximated by the uniform random number U ( ) of the length .
First, we note that
H + H [ δ + γ ] ( X ) 1 n H [ δ + γ ] ( X n ) γ ( n > n 0 )
because of the monotonicity of H [ δ ] ( X ) in δ . Let V n be a random variable subject to P V n B δ + γ ( X n ) , which satisfies
H [ δ + γ ] ( X n ) + γ H ( V n ) .
For γ > 0 , we can choose a c n > 0 so large that
Pr { V n T n } γ
where
T n : = x X n : 1 n log 1 P V n ( x ) c n .
We also define
( x ) : = log 1 P V n ( x ) + n γ f o r x T n 0 otherwise .
For m = 0 , 1 , , β n : = n ( c n + γ ) , set
S n ( m ) : = x X n : ( x ) = m ,
then these sets form a partition of X n , i.e.,
m = 0 β n S n ( m ) = X n a n d m = 1 β n S n ( m ) = T n .
We set L n so that
Pr { L n = m } = Pr { V n S n ( m ) } ,
where it is obvious that m = 0 β n Pr { L n = m } = 1 , and, hence, the probability distribution of the VL uniform random number U ( L n ) is given as
P U ( L n ) ( u , m ) : = Pr { U ( L n ) = u , L n = m } = Pr { V n S n ( m ) } K m ( u U m ) .
(a)
Construction of Mapping φ n : U * X n :
Index the elements in S n ( m ) as x 1 , x 2 , , x | S n ( m ) | ( m = 1 , 2 , ) , where
| S n ( m ) | K m n γ
since for x S n ( m )
log 1 P V n ( x ) m n γ P V n ( x ) K ( m n γ ) ,
and, therefore,
1 x S n ( m ) P V n ( x ) x S n ( m ) K ( m n γ ) = | S n ( m ) | K ( m n γ ) .
For i = 1 , 2 , , | S n ( m ) | , define A ˜ i ( m ) U m as the set of sequences u U m so that
u A ˜ i ( m ) P U ( L n ) ( u , m ) P V n ( x i ) < u A ˜ i ( m ) P U ( L n ) ( u , m ) + Pr { V n S n ( m ) } K m
and
A ˜ i ( m ) A ˜ j ( m ) = ( i j ) .
If
i = 1 | S n ( m ) | u A ˜ i ( m ) P U ( L n ) ( u , m ) < i = 1 | S n ( m ) | P V n ( x i ) = Pr { V n S n ( m ) } ,
then add a u i U m ( j A ˜ j ( m ) ) to obtain
A i ( m ) = A ˜ i ( m ) { u i }
for i = 1 , 2 , in order, until it holds that with some 1 c | S n ( m ) |
i = 1 c A i ( m ) i = c + 1 | S n ( m ) | A ˜ i ( m ) = U m ,
where u 1 , u 2 , are selected to be all different. Since | U m | = K m and
u U m P U ( L n ) ( u , m ) = u U m Pr { V n S n ( m ) } K m = Pr { V n S n ( m ) } ,
such a 1 c | S n ( m ) | always exists. For simplicity, we set for i = c + 1 , c + 2 , , | S n ( m ) |
A i ( m ) = A ˜ i ( m )
and for i = 1 , 2 , , | S n ( m ) |
φ n ( u ) = x i f o r u A i ( m ) ,
which defines the random variable X ˜ n with values in X n such that
P X ˜ n ( x i ) = u A i ( m ) P U ( L n ) ( u , m ) ( x i S n ( m ) ) ,
that is, X ˜ n = φ n ( U ( L n ) ) , where if X n T n , we choose some x 0 X n T n and set
P X ˜ n ( x 0 ) = Pr { V n T n } and φ n ( λ ) = x 0 .
Notice that, by this construction, we have
| P X ˜ n ( x i ) P V n ( x i ) | Pr { V n S n ( m ) } K m
for i = 1 , 2 , , | S n ( m ) | ; m = 1 , 2 , , β n , and
Pr { X ˜ n T n } = Pr { V n T n } γ .
(b)
Evaluation of Average Length:
Since m = 0 does not contribute to the average length E [ L n ] , it is evaluated as follows:
E [ L n ] = m = 1 β n u U m P U ( L n ) ( u , m ) · m = m = 1 β n i = 1 | S n ( m ) | u A i ( m ) P U ( L n ) ( u , m ) · m = m = 1 β n x i S n ( m ) P X ˜ n ( x i ) · m ,
where we have used U m = i = 1 | S n ( m ) | A i ( m ) and (72). For x i S n ( m ) , we obtain from (74)
P X ˜ n ( x i ) P V n ( x i ) + Pr { V n S n ( m ) } K m P V n ( x i ) 1 + 1 P V n ( x i ) K m P V n ( x i ) 1 + 1 K n γ ,
where, to derive the last inequality, we have used (62). Plugging the inequality
m log 1 P V n ( x i ) + n γ + 1 ( x i S n ( m ) )
and (77) into (76), we obtain
E [ L n ] 1 + 1 K n γ m = 1 β n x i S n ( m ) P V n ( x i ) log 1 P V n ( x i ) + n γ + 1 1 + 1 K n γ H ( V n ) + n γ + 1 ,
which yields
lim sup n 1 n E [ L n ] lim sup n 1 n H ( V n ) + 2 γ lim sup n 1 n H [ δ + γ ] ( X n ) + 3 γ = H [ δ + γ ] ( X ) + 3 γ H + + 3 γ = R ,
where the second inequality follows from (53) and the last one is due to (52).
(c)
Evaluation of Variational Distance:
From (61) and (74), we have
x S n ( m ) | P X ˜ n ( x ) P V n ( x ) | | S n ( m ) | Pr { V n S n ( m ) } K m Pr { V n S n ( m ) } K n γ ,
which, in view of (58), leads to
d ( P X ˜ n , P V n ) = 1 2 x T n | P X ˜ n ( x ) P V n ( x ) | + 1 2 x T n | P X ˜ n ( x ) P V n ( x ) | 1 2 m = 1 β n x S n ( m ) | P X ˜ n ( x ) P V n ( x ) | + 1 2 Pr { X ˜ n T n } + Pr { V n T n } 1 2 m = 1 β n Pr { V n S n ( m ) } K n γ + γ 1 2 K n γ + γ ,
where we have used (75) to obtain the leftmost inequality in (82). By the triangle inequality, we obtain
d ( P X n , P X ˜ n ) d ( P X n , P V n ) + d ( P X ˜ n , P V n ) δ + 2 γ + 1 2 K n γ ,
where the last inequality follows because P V n B δ + γ ( X n ) . Thus, we obtain from (83)
lim sup n d ( P X n , P X ˜ n ) δ + 2 γ .
Since γ > 0 is arbitrary and we have (80), we conclude that R is v ( δ ) -achievable.

3.4. General Formula for δ = 0

In this subsection, we consider the special case with δ = 0 . In this case, we can elucidate the relationship between the minimum achievable rates for VL source codes with an asymptotically vanishing decoding error probability and the FL source codes.
We obtain the following corollary from Theorem 3 and Proposition 1:
Corollary 2.
For any general target source X ,
S v ( X ) = lim γ 0 G [ γ ] ( X ) ,
where G [ γ ] ( X ) is defined in (37).
It has been shown by Han [2] that any source X = { X n } n = 1 satisfying the uniform integrability (cf. Han [2]) satisfies
lim γ 0 G [ γ ] ( X ) = H ( X ) : = lim sup n 1 n H ( X n ) ,
where H ( X ) is called the sup-entropy rate. Notice here, in particular, that the finiteness of an alphabet implies the uniform integrability [2]. Thus, we obtain the following corollary:
Corollary 3.
For any finite alphabet target source X ,
S v ( X ) = H ( X ) .
Remark 8.
As in the case of FL resolvability and FL source coding problems, S v ( X ) is tightly related to VL source codes with vanishing decoding error probabilities. Denoting by R v * ( X ) the minimum error-vanishing VL -achievable rate for a source X , Han [36] has shown that
R v * ( X ) = lim γ 0 G [ γ ] ( X ) ,
and, hence, from Corollary 3, it is concluded that
S v ( X ) = R v * ( X ) .
In addition, if a general source X  satisfies the uniform integrability and the strong converse property (cf. Han [2]), then equation (86) holds and hence it follows from ([2], Theorem 1.7.1) that
S f ( X ) = S v ( X ) = R v * ( X ) = R v ( X ) = R f ( X ) = H ( X ) ,
where R f ( X ) : = R f ( 0 | X ) and R v ( X ) denotes the minimum achievable rate of VL source codes with zero error probabilities for all n = 1 , 2 , .
Remark 9.
Han and Verdú [6] have discussed the problem of mean-resolvability for the target distribution P X n . In this problem, the coin distribution may be a general source X ˜ = { X ˜ n } n = 1 , where X ˜ n is a random variable that takes values in X n with the average length rate 1 n E [ L n ] in (18) replaced with the entropy rate 1 n H ( X ˜ n ) . Denoting by S ¯ v ( X ) the mean-resolvability, which is defined as the infimum of v-achievable rates for a general source X (with the countably infinite alphabet), we can easily verify that any mean-resolution rate R > S ¯ v ( X ) must satisfy R lim γ 0 G [ γ ] ( X ) so that S v ( X ) S ¯ v ( X ) . On the other hand, S v ( X ) S ¯ v ( X ) by definition. Thus, in view of Corollary 2, we have
Corollary 4.
For any general target source X ,
S v ( X ) = S ¯ v ( X ) = lim γ 0 G [ γ ] ( X ) .

4. VL Resolvability: Divergence

So far, we have considered the problem of VL resolvability, in which the approximation level is measured by the variational distance between X n and X ˜ n . It is sometimes of use to deal with another quantity as an approximation measure. In this section, we use the (unnormalized) divergence as the approximation measure.

4.1. Definitions

In this subsection, we address the following problem.
Definition 4
(VL δ -resolvability: divergence). A resolution rate R 0 is said to be VL δ-achievable or simply v D ( δ ) -achievable (under the divergence) with δ 0 , if there exists a VL uniform random number U ( L n ) and a deterministic mapping φ n : U * X n satisfying
lim sup n 1 n E [ L n ] R ,
lim sup n D ( X ˜ n | | X n ) δ ,
where X ˜ n = φ n ( U ( L n ) ) and D ( X ˜ n | | X n ) denotes the divergence between P X ˜ n and P X n :
D ( X ˜ n | | X n ) = x X n P X ˜ n ( x ) log P X ˜ n ( x ) P X n ( x ) .
The infimum of all v D ( δ ) -achievable rates, i.e.,
S v D ( δ | X ) : = inf { R : R i s   v D ( δ ) - a c h i e v a b l e }
is called the VL δ -resolvability or simply the v D ( δ ) -resolvability.
Remark 10.
The measure of approximation is now the divergence D ( X ˜ n | | X n ) but not its reversed version D ( X n | | X ˜ n ) . In the context of resolvability, divergence D ( X ˜ n | | X n ) (and its counterpart in the case of channel resolvability) is usually employed as in [6,7,9]. We also use this type of divergence in the subsequent sections.
To establish the general formula for S v D ( δ | X ) , we introduce the following quantity for a general source X = { X n } n = 1 . Recall that P ( X n ) denotes the set of all probability distributions on X n . For δ 0 , defining the δ-ball using the divergence as
B δ D ( X n ) = P V n P ( X n ) : D ( V n | | X n ) δ ,
we introduce the following quantity, referred to as the smooth entropy using the divergence:
H [ δ ] D ( X n ) : = inf P V n B δ D ( X n ) H ( V n ) ,
where H ( V n ) denotes the Shannon entropy of P V n . Obviously, H [ δ ] D ( X n ) is a nonincreasing function of δ . Now, we define
H [ δ ] D ( X ) = lim sup n 1 n H [ δ ] D ( X n ) .
The following lemma is used to derive Corollary 5 of Theorem 4 below in the next subsection.
Lemma 1.
For any general source X ,
H [ δ ] ( X ) H [ g ( δ ) ] D ( X ) ( δ 0 ) ,
where we define g ( δ ) = 2 δ 2 / ln K , and
lim δ 0 G [ δ ] ( X ) = lim δ 0 H [ δ ] ( X ) = lim δ 0 H [ δ ] D ( X ) H ¯ ( X ) .
(Proof) See Appendix E.

4.2. General Formula

Here, we establish another main theorem, which characterizes S v D ( δ | X ) for all δ 0 in terms of the smooth entropy using the divergence.
Theorem 4.
For any general target source X ,
S v D ( δ | X ) = lim γ 0 H [ δ + γ ] D ( X ) ( δ 0 ) .
Remark 11.
It should be noticed that the approximation measure considered here is not the normalized divergence
1 n D ( φ n ( U ( L n ) ) | | X n ) ,
which has been used in the problem of FL δ-resolvability [7]. The achievability scheme given in the proof of the direct part of Theorem 4 can also be used in the case of this relaxed measure. Indeed, denoting the VL δ-resolvability with the normalized divergence by S ˜ v D ( δ | X ) , the general formula for S ˜ v D ( δ | X ) is given in the same form as (101), if the radius of the δ-ball B δ D ( X n ) in the definition of H [ δ ] D ( X n ) is replaced with the normalized divergence. It generally holds that S v D ( δ | X ) S ˜ v D ( δ | X ) for all δ 0 because the normalized divergence is smaller than the unnormalized divergence.
As we have seen in Lemma 1, we generally have S v D ( g ( δ ) | X ) S v ( δ | X ) for any δ [ 0 , 1 ) with g ( δ ) = 2 δ 2 / ln K . In particular, in the case that δ = 0 , we obtain the following corollary of Theorems 3 and 4.
Corollary 5.
For any general target source X ,
S v D ( 0 | X ) = S v ( X ) .
Corollary 5 indicates that the v D ( 0 ) -resolvability S v D ( 0 | X ) coincides with the v -resolvability S v ( X ) and is also characterized by the r.h.s. of (85). By (88), it also implies that S v D ( 0 | X ) = R v * ( X ) , where R v * ( X ) denotes the minimum error-vanishing achievable rate with VL source codes for X .
Proof of Theorem 4.
(1)
Converse Part:
Let R be v D ( δ ) -achievable. Then, there exists U ( L n ) and φ n satisfying (92) and
lim sup n δ n δ ,
where we define δ n = D ( X ˜ n | | X n ) with X ˜ n = φ n ( U ( L n ) ) . Equation (104) implies that, for any given γ > 0 , it holds that δ n δ + γ for all n n 0 with some n 0 > 0 , and therefore
H [ δ + γ ] D ( X n ) H [ δ n ] D ( X n ) ( n n 0 )
since H [ δ ] D ( X n ) is a nonincreasing function of δ . Since P X ˜ n B δ n D ( X n ) , we have
H [ δ n ] D ( X n ) H ( X ˜ n ) .
On the other hand, it follows from (23) that
H ( X ˜ n ) H ( U ( L n ) ) = E [ L n ] + H ( L n ) ,
where the inequality is due to the fact that φ n is a deterministic mapping and X ˜ n = φ n ( U ( L n ) ) .
Combining (105)–(107) yields
H [ δ + γ ] D ( X ) = lim sup n 1 n H [ δ + γ ] D ( X n ) lim sup n 1 n E [ L n ] + lim sup n 1 n H ( L n ) R ,
where we used (25) and (92) for the last inequality. Since γ > 0 is arbitrary, we have
lim γ 0 H [ δ + γ ] D ( X ) R .
(2)
Direct Part:
We modify the achievability scheme in the proof of the direct part of Theorem 3. Although the proof of this part is quite similar to that of the direct part of Theorem 3, we give here the full proof in order to avoid subtle possible confusions. We may assume that H + : = lim γ 0 H [ δ + γ ] D ( X ) is finite ( H + < + ). Letting R = H + + μ , where μ > 0 is an arbitrary constant, we shall show that R is v D ( δ ) -achievable.
Let V n be a random variable subject to P V n B δ + γ D ( X n ) satisfying
H [ δ + γ ] D ( X n ) + γ H ( V n )
for any fixed γ ( 0 , 1 2 ] . We can choose a c n > 0 so large that
γ 0 : = Pr { V n T n } γ
where
T n : = x X n : 1 n log 1 P V n ( x ) c n .
We also define
( x ) : = log 1 P V n ( x ) + n γ f o r x T n .
Letting, for m = 1 , 2 , , β n : = n ( c n + γ ) ,
S n ( m ) : = x X n : ( x ) = m ,
these sets form a partition of T n :
m = 1 β n S n ( m ) = T n .
We set L n so that
Pr { L n = m } = Pr { V n S n ( m ) } Pr { V n T n } = Pr { V n S n ( m ) } 1 γ 0 ,
which satisfies
m = 1 β n Pr { L n = m } = Pr { V n T n } 1 γ 0 = 1 ,
and, hence, the probability distribution of U ( L n ) is given as
P U ( L n ) ( u , m ) : = Pr { U ( L n ) = u , L n = m } = Pr { V n S n ( m ) } ( 1 γ 0 ) K m ( u U m ) .
(a)
Construction of Mapping φ n : U * X n :
Index the elements in S n ( m ) as x 1 , x 2 , , x | S n ( m ) | ( m = 1 , 2 , , β n ) , where it holds that
| S n ( m ) | K m n γ
(cf. (61)–(63)). For i = 1 , 2 , , | S n ( m ) | , define A ˜ i ( m ) U m as the set of sequences u U m so that
u A ˜ i ( m ) P U ( L n ) ( u , m ) P V n ( x i ) 1 γ 0 < u A ˜ i ( m ) P U ( L n ) ( u , m ) + Pr { L n = m } K m
and
A ˜ i ( m ) A ˜ j ( m ) = ( i j ) .
If
i = 1 | S n ( m ) | u A ˜ i ( m ) P U ( L n ) ( u , m ) < 1 1 γ 0 i = 1 | S n ( m ) | P V n ( x i ) = Pr { L n = m } ,
then add a u i U m ( j A ˜ j ( m ) ) to obtain
A i ( m ) = A ˜ i ( m ) { u i }
for i = 1 , 2 , in order, until it holds that with some 1 c | S n ( m ) |
i = 1 c A i ( m ) i = c + 1 | S n ( m ) | A ˜ i ( m ) = U m ,
where u 1 , u 2 , are selected to be all distinct. Since | U m | = K m and
u U m P U ( L n ) ( u , m ) = u U m Pr { V n S n ( m ) } ( 1 γ 0 ) K m = Pr { L n = m } ,
such a 1 c | S n ( m ) | always exists. For simplicity, we set for i = c + 1 , c + 2 , , | S n ( m ) |
A i ( m ) = A ˜ i ( m )
and for i = 1 , 2 , , | S n ( m ) |
φ n ( u ) = x i f o r u A i ( m ) ,
which defines the random variable X ˜ n with values in X n such that
P X ˜ n ( x i ) = u A i ( m ) P U ( L n ) ( u , m ) ( x i S n ( m ) ) ,
that is, X ˜ n = φ n ( U ( L n ) ) . Notice that, by this construction, we have
P X ˜ n ( x i ) P V n ( x i ) 1 γ 0 Pr { L n = m } K m = Pr { V n S n ( m ) } ( 1 γ 0 ) K m
for i = 1 , 2 , , | S n ( m ) | ; m = 1 , 2 , , β n , and
Pr { X ˜ n T n } = 0 and Pr { V n T n } γ .
(b)
Evaluation of Average Length:
The average length E [ L n ] is evaluated as follows:
E [ L n ] = m = 1 β n u U m P U ( L n ) ( u , m ) · m = m = 1 β n i = 1 | S n ( m ) | u A i ( m ) P U ( L n ) ( u , m ) · m = m = 1 β n x i S n ( m ) P X ˜ n ( x i ) · m ,
where we have used U m = i = 1 | S n ( m ) | A i ( m ) and (128). For x i S n ( m ) we obtain from (129) and the right inequality of (130)
P X ˜ n ( x i ) P V n ( x i ) 1 γ 0 + Pr { V n S n ( m ) } ( 1 γ 0 ) K m = 1 1 γ 0 P V n ( x i ) + Pr { V n S n ( m ) } K m 1 1 γ 0 1 + 1 P V n ( x i ) K m P V n ( x i ) 1 + 2 γ 1 + 1 K n γ P V n ( x i ) ,
where, to derive the last inequality, we have used the fact 0 γ 0 γ 1 2 and
P V n ( x i ) K ( m n γ ) ( x i S n ( m ) ) .
It should be noticed that (132) also implies that
P X ˜ n ( x ) 1 + 2 γ 1 + 1 K n γ P V n ( x ) ( x X n )
since P X ˜ n ( x ) = 0 for x T n = m = 1 β n S n ( m ) . Plugging the inequality
m log 1 P V n ( x i ) + n γ + 1 ( x i S n ( m ) )
and (132) into (131), we obtain
E [ L n ] ( 1 + 2 γ ) 1 + 1 K n γ · m = 1 β n x i S n ( m ) P V n ( x i ) log 1 P V n ( x i ) + n γ + 1 ( 1 + 2 γ ) 1 + 1 K n γ H ( V n ) + n γ + 1 .
Thus, we obtain from (136)
lim sup n 1 n E [ L n ] ( 1 + 2 γ ) lim sup n 1 n H ( V n ) + γ ( 1 + 2 γ ) lim sup n 1 n H [ δ + γ ] D ( X n ) + 2 γ ( 1 + 2 γ ) ( H + + 2 γ ) ,
where the second inequality follows from (110). Since we have assumed that H + is finite and γ ( 0 , 1 2 ] is arbitrary, the r.h.s. of (137) can be made as close to H + as desired. Therefore, for all sufficiently small γ > 0 , we obtain
lim sup n 1 n E [ L n ] H + + μ = R
(c)
Evaluation of Divergence:
The divergence D ( X ˜ n | | X n ) can be rewritten as
D ( X ˜ n | | X n ) = D ( X ˜ n | | V n ) + E log P V n ( X ˜ n ) P X n ( X ˜ n ) .
In view of (132), we obtain
D ( X ˜ n | | V n ) = m = 1 β n x S n ( m ) P X ˜ n ( x ) log P X ˜ n ( x ) P V n ( x ) m = 1 β n x S n ( m ) P X ˜ n ( x ) log 1 + 2 γ 1 + 1 K n γ 2 γ ln K + log 1 + 1 K n γ
and
E log P V n ( X ˜ n ) P X n ( X ˜ n ) = x X n P X ˜ n ( x ) log P V n ( x ) P X n ( x ) 1 + 2 γ 1 + 1 K n γ D ( V n | | X n ) 1 + 2 γ ( δ + γ ) 1 + 1 K n γ ,
where to obtain the last inequality we used the fact that P V n B δ + γ D ( X n ) . Plugging (140) and (141) into (139) yields
lim sup n D ( X ˜ n | | X n ) 2 γ ln K + ( 1 + 2 γ ) ( δ + γ ) δ + γ ( 2 δ + 5 ) ,
where we have used the fact that 2 γ ln K 3 γ for all K 2 and the assumption 0 < γ 1 2 to derive the last inequality. Since γ ( 0 , 1 2 ] is arbitrary and we have (138), R is v D ( δ ) -achievable.

5. Mean and VL Channel Resolvability

So far, we have studied the problem of source resolvability, whereas the problem of channel resolvability has been introduced by Han and Verdú [6] to investigate the capacity of identification codes [11]. In a conventional problem of this kind, a target output distribution  P Y n via a channel W n due to an input X n is approximated by encoding the FL uniform random number U M n as a channel input. In this section, we generalize the problem of such channel resolvability to that in the variable-length setting.

5.1. Definitions

Let X and Y be finite or countably infinite alphabets. Let W = { W n } n = 1 be a general channel, where W n : X n Y n denotes a stochastic mapping. We denote by Y = { Y n } n = 1 the output process via W due to an input process X = { X n } n = 1 , where X n and Y n take values in X n and Y n , respectively. Again, we do not impose any assumptions such as stationarity or ergodicity on either X or W . As in the previous sections, we will identify X n and Y n with their probability distributions P X n and P Y n , respectively, and these symbols are used interchangeably.
In this section, we consider several types of problems of approximating a target output distribution P Y n . The first one is the problem of mean-resolvability [6], in which the channel input is allowed to be an arbitrary general source.
Definition 5
(mean δ -channel resolvability: variational distance). Let δ [ 0 , 1 ) be fixed arbitrarily. A resolution rate R 0 is said to be mean δ-achievable for X (under the variational distance) if there exists a general source X ˜ = { X ˜ n } n = 1 satisfying
lim sup n 1 n H ( X ˜ n ) R ,
lim sup n d ( P Y n , P Y ˜ n ) δ ,
where Y ˜ n denotes the output via W n due to input X ˜ n . The infimum of all mean δ-achievable rates for X , i.e.,
S ¯ v ( δ | X , W ) : = inf { R : R i s   m e a n   δ a c h i e v a b l e   f o r X }
is referred to as the mean δ-resolvability for X . We also define the mean δ-resolvability for the worst input as
S ¯ v ( δ | W ) : = sup X S ¯ v ( δ | X , W ) .
On the other hand, we may also consider the problem of VL channel resolvability. Here, the VL uniform random number U ( L n ) is defined as in the foregoing sections. Consider the problem of approximating the target output distribution P Y n via W n due to X n by using another input X ˜ n = φ n ( U ( L n ) ) with a deterministic mapping φ n : U * X n .
Definition 6
(VL δ -channel resolvability: variational distance). Let δ [ 0 , 1 ) be fixed arbitrarily. A resolution rate R 0 is said to be VL δ-achievable or simply v ( δ ) -achievable for X (under the variational distance) if there exists a VL uniform random number U ( L n ) and a deterministic mapping φ n : U * X n satisfying
lim sup n 1 n E [ L n ] R ,
lim sup n d ( P Y n , P Y ˜ n ) δ ,
where E [ · ] denotes the expected value and Y ˜ n denotes the output via W n due to input X ˜ n = φ n ( U ( L n ) ) . The infimum of all v ( δ ) -achievable rates for X , i.e.,
S v ( δ | X , W ) : = inf { R : R i s   v ( δ ) - a c h i e v a b l e   f o r X }
is called the VL δ-channel resolvability or simply v ( δ ) -channel resolvability for X . We also define the VL δ-channel resolvability or simply v ( δ ) -channel resolvability for the worst input as
S v ( δ | W ) : = sup X S v ( δ | X , W ) .
When W n is the identity mapping, the problem of channel resolvability reduces to that of source resolvability, which has been investigated in the foregoing sections. In this sense, the problem of channel resolvability is a generalization of the problem of source resolvability.
Similarly to the problem of source resolvability, we may also use the divergence between the target output distribution P Y n and the approximated output distribution P Y ˜ n as an approximation measure.
Definition 7
(mean δ -channel resolvability: divergence). Let δ 0 be fixed arbitrarily. A resolution rate R 0 is said to be mean δ-achievable for X (under the divergence) if there exists a general source X ˜ = { X ˜ n } n = 1 satisfying
lim sup n 1 n H ( X ˜ n ) R ,
lim sup n D ( Y ˜ n | | Y n ) δ ,
where Y ˜ n denotes the output via W n due to input X ˜ n . The infimum of all mean δ-achievable rates for X , i.e.,
S ¯ v D ( δ | X , W ) : = inf { R : R i s   m e a n   δ a c h i e v a b l e   f o r X }
is referred to as the mean δ-channel resolvability for X . We also define the mean δ-channel resolvability for the worst input as
S ¯ v D ( δ | W ) : = sup X S ¯ v D ( δ | X , W ) .
Definition 8
(VL δ -channel resolvability: divergence). Let δ 0 be fixed arbitrarily. A resolution rate R 0 is said to be VL δ-achievable or simply v D ( δ ) -achievable for X (under the divergence) if there exists a VL uniform random number U ( L n ) and a deterministic mapping φ n : U * X n satisfying
lim sup n 1 n E [ L n ] R ,
lim sup n D ( Y ˜ n | | Y n ) δ ,
where E [ · ] denotes the expected value and Y ˜ n denotes the output via W n due to input X ˜ n = φ n ( U ( L n ) ) . The infimum of all v D ( δ ) -achievable rates for X , i.e.,
S v D ( δ | X , W ) : = inf { R : R i s   v D ( δ ) - a c h i e v a b l e   f o r X }
is called the VL δ-channel resolvability or simply v D ( δ ) -channel resolvability for X . We also define the VL δ-channel resolvability or simply v D ( δ ) -channel resolvability for the worst input as
S v D ( δ | W ) : = sup X S v D ( δ | X , W ) .
Remark 12.
Since the outputs of a deterministic mapping X ˜ n = φ n ( U ( L n ) ) form a general source X ˜ , it holds that
S ¯ v ( δ | X , W ) S v ( δ | X , W ) ( δ [ 0 , 1 ) ) ,
S ¯ v D ( δ | X , W ) S v D ( δ | X , W ) ( δ 0 )
for any general source X and general channel W . These relations lead to the analogous relation for the mean/VL δ-channel resolvability for the worst input:
S ¯ v ( δ | W ) S v ( δ | W ) ( δ [ 0 , 1 ) ) ,
S ¯ v D ( δ | W ) S v D ( δ | W ) ( δ 0 ) .

5.2. General Formulas

For a given general source X = { X n } n = 1 and a general channel W = { W n } n = 1 , let Y = { Y n } n = 1 be the channel output via W due to input X . We define
H [ δ ] , W n ( X n ) = inf P V n B δ ( X n , W n ) H ( V n ) ,
H [ δ ] , W n D ( X n ) = inf P V n B δ D ( X n , W n ) H ( V n )
where H ( V n ) denotes the Shannon entropy of V n and B δ ( X n , W n ) and B δ D ( X n , W n ) are defined as
B δ ( X n , W n ) = P V n P ( X n ) : d ( P Y n , P Z n ) δ ,
B δ D ( X n , W n ) = P V n P ( X n ) : D ( Z n | | Y n ) δ ,
respectively, with Z n defined as the output via W n due to input V n . Both H [ δ ] , W n ( X n ) and H [ δ ] , W n D ( X n ) are nonincreasing functions of δ . In addition, we define
H [ δ ] , W ( X ) = lim sup n 1 n H [ δ ] , W n ( X n ) ,
H [ δ ] , W D ( X ) = lim sup n 1 n H [ δ ] , W n D ( X n ) ,
which play an important role in characterizing the mean/VL δ -channel resolvability.
We show the general formulas for the mean/VL δ -channel resolvability.
Theorem 5
(with variational distance). For any input process X and any general channel W ,
S ¯ v ( δ | X , W ) = S v ( δ | X , W ) = lim γ 0 H [ δ + γ ] , W ( X ) ( δ [ 0 , 1 ) ) .
In particular,
S ¯ v ( δ | W ) = S v ( δ | W ) = sup X lim γ 0 H [ δ + γ ] , W ( X ) ( δ [ 0 , 1 ) ) .
Theorem 6
(with divergence). For any input process X and any general channel W ,
S ¯ v D ( δ | X , W ) = S v D ( δ | X , W ) = lim γ 0 H [ δ + γ ] , W D ( X ) ( δ 0 ) .
In particular,
S ¯ v D ( δ | W ) = S v D ( δ | W ) = sup X lim γ 0 H [ δ + γ ] , W D ( X ) ( δ 0 ) .
Remark 13.
It can be easily verified that the variational distance satisfies
d ( P Y n , P Z n ) d ( P X n , P V n ) ,
and, therefore, we have B δ ( X n ) B δ ( X n , W n ) . This relation and formulas (32) and (169) indicate that
S v ( δ | X , W ) S v ( δ | X ) ( δ [ 0 , 1 ) )
for any given channel W . Likewise, it is well known that the divergence satisfies the data processing inequality D ( Y ˜ n | | Y n ) D ( X ˜ n | | X n ) [33], and formulas (101) and (171) lead to
S v D ( δ | X , W ) S v D ( δ | X ) ( δ 0 ) ,
regardless of channel W .
Remark 14.
It is obvious that Theorems 5 and 6 reduce to Theorems 3 and 4, respectively, when the channel W is the identity mapping. Precisely, for the identity mapping W = I , the mean δ-resolvability and the v ( δ ) -channel resolvability for X are given by
S ¯ v ( δ | X ) = S v ( δ | X ) = lim γ 0 H [ δ + γ ] ( X ) ,
where S ¯ v ( δ | X ) denotes the mean δ-resolvability S ¯ v ( δ | X , W ) for the identity mapping W . The analogous relationship holds under the divergence:
S ¯ v D ( δ | X ) = S v D ( δ | X ) = lim γ 0 H [ δ + γ ] D ( X ) ,
where S ¯ v D ( δ | X ) denotes the mean δ-resolvability S ¯ v D ( δ | X , W ) for the identity mapping W = I . Thus, it turns out that Theorems 5 and 6 are indeed generalizations of Theorems 3 and 4.
Proof of Theorems 5 and 6.
(1)
Converse Part:
Because of the general relationship (159), to prove the converse part of Theorem 5, it suffices to show that
S ¯ v ( δ | X , W ) lim γ 0 H [ δ + γ ] , W ( X ) .
Let R be mean δ -achievable for X under the variational distance. Then, there exists a general source X ˜ = { X ˜ n } n = 1 satisfying (143) and
lim sup n δ n δ ,
where δ n : = d ( P Y n , P Y ˜ n ) . Fixing γ > 0 arbitrarily, we have δ n δ + γ for all n n 0 with some n 0 > 0 and then
H [ δ + γ ] , W n ( X n ) H [ δ n ] , W n ( X n ) ( n n 0 )
since H [ δ ] , W n ( X n ) is a nonincreasing function of δ . Since P X ˜ n B δ n ( X n , W n ) , we have H [ δ n ] , W n ( X n ) H ( X ˜ n ) . Thus, we obtain from (143)
H [ δ + γ ] , W ( X ) lim sup n 1 n H ( X ˜ n ) R .
Since γ > 0 is an arbitrary constant, this implies (178).
The converse part of Theorem 6 can be proven in an analogous way with due modifications.
(2)
Direct Part:
Because of the general relationship (159), to prove the direct part (achievability) of Theorem 5, it suffices to show that, for any fixed γ > 0 , the resolution rate
R = lim γ 0 H [ δ + γ ] , W ( X ) + 3 γ
is v ( δ ) -achievable for X under the variational distance.
Let P V n B δ + γ ( X n , W n ) be a source satisfying
H ( V n ) H [ δ + γ ] , W n ( X n ) + γ .
Then, by the same argument to derive (80) and (82) as developed in the proof of the direct part of Theorem 3, we can construct a VL uniform random number U ( L n ) and a deterministic mapping φ n : U * X n satisfying
lim sup n 1 n E [ L n ] lim γ 0 H [ δ + γ ] , W ( X ) + 3 γ = R
and
d ( P X ˜ n , P V n ) 1 2 K n γ + γ ,
where X ˜ n = φ n ( U ( L n ) ) . Let Z n denote the output random variable via W n due to input V n . Then, letting Y ˜ n be the output via channel W n due to input X ˜ n , we can evaluate d ( P Y ˜ n , P Z n ) as
d ( P Y ˜ n , P Z n ) = 1 2 y Y n | P Y ˜ n ( y ) P Z n ( y ) | = 1 2 y Y n x X n W ( y | x ) P X ˜ n ( x ) P V n ( x ) 1 2 y Y n x X n W ( y | x ) P X ˜ n ( x ) P V n ( x ) = d ( P X ˜ n , P V n ) 1 2 K n γ + γ .
Thus, we obtain
lim sup n d ( P Y n , P Y ˜ n ) lim sup n d ( P Y n , P Z n ) + lim sup n d ( P Y ˜ n , P Z n ) δ + 2 γ ,
where we have used the fact P V n B δ + γ ( X n , W n ) to derive the last inequality. Since γ > 0 is an arbitrary constant, we can conclude that R is v ( δ ) -achievable for X .
The direct part of Theorem 6 can be proven in the same way as Theorem 4 with due modifications. Fixing P V n B δ + γ D ( X n , W n ) and using the encoding scheme as developed in the proof of Theorem 4, the evaluation of the average length rate is exactly the same, and we can obtain (138). A key step is to evaluate the divergence D ( Y ˜ n | | Y n ) , which can be rewritten as
D ( Y ˜ n | | Y n ) = D ( Y ˜ n | | Z n ) + E log P Z n ( Y ˜ n ) P Y n ( Y ˜ n ) .
The first term on the r.h.s. can be bounded as
D ( Y ˜ n | | Z n ) D ( X ˜ n | | V n ) 2 γ ln K + log 1 + 1 K n γ
as in (140), where the left inequality is due to the data processing inequality. Similarly to the derivation of (141), the second term can be bounded as
E log P Z n ( Y ˜ n ) P Y n ( Y ˜ n ) = y Y n x X n P X ˜ n ( x ) W n ( y | x ) log P Z n ( y ) P Y n ( y ) 1 + 2 γ 1 + 1 K n γ y Y n x X n P V n ( x ) W n ( y | x ) log P Z n ( y ) P Y n ( y ) = 1 + 2 γ 1 + 1 K n γ D ( Z n | | Y n ) ,
where we have used (134). Here, D ( Z n | | Y n ) δ + γ because Z n is the output via W n due to input V n B δ + γ D ( X n , W n ) . The rest of the steps are the same as in the proof of Theorem 4.

6. Second-Order VL Channel Resolvability

So far, we have analyzed the first-order VL resolvabilities and established various first-order resolvability theorems. One of the next important steps is the second-order analysis, and so, in this section, we generalize VL resolvabilities in the second-order setting.

6.1. Definitions

We now turn to considering the second-order resolution rates [26,27,29]. First, we consider the VL resolvability based on the variational distance.
Definition 9
(VL ( δ , R ) -channel resolvability: variational distance). A second-order resolution rate L ( , + ) is said to be VL ( δ , R ) -achievable (under the variational distance) for X with δ [ 0 , 1 ) if there exist a VL uniform random number U ( L n ) and a deterministic mapping φ n : U * X n satisfying
lim sup n 1 n E [ L n ] n R L ,
lim sup n d ( P Y n , P Y ˜ n ) δ ,
where Y ˜ n denotes the output via W n due to input X ˜ n = φ n ( U ( L n ) ) . The infimum of all VL ( δ , R ) -achievable rates for X is denoted by
T v ( δ , R | X , W ) : = inf { L : L i s V L ( δ , R ) - a c h i e v a b l e f o r X } .
When W is the identity mapping I , T v ( δ , R | X , W ) is simply denoted by T v ( δ , R | X ) (source resolvability).
Next, we may consider the VL resolvability with the divergence instead of the variational distance.
Definition 10
(VL ( δ , R ) -channel resolvability: divergence). A second-order resolution rate L ( , + ) is said to be VL ( δ , R ) -achievable for X (with the divergence) where δ 0 if there exists a VL uniform random number U ( L n ) and a deterministic mapping φ n : U * X n satisfying
lim sup n 1 n E [ L n ] n R L ,
lim sup n D ( Y ˜ n | | Y n ) δ ,
where Y ˜ n denotes the output via W n due to input X ˜ n = φ n ( U ( L n ) ) . The infimum of all VL ( δ , R ) -achievable rates for X is denoted as
T v D ( δ , R | X , W ) : = inf { L : L i s   V L ( δ , R ) a c h i e v a b l e f o r X } .
When W is the identity mapping I , T v D ( δ , R | X , W ) is simply denoted by T v D ( δ , R | X ) (source resolvability).
Remark 15.
It is easily verified that
T v ( δ , R | X , W ) = + f o r R < S v ( δ | X , W ) f o r R > S v ( δ | X , W ) .
Hence, only the case R = S v ( δ | X , W ) is of interest to us. The same remark also applies to T v D ( δ , R | X , W ) .

6.2. General Formulas

We establish general formulas for the second-order resolvability. The proofs of the following theorems are given below subsequently to Remark 17.
Theorem 7
(with variational distance). For any input process X and general channel W ,
T v ( δ , R | X , W ) = lim γ 0 lim sup n 1 n H [ δ + γ ] , W n ( X n ) n R ( δ [ 0 , 1 ) , R 0 ) .
In particular, in the case where W is the identity mapping I ,
T v ( δ , R | X ) = lim γ 0 lim sup n 1 n H [ δ + γ ] ( X n ) n R ( δ [ 0 , 1 ) , R 0 ) .
Theorem 8
(with divergence). For any input process X and general channel W ,
T v D ( δ , R | X , W ) = lim γ 0 lim sup n 1 n H [ δ + γ ] , W n D ( X n ) n R ( δ 0 , R 0 ) .
In particular, in the case where W is the identity mapping I ,
T v D ( δ , R | X ) = lim γ 0 lim sup n 1 n H [ δ + γ ] D ( X n ) n R ( δ 0 , R 0 ) .
Remark 16.
As discussed in Section 5, we may also consider using a general source X ˜ as an input to channel W , and we can define L to be a mean ( δ , R ) -achievable rate for X by replacing (191) and (194) with
lim sup n 1 n H ( X ˜ n ) n R L .
Let T ¯ v ( δ , R | X , W ) and T ¯ v D ( δ , R | X , W ) denote the infimum of all mean ( δ , R ) -achievable rates for X under the variational distance and the divergence, respectively. Then, it is not difficult to verify that
T ¯ v ( δ , R | X , W ) = T v ( δ , R | X , W ) ( δ [ 0 , 1 ) ) ,
T ¯ v D ( δ , R | X , W ) = T v D ( δ , R | X , W ) ( δ 0 ) .
Thus, there is no loss in the ( δ , R ) -achievable resolution rate even if the channel input X ˜ is restricted to be generated by the VL uniform random number U ( L n ) .
Remark 17.
As in the first-order case, when the channel W is the identity mapping I , T v ( δ , R | X ) coincides with the minimum second-order length rate of the VL source codes. More precisely, we denote by R v * ( δ , R | X ) the minimum second-order length rate of a sequence of VL source codes with the first-order average length rate R and an average error probability asymptotically not exceeding δ. Yagi and Nomura [31] have shown that
R v * ( δ , R | X ) = lim γ 0 lim sup n 1 n G [ δ + γ ] ( X n ) n R ( δ [ 0 , 1 ) , R 0 ) .
Modifying the proof of Proposition 1 (cf. Appendix C), we can show that the r.h.s. of (199) coincides with that of (205), and, therefore, it generally holds that
T v ( δ , R | X ) = R v * ( δ , R | X ) ( δ [ 0 , 1 ) , R 0 ) .
As a special case, suppose that X is a stationary and memoryless source X with the finite third absolute moment of log 1 P X ( X ) . In this case, Kostina et al. [25] have recently given a single-letter characterization for R v * ( δ , R | X ) with R = H [ δ ] ( X ) = ( 1 δ ) H ( X ) as
R v * ( δ , R | X ) = V ( X ) 2 π e ( Q 1 ( δ ) ) 2 2 ,
where V ( X ) denotes the variance of log 1 P X ( X ) (varentropy) and Q 1 is the inverse of the complementary cumulative distribution function of the standard Gaussian distribution. In view of the general relation (206), we can also obtain the single-letter characterization for T v ( δ , R | X ) :
T v ( δ , R | X ) = V ( X ) 2 π e ( Q 1 ( δ ) ) 2 2 .
It has not yet been made clear whether we can also have a single-letter formula for T v ( δ , R | X , W ) when the channel W is memoryless but not necessarily the identity mapping.
Proof of Theorems 7 and 8.
(1)
Converse Part:
We will show the converse part of Theorem 7. The converse part of Theorem 8 can be proved in an analogous way.
Let L be VL ( δ , R ) -achievable for X under the variational distance. Then, there exists U ( L n ) and φ n satisfying (191) and
lim sup n δ n δ ,
where we define δ n = d ( P Y n , P Y ˜ n ) , and Y ˜ n is the output via W n due to input X ˜ n = φ n ( U ( L n ) ) . Equation (209) implies that, for any given γ > 0 , it holds that δ n δ + γ for all n n 0 with some n 0 > 0 , and, therefore,
H [ δ + γ ] , W n ( X n ) H [ δ n ] , W n ( X n ) ( n n 0 ) .
Since P X ˜ n B δ n ( X n , W n ) , we have
H [ δ n ] , W n ( X n ) H ( X ˜ n ) .
On the other hand, it follows from (23) that
H ( X ˜ n ) H ( U ( L n ) ) = E [ L n ] + H ( L n ) ,
where the inequality is due to the fact that φ n is a deterministic mapping and X ˜ n = φ n ( U ( L n ) ) .
Combining (210)–(212) yields
lim sup n 1 n H [ δ + γ ] , W n ( X n ) n R lim sup n 1 n H ( X ˜ n ) n R lim sup n 1 n E [ L n ] n R + lim sup n 1 n H ( L n ) L ,
where we have used (24) and (191) for the last inequality. Since γ > 0 is arbitrary, we have
lim γ 0 lim sup n 1 n H [ δ + γ ] , W n ( X n ) n R L .
(2)
Direct Part:
We will show the direct part (achievability) of Theorem 7 by modifying the argument of Theorems 3 and 5, whereas the direct part of Theorem 8 can be proved in a similar manner by modifying that of Theorem 4 instead of Theorem 3.
Letting
L = lim γ 0 lim sup n 1 n H [ δ + γ ] , W n ( X n ) n R + 2 γ ,
where γ > 0 is an arbitrary constant, we shall show that L is VL ( δ , R ) -achievable for X under the variational distance.
We use the same achievability scheme as in the proof of Theorem 3 with slightly different parameter settings. For γ > 0 , we choose a c n > 0 so that
Pr { V n T n } γ
where P V n B δ + γ ( X n , W n ) with H [ δ + γ ] , W n ( X n ) + γ H ( V n ) and
T n : = x X n : 1 n log 1 P V n ( x ) c n .
We here define
( x ) : = log 1 P V n ( x ) + n γ f o r x T n 0 otherwise
and β n = ( n c n + n γ ) . Arguing similarly to the proof of Theorems 3 and 5, we can show that there exist φ n : U * X n and U ( L n ) such that
d ( P Y n , P Y ˜ n ) δ + 2 γ + 1 2 K n γ
and
E [ L n ] 1 + 1 K n γ H [ δ + γ ] , W n ( X n ) + 2 n γ + 1 .
Therefore, we obtain
lim n d ( P Y n , P Y ˜ n ) δ + 2 γ
and
lim sup n 1 n E [ L n ] n R lim sup n 1 n H [ δ + γ ] , W n ( X n ) n R + 2 γ lim γ 0 lim sup n 1 n H [ δ + γ ] , W n ( X n ) n R + 2 γ = L .
Since γ > 0 is arbitrary, L is VL ( δ , R ) -achievable for X .

7. Conclusions

We have investigated the problem of VL source/channel resolvability, in which a given target probability distribution is approximated by transforming VL uniform random numbers. Table 1 summarizes various first-order resolvabilities and their characterizations in terms of information quantities. In this table, the theorem numbers that contain the corresponding characterization are also indicated.
In this paper, we have first analyzed the fundamental limits on the VL δ -source resolvability with the variational distance in Theorem 3. The VL δ -source resolvability is essentially characterized in terms of smooth Shannon entropies. In the proof of the direct part, we have developed a simple method for information spectrum slicing, in which sliced information densities quantized to the same integer are approximated by an FL uniform random number of the same length. Next, we have extended the analysis to the δ -source resolvability under the unnormalized divergence in Theorem 4. The smoothed entropy with the divergence again plays an important role in characterizing the δ -source resolvability.
Then, we have addressed the problem of δ -channel resolvability. It has been revealed in Theorems 5 and 6 that using an arbitrary general source as a coin distribution (mean-resolvability problem) cannot go beyond the fundamental limits of the VL resolvability, in which only VL uniform random numbers are allowed to be a coin distribution. As in the case of source resolvability, we have discussed the δ -channel resolvability under the variational distance and the unnormalized divergence. The second-order channel resolvability has been characterized in Theorems 7 and 8 as well as the first-order case. We notice here that a counterpart of the VL uniform random number is the problem of VL source coding, for which the general treatment, focused on overflow/underflow probabilities, is found in [38]. Indeed, when the variational distance is used as an approximation measure, it turns out that the δ -source resolvability is equal to the minimum achievable rate of VL source codes with an error probability of less than or equal to δ . This is a parallel relationship between FL source resolvability and the minimum achievable rate of FL source codes [6,7]. It is of interest to investigate whether there is a coding problem to which the δ -channel resolvability is closely related.
When δ = 0 , asymptotically exact approximation is required. In the case where the channel W is the identity mapping I , it turned out that the source resolvability under the variational distance and the unnormalized divergence coincides and is given by lim γ 0 H [ γ ] ( X ) , where X is the general target source. This result is analogous to the dual problem of VL intrinsic randomness [5,36], in which the maximum achievable rates of VL uniform random numbers extracted from a given source X are the same under two kinds of approximation measures. It should be emphasized that in the case of VL intrinsic randomness, the use of normalized divergence as an approximation measure results in the same general formula as with the variational distance and the unnormalized divergence, which does not necessarily hold in the case of mean/VL resolvability (cf. Remark 11). It is also noteworthy that whereas only the case of δ = 0 has been completely solved for the VL intrinsic randomness, we have also dealt with the case of a δ > 0 for the VL source/channel resolvability.
When X is a stationary and memoryless source or is even with a one-point spectrum (cf. Corollary 1), the formulas established reduce to a single-letter characterization for the first- and second-order source resolvability under the variational distance. In the case where the divergence is an approximation measure and/or the channel W is a non-identity mapping, however, it has not yet been made clear whether we can derive a single-letter characterization for the δ -source/channel resolvability. This question remains to be studied.
As noted in Remark 10, the order of arguments in divergence D ( X ˜ n | | X n ) is important, and it seems difficult to extend the analyses in this paper to the case with the reversed D ( X n | | X ˜ n ) used as an approximation measure. In the context of intrinsic randomness, the reversed divergence is also discussed [23]. Investigating the problem of source/channel resolvability using such divergence is an interesting research topic.

Author Contributions

The author H.Y. contributed to the conceptualization of the research goals and aims, the visualization/presentation of the works, the formal analysis of the results, and the writing of the paper. The author T.S.H. contributed to the formal analysis of the results, the validation of obtained results, and the review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by JSPS KAKENHI Grant Numbers JP16K06340, JP18H01438, and JP20K04462.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Proof of Equation (31)

(i)
We first show lim α 1 H [ δ ] α ( X n ) H [ δ ] ( X n ) .
Fix γ > 0 arbitrarily. We choose P V n B δ ( X n ) satisfying
H [ δ ] ( X n ) + γ H ( V n ) .
It is well known that the Rényi entropy of order α defined as
H α ( V n ) = 1 1 α log x X n P V n ( x ) α ( α ( 0 , 1 ) ( 1 , + ) )
satisfies
H ( V n ) = lim α 1 H α ( V n ) .
By definition, we have
H α ( V n ) H [ δ ] α ( X n ) ( α ( 0 , 1 ) ( 1 , + ) ) ,
leading to
H ( V n ) = lim α 1 H α ( V n ) lim α 1 H [ δ ] α ( X n ) .
Combining (A1) and (A4) yields
H [ δ ] ( X n ) + γ H ( V n ) lim α 1 H [ δ ] α ( X n ) .
Since γ > 0 is an arbitrary constant, we obtain lim α 1 H [ δ ] α ( X n ) H [ δ ] ( X n ) .
(ii)
Next, we shall show lim α 1 H [ δ ] α ( X n ) H [ δ ] ( X n ) .
Fix γ > 0 arbitrarily. We choose some α 0 ( 0 , 1 ) satisfying
lim α 1 H [ δ ] α ( X n ) + γ H [ δ ] α 0 ( X n ) .
For this α 0 we choose P V n B δ ( X n ) satisfying
H [ δ ] α 0 ( X n ) + γ H α 0 ( V n ) .
Since H α ( V n ) is a nonincreasing function of α , we have
H α 0 ( V n ) H ( V n ) ,
and it follows from (A6)–(A8) that
lim α 1 H [ δ ] α ( X n ) + 2 γ H α 0 ( V n ) H ( V n ) .
Since H ( V n ) H [ δ ] ( X n ) , due to P V n B δ ( X n ) , and γ > 0 is arbitrarily fixed, we obtain the desired inequality.

Appendix B. Proof of Equation (34)

To prove the alternative formula (34) for the v ( δ ) -resolvability S v ( δ | X ) , we shall show
lim γ 0 H [ δ + γ ] ( X ) = inf V B δ ( X ) H ( V ) ( δ [ 0 , 1 ) ) .
(i)
We first show lim γ 0 H [ δ + γ ] ( X ) inf V B δ ( X ) H ( V ) .
Fix γ > 0 arbitrarily. We choose V ˜ = { V ˜ n } n = 1 B δ ( X ) satisfying
H ( V ˜ ) inf V B δ ( X ) H ( V ) + γ .
For V ˜ B δ ( X ) , we have d ( X n , V ˜ n ) δ + γ for all n n 0 with some n 0 > 0 , yielding
H [ δ + γ ] ( X n ) H ( V ˜ n ) ( n n 0 ) .
Thus, it follows from (A11) and (A12) that
H [ δ + γ ] ( X ) inf V B δ ( X ) H ( V ) + γ .
Since γ > 0 is an arbitrary constant, letting γ 0 on both sides yields the desired inequality.
(ii)
Next, we shall show lim γ 0 H [ δ + γ ] ( X ) inf V B δ ( X ) H ( V ) .
Fix λ > 0 arbitrarily. We choose an arbitrary decreasing sequence of positive numbers { γ i } i = 1 satisfying γ 1 > γ 2 > 0 . Then, we have
lim γ 0 H [ δ + γ ] ( X ) = lim i H [ δ + γ i ] ( X ) .
Additionally, by the definition of the limit superior, for each i = 1 , 2 , we have
1 n H [ δ + γ i ] ( X n ) H [ δ + γ i ] ( X ) + λ ( n n i )
with some 0 < n 1 < n 2 < . Now, for each n = 1 , 2 , , we denote by i n the index i satisfying
n i n < n i + 1 .
Then, from (A15), we obtain
1 n H [ δ + γ i n ] ( X n ) H [ δ + γ i n ] ( X ) + λ ( n n 1 ) .
On the other hand, by the definition of H [ δ + γ i n ] ( X n ) , for each n = 1 , 2 , , we can choose some V i n n B δ + γ i n ( X n ) satisfying
1 n H ( V i n n ) 1 n H [ δ + γ i n ] ( X n ) + λ .
We now construct the general source V ˜ = { V i n n } n = 1 . Since V i n n B δ + γ i n ( X n ) for all n n 1 indicates that
lim sup n d ( X n , V i n n ) δ + lim n γ i n = δ ,
the general source satisfies V ˜ B δ ( X ) .
From (A17) and (A18), we obtain
1 n H ( V i n n ) H [ δ + γ i n ] ( X ) + 2 λ ( n n 1 ) .
In view of (A14) and the fact V ˜ B δ ( X ) , taking lim sup n on both sides yields
inf V B δ ( X ) H ( V ) H ( V ˜ ) lim sup n H [ δ + γ i n ] ( X ) + 2 λ = lim γ 0 H [ δ + γ ] ( X ) + 2 λ .
Since λ > 0 is an arbitrary constant, letting λ 0 yields the desired inequality.

Appendix C. Proof of Proposition 1

We shall prove the equality and inequality in (40). Equation (41) is an immediate consequence of (40) because H ¯ δ ( X ) is a right-continuous function [2].
(i)
We first shall show the equality in (40): H [ δ ] ( X ) = G [ δ ] ( X ) .
We show H [ δ ] ( X ) G [ δ ] ( X ) . For any given γ > 0 and P X n , let A n * X n be a subset of X n which satisfies
Pr { X n A n * } 1 δ
and
G [ δ ] ( X n ) + γ x A n * P X n ( x ) log 1 P X n ( x ) = : F ( A n * ) .
Choose x 0 X n A n * arbitrarily. Set P X ˜ n so that
P X ˜ n ( x ) = P X n ( x ) f o r x A n * α 0 f o r x = x 0 0 otherwise ,
where we define α 0 = Pr { X n A n * } .
The variational distance between P X n and P X ˜ n satisfies
d ( P X n , P X ˜ n ) = 1 2 x A n * | P X n ( x ) P X ˜ n ( x ) | 1 2 x A n * ( P X n ( x ) + P X ˜ n ( x ) ) α 0 = 1 Pr { X n A n * } δ ,
where the last inequality is due to (A22). Therefore, P X ˜ n B δ ( X n ) , and this implies
H [ δ ] ( X n ) H ( X ˜ n ) = F ( A n * ) + α 0 log 1 α 0 G [ δ ] ( X n ) + α 0 log 1 α 0 + γ G [ δ ] ( X n ) + log e e + γ ,
where the first inequality is due to (A23) and the last inequality is due to the inequality x log x log e e for all x > 0 . Thus, we obtain the desired inequality: H [ δ ] ( X ) G [ δ ] ( X ) .
Next, we shall show H [ δ ] ( X ) G [ δ ] ( X ) . Assume, without loss of generality, that the elements of X n are indexed as x 1 , x 2 , X n so that
P X n ( x i ) P X n ( x i + 1 ) ( i = 1 , 2 , ) .
For a given δ [ 0 , 1 ) , let j * denote the integer satisfying
i = 1 j * 1 P X n ( x i ) < 1 δ , i = 1 j * P X n ( x i ) 1 δ .
Let V δ n be a random variable taking values in X n whose probability distribution is given by
P V δ n ( x i ) = P X n ( x i ) + δ for i = 1 P X n ( x i ) for i = 2 , 3 , , j * 1 P X n ( x i ) ε n for i = j * 0 otherwise ,
where ε n : = δ i j * + 1 P X n ( x i ) . It is easily checked that 0 ε n P X n ( x j * ) and the probability distribution P V δ n  majorizes (for a sequence u = ( u 1 , u 2 , , u L ) of length L, we denote by u ˜ = ( u ˜ 1 , u ˜ 2 , , u ˜ L ) a permuted version of u satisfying u ˜ i u ˜ i + 1 for all i = 1 , 2 , , L 1 , where ties are arbitrarily broken. We say u = ( u 1 , u 2 , , u L )  majorizes  v = ( v 1 , v 2 , , v L ) if i = 1 j u ˜ i i = 1 j v ˜ i for all j = 1 , 2 , , L .) any P V n B δ ( X n ) [39]. Since the Shannon entropy is a Schur concave function (The function f ( u ) is said to be Schur concave if f ( u ) f ( v ) for any pair ( u , v ) , where v is majorized by u .) According to [40], we immediately obtain the following lemma, which is of use to compute H [ δ ] ( X n ) .
Lemma A1
(Ho and Yeung [39]).
H [ δ ] ( X n ) = H ( V δ n ) ( δ [ 0 , 1 ) ) .
By the definition of G [ δ ] ( X n ) , we obtain
G [ δ ] ( X n ) i = 1 j * P X n ( x i ) log 1 P X n ( x i )
H ( V δ n ) + P X n ( x 1 ) log 1 P X n ( x 1 ) + P X n ( x j * ) log 1 P X n ( x j * )
H ( V δ n ) + 2 log e e ,
where the last inequality is due to x log x log e e for all x > 0 . Thus, it follows from Lemma A1 that
G [ δ ] ( X ) lim sup n 1 n H ( V δ n ) = H [ δ ] ( X ) ,
which is the desired inequality.
(ii)
Next, we show the inequality in (40): H [ δ ] ( X ) ( 1 δ ) H ¯ δ γ ( X ) . By the definition of H ¯ δ γ ( X ) , for any η > 0 , there exists some n 0 > 0 such that
Pr { X n T n } 1 δ ( n n 0 ) ,
where we define
T n = x X n : 1 n log 1 P X n ( x ) H ¯ δ γ ( X ) + η .
Choose a sequence x 0 T n arbitrarily. Set P V n so that
P V n ( x ) = α n P X n ( x ) + δ f o r x = x 0 α n P X n ( x ) f o r x x 0 , x T n 0 otherwise ,
where we define α n = ( 1 δ ) / Pr { X n T n } . Then, the variational distance between P X n and P V n satisfies
d ( P X n , P V n ) δ 2 + 1 2 x T n P X n ( x ) | 1 α n | + 1 2 x T n c P X n ( x ) = δ 2 + 1 2 Pr { X n T n } ( 1 α n ) + Pr { X n T n c } = δ ,
which indicates that P V n B δ ( X n ) . The normalized Shannon entropy 1 n H ( V n ) can be upper bounded as
1 n H ( V n ) = 1 n ( α n P X n ( x 0 ) + δ ) log 1 α n P X n ( x 0 ) + δ + 1 n x T n { x 0 } P V n ( x ) log 1 P V n ( x ) δ n log 1 δ + 1 n x T n α n P X n ( x ) log 1 α n P X n ( x ) = δ n log 1 δ + α n n x T n P X n ( x ) log 1 α n + log 1 P X n ( x ) δ n log 1 δ + ( 1 δ ) n log 1 1 δ + α n x T n P X n ( x ) H ¯ δ γ ( X ) + η = 1 n h 2 ( δ ) + ( 1 δ ) H ¯ δ γ ( X ) + η ,
where we have used the fact that 1 n log 1 P X n ( x ) H ¯ δ γ ( X ) + η for x T n to obtain the second inequality, and h 2 ( δ ) : = δ log δ ( 1 δ ) log ( 1 δ ) denotes the binary entropy function. On the other hand, because P V n B δ ( X n ) , there exists some n 1 > 0 such that
H [ δ ] ( X ) 1 n H ( V n ) + η ( n > n 1 ) .
Combining (A39) and (A40) yields
H [ δ ] ( X ) ( 1 δ ) H ¯ δ γ ( X ) + 3 η ,
and since η > 0 is arbitrarily fixed, (A41) means the desired inequality.

Appendix D. Proof of Corollary 1

We first show (44), i.e.,
S f ( δ | X ) = ( a ) H ¯ δ ( X ) = ( b ) H * ( X ) .
Equality (a) is due to Theorem 2. To prove equality (b), we notice that
H ̲ ( X ) H ¯ δ ( X ) H ¯ ( X ) ( δ [ 0 , 1 ) )
for general source X by definition. If source X is with one-point spectrum, the left-hand side (l.h.s.) and the r.h.s. in (A43) match, i.e., H * ( X ) : = H ̲ ( X ) = H ¯ ( X ) , and the squeeze theorem indicates equality (b).
Next, we show (45), i.e.,
S v ( δ | X ) = ( c ) H [ δ ] ( X ) = ( d ) G [ δ ] ( X ) = ( e ) ( 1 δ ) H * ( X ) .
Equality (d) is a direct consequence of Proposition 1. To prove equality (e), we notice a general relationship
( 1 δ ) H ̲ ( X ) G [ δ ] ( X ) ( 1 δ ) H ¯ ( X ) ( δ [ 0 , 1 ) )
for general source X , where the first inequality is due to ([35], Theorem 4) and the second one is due to Proposition 1 (cf. Equation (42)). In view of H * ( X ) = H ̲ ( X ) = H ¯ ( X ) for source X with one-point spectrum, the squeeze theorem for (A45) indicates equality (e), i.e.,
G [ δ ] ( X ) = ( 1 δ ) H * ( X ) ( δ [ 0 , 1 ) ) .
The r.h.s. of (A46) is obviously right-continuous in δ [ 0 , 1 ) , and so is the l.h.s. side G [ δ ] ( X ) = H [ δ ] ( X ) if source X is with one-point spectrum. The right-continuity of H [ δ ] ( X ) and Theorem 3 indicate equality (c).

Appendix E. Proof of Lemma 1

We first show (99). For two general sources X = { X n } n = 1 and X ˜ = { X ˜ n } n = 1 , the following well-known inequality (cf. ([33], Problem 3.18)) holds between the variational distance and the divergence:
2 d ( P X n , P X ˜ n ) 2 ln K D ( X ˜ n | | X n ) .
This inequality implies that any P V n B g ( δ ) D ( X n ) satisfies P V n B δ ( X n ) . Thus, we have
H [ δ ] ( X n ) H [ g ( δ ) ] D ( X n ) .
Now, we shall show the rightmost equality of (100). It obviously follows from (99) that
lim δ 0 H [ δ ] ( X ) lim δ 0 H [ δ ] D ( X ) .
To show the opposite inequality, in view of (41), it suffices to show
lim δ 0 G [ δ ] ( X ) lim δ 0 H [ δ ] D ( X ) .
Fix δ ( 0 , 1 ) and γ > 0 arbitrarily. We choose A n X n satisfying
G [ δ ] ( X n ) + γ x A n P X n ( x ) log 1 P X n ( x ) ,
α 0 : = Pr { X n A n } 1 δ .
We arrange a new random variable V n subject to
P V n ( x ) = P X n ( x ) α 0 i f x A n 0 otherwise .
Then, we obtain
D ( V n | | X n ) = x A n P V n ( x ) log P V n ( x ) P X n ( x ) = log 1 α 0 log 1 1 δ ,
and, thus, letting h ( δ ) = log 1 1 δ , it holds that P V n B h ( δ ) D ( X n ) . We can expand (A51) as
G [ δ ] ( X n ) + γ α 0 x A n P V n ( x ) log 1 α 0 P V n ( x ) α 0 H ( V n ) ( 1 δ ) H [ h ( δ ) ] D ( X n ) ,
where the last inequality is due to (A52) and P V n B h ( δ ) D ( X n ) . Thus, as γ > 0 is arbitrary,
G [ δ ] ( X ) ( 1 δ ) H [ h ( δ ) ] D ( X ) .
Since δ ( 0 , 1 ) is arbitrary, in view of lim δ 0 h ( δ ) = 0 , we obtain (A50).

References

  1. Elias, P. The efficient construction of an unbiased random sequence. Ann. Math. Statist. 1972, 43, 865–870. [Google Scholar] [CrossRef]
  2. Han, T.S. Information-Spectrum Methods in Information Theory; Springer: Berlin/Heidelberg, Germany, 2003. [Google Scholar]
  3. Knuth, D.E.; Yao, A.C.-C. The complexity of nonuniform random number generation. In Algorithms and Complexity: New Directions and Recent Results; Academic Press: New York, NY, USA, 1976; pp. 357–428. [Google Scholar]
  4. Han, T.S. Theorems on variable-length intrinsic randomness. IEEE Trans. Inf. Theory 2000, 46, 2108–2116. [Google Scholar]
  5. Vembu, S.; Verdú, S. Generating random bits from an arbitrary source: Fundamental limits. IEEE Trans. Inf. Theory 1995, 41, 1322–1332. [Google Scholar] [CrossRef]
  6. Han, T.S.; Verdú, S. Approximation theory of output statistics. IEEE Trans. Inf. Theory 1993, 39, 752–771. [Google Scholar] [CrossRef]
  7. Steinberg, Y.; Verdú, S. Simulation of random processes and rate-distortion theory. IEEE Trans. Inf. Theory 1996, 42, 63–86. [Google Scholar] [CrossRef]
  8. Renner, R.; Wolf, S. Smooth Rényi entropy and applications. In Proceedings of the International Symposium onInformation Theory, ISIT 2004, Chicago, IL, USA, 27 June–2 July 2004; p. 232. [Google Scholar]
  9. Hou, J.; Kramer, G. Informational divergence approximations to product distributions. In Proceedings of the 13th Canadian Workshop on Information Theory, Toronto, ON, Canada, 18–21 June 2013. [Google Scholar]
  10. Yagi, H. Channel resolvability theorems for general sources and channels. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017. [Google Scholar]
  11. Ahlswede, R.; Dueck, G. Identification via channels. IEEE Trans. Inf. Theory 1989, 35, 15–29. [Google Scholar] [CrossRef]
  12. Oohama, Y. Converse coding theorems for identification via channels. IEEE Trans. Inf. Theory 2013, 59, 744–759. [Google Scholar] [CrossRef]
  13. Steinberg, Y. New converses in the theory of identification via channels. IEEE Trans. Inf. Theory 1998, 44, 984–998. [Google Scholar] [CrossRef]
  14. Sreekumar, S.; Cohen, A.; Gündüz, D. Privacy-aware distributed hypothesis testing. Entropy 2020, 22, 665. [Google Scholar] [CrossRef]
  15. Zheng, F.; Xiao, Z.; Zhou, S.; Wang, J.; Huang, L. Message authentication over noisy channels. Entropy 2015, 17, 368–383. [Google Scholar] [CrossRef]
  16. Lin, P.-H.; Janda, C.R.; Jorswieck, E.A.; Schaefer, R.F. Stealthy secret key generation. Entropy 2020, 22, 679. [Google Scholar] [CrossRef]
  17. Hayashi, M. Nonasymptotic and asymptotic formulas in channel resolvability and identification capacity and their application to the wiretap channel. IEEE Trans. Inf. Theory 2006, 52, 1562–1575. [Google Scholar] [CrossRef]
  18. Bloch, M.R.; Laneman, J.N. Strong secrecy from channel resolvability. IEEE Trans. Inf. Theory 2013, 59, 8077–8098. [Google Scholar] [CrossRef]
  19. Frèche, G.; Bloch, M.R.; Barret, M. Polar codes for covert communications over asynchronous discrete memoryless channels. Entropy 2022, 22, 679. [Google Scholar]
  20. Strassen, V. Asymptotische abschätzungen in Shannon’s informationstheorie. In Proceedings of the Transactions of the Third Prague Conference on Information Theory, Statistical Decision Functions, Random Processes, Prague, Czech Republic, 5–13 June 1962; pp. 689–723. [Google Scholar]
  21. Han, T.S. Second-order information theory and hypothesis testing. In Proceedings of the 2015 IEEE Information Theory Workshop, Jeju, Republic of Korea, 11–15 October 2015. [Google Scholar]
  22. Han, T.S.; Nomura, R. First- and second-order hypothesis testing for mixed memoryless sources. Entropy 2018, 20, 174. [Google Scholar] [CrossRef]
  23. Hayashi, M. Second-order asymptotics in fixed-length source coding and intrinsic randomness. IEEE Trans. Inf. Theory 2008, 54, 4619–4637. [Google Scholar] [CrossRef]
  24. Hayashi, M. Information spectrum approach to second-order coding rate in channel coding. IEEE Trans. Inf. Theory 2009, 55, 4947–4966. [Google Scholar] [CrossRef]
  25. Kostina, V.; Polyanskiy, Y.; Verdú, S. Variable-length compression allowing errors. IEEE Trans. Inf. Theory 2015, 61, 4316–4330. [Google Scholar] [CrossRef]
  26. Nomura, R. Source resolvability and intrinsic randomness: Two random number generation problems with respect to a subclass of f-divergences. IEEE Trans. Inf. Theory 2020, 66, 7588–7601. [Google Scholar] [CrossRef]
  27. Nomura, R.; Han, T.S. Second-order resolvability, intrinsic randomness, and fixed-length source coding for mixed sources: Information spectrum approach. IEEE Trans. Inf. Theory 2013, 59, 1–16. [Google Scholar] [CrossRef]
  28. Polyanskiy, Y.; Poor, H.V.; Verdú, S. Channel coding rate in the finite blocklength regime. IEEE Trans. Inf. Theory 2010, 59, 2307–2358. [Google Scholar] [CrossRef]
  29. Watanabe, S.; Hayashi, M. Strong converse and second-order asymptotics of channel resolvability. In Proceedings of the 2014 IEEE International Symposium on Information Theory, Honolulu, HI, USA, 29 June–4 July 2014. [Google Scholar]
  30. Yagi, H.; Han, T.S.; Nomura, R. First- and second-order coding theorems for mixed memoryless channels with general mixture. IEEE Trans. Inf. Theory 2016, 62, 4395–4412. [Google Scholar] [CrossRef]
  31. Yagi, H.; Nomura, R. Variable-length coding with cost allowing non-vanishing error probability. IEICE Trans. Fundam. 2017, E100-A, 1683–1692. [Google Scholar] [CrossRef]
  32. Uyematsu, R. A new unified method for fixed-length source coding problems of general sources. IEICE Trans. Fundam. 2010, E93-A, 1868–1877. [Google Scholar] [CrossRef]
  33. Csiszár, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems, 2nd ed.; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  34. Tomita, T.; Uyematsu, T.; Matsumoto, R. Bounds on a variable-to-fixed type δ-resolvability problem. In Proceedings of the Poster Presentation at 2016 Symposium on Information Theory and its Applications, Gifu, Japan, 30 October–2 November 2016. [Google Scholar]
  35. Koga, H.; Yamamoto, H. Asymptotic properties on codeword length of an optimal FV code for general sources. IEEE Trans. Inf. Theory 2005, 51, 1546–1555. [Google Scholar] [CrossRef]
  36. Han, T.S. Weak variable-length source coding theorem. IEEE Trans. Inf. Theory 2000, 46, 1217–1226. [Google Scholar]
  37. Kuzuoka, S.; Watanabe, S. An information-spectrum approach to weak variable-length source coding with side information. IEEE Trans. Inf. Theory 2015, 61, 3559–3573. [Google Scholar] [CrossRef]
  38. Uchida, O.; Han, T.S. The optimal overflow and underflow probabilities of variable-length coding for general source. IEICE Trans. Fundam. 2001, E84-A, 2457–2465. [Google Scholar]
  39. Ho, S.-W.; Yeung, R.W. The interplay between entropy and variational distance. IEEE Trans. Inf. Theory 2010, 56, 5906–5929. [Google Scholar] [CrossRef]
  40. Marshall, A.W.; Olkin, I.; Arnold, B.C. Inequalities: Theory of Majorization and Its Applications, 2nd ed.; Springer: New York, NY, USA, 2011. [Google Scholar]
Figure 1. Illustration of the problem of FL resolvability.
Figure 1. Illustration of the problem of FL resolvability.
Entropy 25 01466 g001
Table 1. Summary of First-Order Resolvability and Information Quantities.
Table 1. Summary of First-Order Resolvability and Information Quantities.
Approximation MeasureResolvabilityCharacterizationTheorem #
Fixed-Length Resolvability
Variational Distance S f ( X ) H ¯ ( X ) Theorem 1 [6]
S f ( δ X ) H ¯ δ ( X ) Theorem 2 [7]
Variable-Length Resolvability
Variational Distance S v ( δ X ) lim γ 0 H [ δ + γ ] ( X ) Theorem 3
S v ( δ X , W ) lim γ 0 H [ δ + γ ] , W ( X ) Theorem 5
Divergence S v D ( δ X ) lim γ 0 H [ δ + γ ] D ( X ) Theorem 4
S v D ( δ X , W ) lim γ 0 H [ δ + γ ] , W D ( X ) Theorem 6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yagi, H.; Han, T.S. Variable-Length Resolvability for General Sources and Channels. Entropy 2023, 25, 1466. https://doi.org/10.3390/e25101466

AMA Style

Yagi H, Han TS. Variable-Length Resolvability for General Sources and Channels. Entropy. 2023; 25(10):1466. https://doi.org/10.3390/e25101466

Chicago/Turabian Style

Yagi, Hideki, and Te Sun Han. 2023. "Variable-Length Resolvability for General Sources and Channels" Entropy 25, no. 10: 1466. https://doi.org/10.3390/e25101466

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop