Next Article in Journal
Revocable and Traceable Decentralized ABE for P2P Networks
Previous Article in Journal
Optimization Method for Robustness of Hypernetwork Communication with Integrated Structural Features
Previous Article in Special Issue
Non-Uniform Entropy-Constrained L Quantization for Sparse and Irregular Sources
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Refinements and Generalizations of the Shannon Lower Bound via Extensions of the Kraft Inequality

The Viterbi Faculty of Electrical and Computer Engineering, Technion—Israel Institute of Technology, Technion City, Haifa 3200003, Israel
Entropy 2026, 28(1), 76; https://doi.org/10.3390/e28010076
Submission received: 12 December 2025 / Revised: 31 December 2025 / Accepted: 7 January 2026 / Published: 9 January 2026
(This article belongs to the Special Issue Information Theory and Data Compression)

Abstract

We derive a few extended versions of the Kraft inequality for lossy compression, which pave the way to the derivation of several refinements and extensions of the well-known Shannon lower bound in a variety of instances of rate-distortion coding. These refinements and extensions include sharper bounds for one-to-one codes and D-semifaithful codes, a Shannon lower bound for distortion measures based on sliding-window functions, and an individual-sequence counterpart of the Shannon lower bound.

1. Introduction

The Shannon lower bound (SLB) is one of the most important analytic tools in rate-distortion theory because it provides a simple, explicit, and often very tight lower bound to the rate-distortion function for a wide class of sources and distortion measures; see, e.g., Subsection 3.4.1 of [1], Sections 4.3 and 4.6 of [2], and Problem 10.6 of [3]. Its significance lies in giving a universal benchmark that connects rate-distortion tradeoffs to the entropy or the differential entropy of the source (depending on whether the source has a discrete or continuous alphabet), thereby offering an intuitively transparent approximation in regimes where exact evaluation of the rate-distortion function is intractable. The SLB is particularly powerful at low distortion, where it frequently coincides with the exact rate-distortion function for smooth sources, and it serves as a foundation for many refinements and asymptotic approximations (e.g., high-resolution analysis, corrections for lattice quantizers, and recent non-asymptotic bounds). In fact, the literature contains reported results on the asymptotic tightness of the SLB in the limit of low distortion under fairly mild regularity conditions [4,5]. Because computing the rate-distortion function exactly is generally difficult, the SLB plays a central role in the analysis, design, and performance assessment of lossy compression schemes. More recent studies include the finite block-length regime [6,7] and further developments concerning the quadratic distortion function [8].
In [9] Campbell derived an extension of Kraft’s inequality that leads to the SLB in both the discrete and the continuous alphabet cases. However, his results were claimed to apply to rate-distortion codes whose distortion level is defined by the distance between the two most distant source-space vectors that are mapped to the same codeword, namely, the diameter (as opposed to the radius) of the distortion ball centered at the reproduction vector. In [10], a similar Kraft inequality was derived for D-semifaithful codes, namely, codes that incur per-letter distortion that never exceeds D in the usual sense. An additional benefit of the derivation in [10] was that it could also deliver an O log n n redundancy term on top of the SLB, at least for certain distortion measures for which there is an explicit expression of the cardinality (or the volume, in the continuous case) of a ball of normalized radius D in the source vector space.
In this work, we propose a few other extended versions of Kraft’s inequality that together pave the way to several further refinements and generalizations of the SLB. In these extensions of Kraft’s inequality, the idea is to upper bound the summation (or the integral, in the continuous case) of an exponentiated negative linear combination of the code length and the distortion incurred by each and every vector in the source space. By contrasting an upper bound with a lower bound to this quantity, we obtain several refinements and extensions to the SLB, which apply to the following scenarios.
1. One-to-one codes. Instead of assuming that the reproduction vectors are represented by uniquely decodable (UD) codes, we relax this restriction and allow any one-to-one code in the level of n-vectors. The lower bound then becomes the SLB minus an O log n n term, similarly to the lossless case derived by Rissanen [11].
2. D-semifaithful codes. Similarly to [10], we consider D-semifaithful codes, but here we allow several simultaneous distortion criteria. Using saddle-point integration, we show that the resulting lower bound is given by the SLB plus k log n 2 n + o 1 n , where k is the “effective number” of distortion constraints. The concrete meaning of this term will be clarified in Section 4.2 (see the paragraph that includes Equation (40)). Generally speaking, it means the remaining number of independent constraints after removing the redundant (inactive) ones. This O log n n redundancy term stems from a more accurate evaluation of the volume of a ball of radius n D around each source word, compared to the one obtained by Chernoff bounding techniques and by the method of types in the finite-alphabet case. Interestingly, the O log n n behavior is also coherent with that of earlier derived lower bounds to the rate-distortion function for n-blocks in the finite-alphabet case; see, e.g., [12] and references therein.
3. Sliding-window distortion functions. In some applications, one might wish to shape the spectrum or the memory properties of the reconstruction error signal. This can be performed by imposing additional distortion constraints defined by additive functions that operate on two or more consecutive samples of the error signal in a sliding-window fashion. Our framework is capable of incorporating such distortion functions and allowing a derivation of the generalized SLB for this case.
4. Individual sequences and finite-state encoders. By developing a generalized Kraft inequality, similar (but not identical) to the one by Ziv and Lempel [13] for finite-state encoders, we derive also an individual-sequence counterpart of the SLB, where the source entropy term is replaced by the Lempel–Ziv complexity.
As is well known, the classical SLB can be obtained significantly more simply and easily than going via the Kraft inequality. In particular, it is obtained by a straightforward direct manipulation of the mutual information. But it should be emphasized that the point of this article is not a quest for a simpler proof of the SLB. The point is that the path that goes via the Kraft inequality leads to the above-mentioned extensions and refinements.
The outline of the remaining part of this article is as follows. In Section 2, we establish some notation conventions (Section 2.1) and provide elementary background on the SLB (Section 2.2). In Section 3, we present and prove our extended Kraft inequality in several variations. In Section 4, we derive corresponding lower bounds, first, for UD lossless compression of the reproduction data, then for one-to-one compression thereof (Section 4.1), and finally, for D-semifaithful codes (Section 4.2). In Section 5, we address the case of sliding-window distortion functions. In Section 6, we first provide some background on finite-state compression of individual sequences and the LZ algorithm (Section 6.1) and then derive an individual-sequence counterpart of the SLB (Section 6.2). Finally, in Section 6.2, we summarize and conclude this paper.

2. Notation Conventions and Background

2.1. Notation Conventions

Throughout this paper, scalar random variables (RVs) will be denoted by capital letters, their sample values will be denoted by the respective lower-case letters, and their alphabets will be denoted by the respective calligraphic letters. A similar convention will apply to random vectors and their sample values, which will be denoted with same symbols superscripted by the dimension. Thus, for example, U n (n—positive integer) will denote a random n-vector ( U 1 , , U n ) , and u n = ( u 1 , , u n ) is a specific vector value in U n , the n–th Cartesian power of U , which is the alphabet of each component of u n . In some of our derivations below, there will be a need to refer to multiple copies of the vector u n . In such cases, in order to avoid cumbersome subscripts and superscripts for indexing, we will use the alternative notation u for u n , and then the various copies will be denoted by u 1 , u 2 , etc. Returning to the first notation method, u i j and U i j , where i and j are integers and i j , will designate segments ( u i , , u j ) and ( U i , , U j ) , respectively, where for i = 1 , the subscript will be omitted (as above). For i > j , u i j (or U i j ) will be understood as the null string. Logarithms and exponents, throughout this paper, will be understood to be taken to the base 2 unless specified otherwise. The indicator function of an event A will be denoted by I { A } , i.e., I { A } = 1 if A occurs and I { A } = 0 if not.
Sources and probability distributions associated with them will be denoted generically by the letter P subscripted by the name of the RV and its conditioning, if applicable, exactly like in ordinary textbook notation standards; e.g., P U n ( u n ) is the probability function of U n at the point U n = u n , P X | U n ( x | u n ) is the conditional probability of X = x given U n = u n , and so on. Whenever clear from the context, these subscripts will be omitted. Information theoretic quantities, like entropies and mutual informations, will be denoted following the usual conventions of the information theory literature, e.g., H ( U n ) , I ( S ; U n | V n ) , and so on. The differential entropy of a continuous valued RV, U n , will be denoted by h ( U n ) . The expectation operator will be denoted by E { · } and the probability of an event A will be denoted by Pr { A } .
It should be noted that our derivations will apply to both discrete-alphabet sources and to continuous-alphabet sources. To avoid repetitions, we henceforth carry on under the assumption of a continuous-alphabet source with the understanding that in the discrete-alphabet cases, integrations over U n should simply be replaced by summations.
Let U n = ( U 1 , U 2 , , U n ) denote a source vector, drawn from a stochastic process, P, whose alphabet is U . The source vector is compressed by a lossy fixed-to-variable (F-V) source code, defined by an encoder ϕ n : U n B n { 0 , 1 } and a decoder ψ n : B n V n U n , where B n denotes a certain subset of the set of all binary variable-length strings, { 0 , 1 } * . Without loss of generality (and optimality), it is assumed that the encoder ϕ n can be viewed as a cascade of a reproduction encoder (vector quantizer) U n V n and a uniquely decodable (UD) lossless code V n B n . In a certain part of our results, this unique decodability assumption will be partially relaxed to become the less demanding assumption of a one-to-one mapping. Let L [ ϕ n ( u n ) ] denote the length (in bits) of the compressed codeword, ϕ n ( u n ) .
Assuming that U is a group with certain addition and subtraction operations (e.g., modulo-K addition/subtraction for U = { 0 , 1 , , K 1 } for a finite positive integer K or ordinary addition/subtraction for U = R ) we will focus on additive difference distortion measures, where the distortion between the source vector u n U n and the reproduction vector v n = ψ n ( ϕ n ( u n ) ) V n will be given by
d ( u n , v n ) = i = 1 n d ( u i , v i ) = i = 1 n ρ ( u i v i ) ,
where ρ ( z ) , z U is a certain non-negative function, which vanishes if and only if z = 0 , for example, ρ ( z ) = | z | , ρ ( z ) = z 2 , etc. For the sake of convenience, we will sometimes denote d ( u n , v n ) by ρ ( u n v n ) , which for a given encoder–decoder pair is also ρ ( u n ψ n ( ϕ n ( u n ) ) ) . Given an encoder–decoder pair, ( ϕ n , ψ n ) , the expected distortion, i = 1 n E { ρ ( U i V i ) } , will be constrained to be less than or equal to n D , where D > 0 designates the per-letter distortion level allowed. In certain parts of our derivations, a more restrictive pointwise distortion constraint will be imposed, i.e., ρ ( u n ψ n ( ϕ n ( u n ) ) ) n D for all u n U n . In other parts, more than one difference distortion measure will play a role, and accordingly, more than one distortion constraint will be imposed. Given k difference distortion functions, ρ j ( · ) , j = 1 , 2 , k , we then require E { ρ j ( U n ψ n ( ϕ n ( U n ) ) ) } n D j (or max u n U n ρ j ( u n ψ n ( ϕ n ( u n ) ) ) } n D j ) for all j = 1 , 2 , , k . In another part of our results, we also allow distortion functions to be additive sliding-window functions operating on m consecutive symbols of the difference z n = u n v n . I.e.,
ρ ( z n ) = i = m n ρ ( z i m + 1 i ) .
This corresponds to situations where we wish to shape not only the ‘intensity’ of the error signal, z n , but also its memory properties, for example, the correlations between consecutive symbols of z n . We will elaborate more on this in Section 5.
Similarly to u n , for which we adopt the alternative notation u, the same will apply to v n , which will also be denoted by v, along with its multiple copies v 1 , v 2 , and so on.

2.2. Background

For a continuous-alphabet memoryless source P, the SLB is given by
R ( D ) R S L B ( D ) = h ( U ) Φ ( D ) ,
where h ( U ) is the differential entropy of a single symbol U and
Φ ( D ) = sup { Z :   E { ρ ( Z ) } D } h ( Z ) = inf β 0 β D + log U 2 β ρ ( z ) d z ,
assuming that U 2 β ρ ( z ) d z < for some β > 0 . The equivalence between the two expressions of Φ ( D ) can be easily shown using standard techniques. For the sake of completeness, we prove this equivalence in Appendix A (see also Section 4.3.1 in [1] and Sections 4.3 and 4.6 in [2]). The advantage of the second formula of Φ ( D ) is that it involves optimization over one parameter only, as opposed to variational calculus in the first formula. Clearly, if the source is discrete rather than continuous, the differential entropy, h ( U ) , should be replaced by ordinary entropy, H ( U ) , and the integration over U should be replaced by summation, as indicated above.
For a source with memory, the SLB is given by
R n ( D ) = min { P V n | U n :   E { ρ ( U n V n ) n D } I ( U n ; V n ) n h ( U n ) n Φ ( D ) ,
which in the limit of large n becomes
R ( D ) lim n h ( U n ) n Φ ( D ) = h ¯ ( U ) Φ ( D ) ,
where h ¯ ( U ) is the differential entropy rate of the source. In all cases, the function Φ ( D ) remains as in (4).

3. Extended Kraft Inequalities

For a given encoder–decoder pair, ( ϕ n , ψ n ) , and parameters α > 1 and β 0 , define the extended Kraft integral (or the extended Kraft sum, in the discrete-alphabet case) as follows:
Z n ( α , β ) = U n exp 2 { α L [ ϕ n ( u n ) ] β ρ ( u n ψ n ( ϕ n ( u n ) ) ) } d u n .
Lemma 1.
Let ( ϕ n , ψ n ) induce a UD lossless compression of v n = ψ n ( ϕ n ( u n ) ) . Then, for every α > 1 and β 0 ,
Z n ( α , β ) U 2 β ρ ( z ) d z n .
Proof of Lemma 1.
Let us examine the expression of [ Z n ( α , β ) ] k for an arbitrary positive integer k.
[ Z n ( α , β ) ] k = U n exp 2 { α L [ ϕ n ( u ) ] β ρ ( u ψ n ( ϕ n ( u ) ) ) } d u k   = U n d u 1 · · · U n d u k · exp 2 α i = 1 k L [ ϕ n ( u i ) ] β i = 1 k ρ ( u i ψ n ( ϕ n ( u i ) ) )   l = 1 { { v i } i = 1 k :   i = 1 k L ( v i ) ] = l } { u 1 :   ψ n ( ϕ n ( u 1 ) ) = v 1 } d u 1 · · · { u k :   ψ n ( ϕ n ( u k ) ) = v k } d u k ×     exp 2 α l β i = 1 k ρ ( u i v i )   l = 1 2 α l { { v i } i = 1 k :   i = 1 k L ( v i ) = l } U n d z 1 U n d z k exp 2 β i = 1 k ρ ( z i )   l = 1 2 α l · 2 l U n d z n exp 2 { β ρ ( z n ) } k   = U n · d z n exp 2 { β ρ ( z n ) } k · l = 1 2 ( α 1 ) l   = U d z exp 2 { β ρ ( z ) } n k 2 α 1 1 ,
where in the second inequality we transformed each integration variable u i into z i = u i v i (a transformation with unit Jacobian) and expanded the integration domain from { u i : ψ n ( ϕ n ( u i ) ) = v i } to U n . Consequently,
Z n ( α , β ) U d z exp 2 { β ρ ( z ) } n ( 2 α 1 1 ) 1 / k ,
which upon taking the limit k becomes
Z n ( α , β ) U 2 β ρ ( z ) d z n ,
completing the proof of Lemma 1. □
For one-to-one codes, instead of taking the limit of k in the proof of Lemma 1, we simply set k = 1 , since the one-to-one property is merely imposed at the level of a single n-block rather than at the level of concatenations of blocks, as in UD codes. This yields the following variation of Lemma 1.
Lemma 2.
Let ( ϕ n , ψ n ) induce a one-to-one mapping between ψ n ( u n ) and v n = ψ n ( ϕ n ( u n ) ) . Then, for every α > 1 and β 0 ,
Z 1 1 n ( α , β ) U 2 β ρ ( z ) d z n 2 α 1 1 ,
where  Z 1 1 n ( α , β )  is defined exactly as  Z n ( α , β ) , except that it may apply to the larger class of one-to-one codes, rather than UD codes.
When α ( 1 , 2 ) , the denominator, 2 α 1 1 , is smaller than unity, and then the upper bound to Z 1 1 n ( α , β ) in Lemma 2 is larger than that of Lemma 1. Indeed, the interesting region for selecting the values of α is in the vicinity of unity, where 2 α 1 1 < 1 .
Returning to the class of UD lossless encodings of v n = ψ n ( ϕ n ( u n ) ) , two additional variations of the above extended Kraft inequality can be considered. The first pertains to D-semifaithful codes, namely, codes for which the distortion is restricted to never exceed n D , pointwise, and not merely in expectation, that is, max u n U n ρ ( u n ψ n ( ϕ n ( u n ) ) ) n D , where D is the allowed per-letter distortion. The second variation is associated with fixed-rate codes, i.e., codes for which L [ ϕ n ( u n ) ] = n R for all u n U n , where R > 0 is the allowed coding rate.
For D-semifaithful codes, let us redefine the integrand of the extended Kraft integral by replacing the term β ρ ( u n ψ n ( ϕ n ( u n ) ) ) in the exponent with the function W ( ρ ( u n ψ n ( ϕ n ( u n ) ) ) n D ) , where W [ · ] is the infinite well function (IWF),
W ( t ) = 0 t 0 t > 0
This causes the integrand of the extended Kraft integral to vanish wherever the distortion constraint is violated, and then the extended Kraft integral becomes
Z D s f n ( α ) = S n ( D ) exp 2 { α L [ ϕ n ( u n ) ] } d u n ,
where S n ( D ) = { u n :   ρ ( u n ψ n ( ϕ n ( u n ) ) ) n D } . In this case, a simple modification of the proof of Lemma 1 yields the following version of the extended Kraft inequality (see also [10]).
Lemma 3.
Let ( ϕ n , ψ n ) be a D-semifaithful code that comprises UD lossless encoding of v n = ψ n ( ϕ n ( u n ) ) . Then, for every α > 1 ,
Z D s f n ( α ) V o l { z n :   ρ ( z n ) n D } .
Here too, if the UD property is replaced by the one-to-one property, then the right-hand side should be divided by 2 α 1 1 .
Finally, for fixed-rate codes, let us replace the term L [ ϕ n ( u n ) ] of the Kraft integrand by n R and return the second term therein to β ρ ( u n ψ n ( ϕ n ( u n ) ) ) . By similar manipulations of the proof of Lemma 1, we find that for fixed-rate codes, the Kraft integral becomes
Z fr n ( α , β ) = U n exp 2 { α n R β ρ ( u n ψ n ( ϕ n ( u n ) ) ) } d u n ,
with the following version of the extended Kraft inequality.
Lemma 4.
Let ( ϕ n , ψ n ) be a rate-R fixed-rate code. Then, for every α 0 and β 0 ,
Z f r n ( α , β ) 2 n ( 1 α ) R · U 2 β ρ ( z ) d z n .

4. Lower Bounds

To fix ideas, we first demonstrate how the classical SLB is obtained from Lemma 1. Consider an arbitrary encoder–decoder pair, ( ϕ n , ψ n ) , in which the reproduction vector, v n = ψ n ( ϕ n ( u n ) ) , is losslessly compressed by a UD code. Then, following Lemma 1, we have the following chain of inequalities:
U 2 β ρ ( z ) d z n ( a ) Z n ( α , β )   = U n exp 2 { α L [ ϕ n ( u n ) ] β ρ ( u n ψ n ( ϕ n ( u n ) ) ) } d u n   = U n P ( u n ) exp 2 { α L [ ϕ n ( u n ) ] β ρ ( u n ψ n ( ϕ n ( u n ) ) ) log P ( u n ) } d u n   = E exp 2 { α L [ ϕ n ( U n ) ] β ρ ( U n ψ n ( ϕ n ( U n ) ) ) log P ( U n ) }   ( b ) exp 2 α E { L [ ϕ n ( U n ) ] } β E { ρ ( U n ψ n ( ϕ n ( U n ) ) ) } E { log P ( U n ) }   = exp 2 α E { L [ ϕ n ( U n ) ] } β E { ρ ( U n ψ n ( ϕ n ( U n ) ) ) } + h ( U n ) ,
where (a) stems from Lemma 1 and (b) is due to Jensen’s inequality applied to the convex function, f ( t ) = 2 t . It follows that
α E { L [ ϕ n ( U n ) ] } + β E { ρ ( U n ψ n ( ϕ n ( U n ) ) ) } h ( U n ) n log U 2 β ρ ( z ) d z .
Since this holds true for every α > 1 while the right-hand side is independent of α , we may take the infimum of the left-hand side in the range α > 1 to obtain, after normalization by n,
E { L [ ϕ n ( U n ) ] } n + β · E { ρ ( U n ψ n ( ϕ n ( U n ) ) ) } n h ( U n ) n log U 2 β ρ ( z ) d z .
At this point, there are several possible perspectives that can be adopted regarding Equation (20). The first is, of course, to view this as a lower bound to the Lagrangian of rate and distortion. The second is to impose an expected distortion constraint, E { ρ ( U n ψ n ( ϕ n ( U n ) ) ) } n D , and then Equation (20) would yield a lower bound to the expected rate according to
E { L [ ϕ n ( U n ) ] } n h ( U n ) n log U 2 β ρ ( z ) d z β D .
Since this holds true for every β 0 while the left-hand side is independent of β , we may maximize the right-hand side over β 0 , to obtain
E { L [ ϕ n ( U n ) ] } n h ( U n ) n inf β 0 log U 2 β ρ ( z ) d z + β D = h ( U n ) n Φ ( D ) ,
thus recovering the classical SLB. In case of multiple distortion constraints, say,
E { ρ j ( U n ψ n ( ϕ n ( U n ) ) ) } n D j ,           j = 1 , 2 , , k ,
the above rate lower bound continues to apply, provided that β , D and ρ ( z ) are redefined as k-dimensional vectors, β = ( β 1 , , β k ) , D = ( D 1 , , D k ) , and ρ ( z ) = ( ρ 1 ( z ) , , ρ k ( z ) ) , and accordingly, β D and β ρ ( z ) are understood to be inner products. The infimum over β is then defined across [ 0 , ) k .
Returning to Equation (20) and to a single distortion criterion, we may alternatively apply a rate constraint E { L [ ϕ n ( U n ) ] } n R and then obtain a lower bound to the expected distortion:
E { ρ ( U n ψ n ( ϕ n ( U n ) ) ) } n sup β 0 1 β h ( U n ) n log U 2 β ρ ( z ) d z R = sup γ 0 γ h ( U n ) n log U 2 ρ ( z ) / γ d z R ,
which is the distortion-rate counterpart of the SLB.
Having established the classical SLB for general variable-rate codes with UD lossless compression of the reproduction vectors, we now carry on to derive some refinements and extensions associated with the other types of codes that we mentioned. The idea would be to apply the same chain of inequalities as in (18) but to invoke Lemma 2 or Lemma 3, or Lemma 4, according to the relevant class of codes, instead of Lemma 1. We next implement this plan for one-to-one codes and for D-semifaithful codes. The same methodology can be applied to fixed-rate codes, but we will not delve into this here.

4.1. One-to-One Codes

For one-to-one codes, we repeat the same derivation as in (18) by invoking Lemma 2 instead of Lemma 1. This yields the following modified version of Equation (19):
α E { L [ ϕ n ( U n ) ] } + β E { ρ ( U n ψ n ( ϕ n ( U n ) ) ) } h ( U n ) n log U 2 β ρ ( z ) d z + log ( 2 α 1 1 ) .
Applying the average distortion constraint, optimizing over β , and normalizing by n, we obtain
E { L [ ϕ n ( U n ) ] } n sup α 1 h ( U n ) / n Φ ( D ) α + log ( 2 α 1 1 ) α n .
The maximization of the right-hand side with respect to α does not seem to lend itself to a closed-form expression, but by selecting α = α n = 1 + c log n n (c being an arbitrary positive constant), it is readily seen that the resulting lower bound becomes
E { L [ ϕ n ( U n ) ] } n h ( U n ) n Φ ( D ) O log n n ,
which is in agreement with the subtraction of O ( ( log n ) / n ) below the entropy for lossless one-to-one codes [11]. This reduction of O ( ( log n ) / n ) is due to the fact that the class of one-to-one codes is broader than the class of UD codes, and therefore the former codes are potentially more capable for a given finite n, albeit the difference fades away as n grows.

4.2. D-Semifaithful Codes

For D-semifaithful codes with UD correspondence between the compressed bit-stream and the reproduction, we obtain from Lemma 3, combined with a derivation like (18),
E { L [ ϕ n ( U n ) ] } h ( U n ) log Vol { S n ( D ) } ,
which is in line with the results of [10], and the remaining issue becomes the assessment of the log-volume of S n ( D ) or an evaluation of a good upper bound to this quantity. One way to proceed is to apply a Chernoff bound, as was suggested by Campbell [9]. For U = R , this amounts to the following derivation:
log Vol { S n ( D ) } = log Vol { z :   ρ ( z ) n D }   = log R n I { ρ ( z ) n D } d z   log inf β 0 R n 2 β [ n D ρ ( z ) ] d z   = n · inf β 0 β D + log R 2 β ρ ( z ) d z   = n Φ ( D ) ,
and we are back to the ordinary SLB,
E { L [ ϕ n ( U n ) ] } n h ( U n ) n Φ ( D ) ,
exactly as we had for UD lossless compression of the reproduction data, and once again, this derivation extends straightforwardly to the case of multiple simultaneous distortion constraints by considering β , D, and ρ ( · ) to be vectors rather than scalars, as mentioned before.
However, since the class of D-semifaithful codes is narrower (and hence more limited to a certain extent) than the class of codes that merely comply with an average distortion constraint, it is conceivable to expect a somewhat tighter (larger) lower bound. Indeed, this turns out to be the case if the log-volume of S n ( D ) is estimated using a more sophisticated analysis tool, namely, the saddle-point method (see, e.g., Chapter 5 in [14] and Chapter 3 in [15], as well as its extension to the multivariate case [16]).
To apply the saddle-point method, the idea is to represent the indicator function in the integrand of the second line in (29) as I { ρ ( z ) n D } = u ( n D ρ ( z ) ) , where u ( t ) is the unit step function, defined as
u ( t ) = 0 t < 0 1 t 0
which in turn is represented as the inverse Laplace transform of the complex function U ( s ) = 1 s , i.e.,
u ( t ) = 1 2 π i R e { s } = c e s t d s s ,
where i = 1 and c is an arbitrary positive real. It follows that
Vol { S n ( D ) } = R n u ( n D ρ ( z ) ) d z   = R n d z 2 π i Re { s } = c e s ( n D ρ ( z ) ) d s s   = 1 2 π i Re { s } = c e s n D d s s R n e s ρ ( z ) d z   = 1 2 π i Re { s } = c e s n D d s s R n exp s t = 1 n ρ ( z t ) d z   = 1 2 π i Re { s } = c e s n D d s s R e s ρ ( z ) d z n   = 1 2 π i Re { s } = c d s s exp n s D + ln R e s ρ ( z ) d z .
This path integral in the complex plane complies with the general form A B g ( s ) e n f ( s ) d s , which under certain regularity conditions can be approximated for large n (see, e.g., Equation (5.7.2) in [14]) according to
A B g ( s ) e n f ( s ) d s = e i θ 2 π n | f ( s ) | · g ( s ) e n f ( s ) · 1 + O 1 n ,
provided that the functions f and g are independent of n and analytic within some connected region D that includes A and B (which are also independent of n). Here, s D is a saddle-point, i.e., a point where f ( s ) = 0 , f ( s ) > 0 and g ( s ) 0 . The angle θ is called the axis and is given by θ = ( π a r g { f ( s ) } ) / 2 . In our case, f ( s ) = s D + ln R e s ρ ( z ) d z , g ( s ) = 1 s , θ = π 2 , and s is the point at which f vanishes, which is assumed strictly positive. But this point of zero derivative is also the point that minimizes the convex function f across the positive reals. It follows then that Vol { S n ( D ) } can be approximated as follows:
Vol { S n ( D ) } = 1 s 2 π n | f ( s ) | · exp n s D + ln R e s ρ ( z ) d z · 1 + O 1 n .
As for the exponential factor, observe that
  exp n s D + ln R e s ρ ( z ) d z = exp n · inf s 0 s D + ln R e s ρ ( z ) d z = exp 2 n ( log 2 e ) · inf s 0 s D + 1 log 2 e · log 2 R 2 s ρ ( z ) log 2 e d z = exp 2 n inf s 0 s D log 2 e + log 2 R 2 s ρ ( z ) log 2 e d z = exp 2 n inf β 0 β D + log 2 R 2 β ρ ( z ) d z = 2 n Φ ( D ) ,
and so,
Vol { S n ( D ) } = 2 n Φ ( D ) s 2 π n | f ( s ) | · 1 + O 1 n ,
which yields
log Vol { S n ( D ) } n = Φ ( D ) log n 2 n O 1 n ,
and then, following Equation (28), we end up with
E { L [ ϕ n ( U n ) ] } n h ( U n ) n Φ ( D ) + log n 2 n + O 1 n .
We therefore observe that the SLB for D-semifaithful codes is given by the ordinary SLB plus redundancy, whose leading term is log n 2 n . This is in agreement with findings of [10], where the derivations corresponded to special cases for which simple geometric and/or combinatorial considerations facilitated the accurate characterization of log Vol { S n ( D ) } , but here the conclusion is more general. Note that we could have been even more precise and also specified the O 1 n term to be 1 n · log [ s 2 π | f ( s ) | ] + O 1 n 2 , but this is, of course, less important.
When k simultaneous distortion constraints are imposed pointwise, i.e.,
max u n R n ρ j ( u n ψ n ( ϕ n ( u n ) ) ) n D j ,       j = 1 , 2 , , k ,
then using the multivariate saddle-point method [16], the above derivation extends with the vector version of the definition of Φ ( D ) replacing the scalar one, and the pre-exponential factor of the saddle-point approximation becomes
( 2 π / n ) k / 2 S det { Hess { f } | s } ,
where S is the product of the components of the k-dimensional vector s —the point at which f ( s ) = 0 , where f ( s ) = s · D + ln R e s · ρ ( z ) d z is defined such that s, D, and ρ ( · ) are k-dimensional vectors, with s · D and s · ρ ( z ) being inner products, as was defined before. Here, det { Hess { f } | s is the determinant of the Hessian of f, computed at s = s . It is assumed, of course, that the k × k matrix Hess { f } | s is non-singular; otherwise there might be redundant (inactive) constraints, which should be removed from the calculation. In this case, the redundancy on top of the ordinary SLB is k log n 2 n + O 1 n , i.e.,
E { L [ ϕ n ( U n ] } n h ( U n ) n Φ ( D ) + k log n 2 n + O 1 n ,
where k k is the effective dimension after the possible removal of redundant constraints.
For example, if k = 2 , ρ 1 ( z ) = | z | , and ρ 2 ( z ) = z 2 , then obviously, 1 n t = 1 n | z t | 2 cannot exceed 1 n t = 1 n z t 2 , and so, if D 1 > D 2 , the constraint t = 1 n | z t | n D 1 is inactive (and hence removable) in the presence of the constraint t = 1 n z t 2 n D 2 . In this case, although k = 2 , the effective dimension is k = 1 . Another obvious example of a superfluous distortion constraint occurs when there is a linear dependence. For instance, let k = 3 and ρ 3 ( z ) = a ρ 1 ( z ) + b ρ 2 ( z ) , where a and b are fixed positive reals. Then, whenever D 3 a D 1 + b D 2 , the third distortion constraint is redundant and then k = 2 (or even less, if there are additional superfluous constraints). More generally, inactive constraints can be identified as those whose corresponding components of s vanish.
Another type of situation where the effective dimension k is smaller than the formal dimension k occurs when the dependence of f ( s ) on some of the components of s disappears in the first place. Consider, for example, the case where k = 2 , ρ 1 ( z ) = z 2 , and the other distortion constraint is max 1 t n | z t | A ; in other words, we impose both an average quadratic constraint and a peak-limited distortion constraint with the motivation of avoiding large spikes in the reconstruction error signal. The peak-limited distortion constraint can be represented as an additive distortion constraint if we select ρ 2 ( z ) = W ( | z | A ) , where W ( · ) is the IWF defined in (13) and the value of D 2 can then be selected to be an arbitrary finite, non-negative real, say D 2 = 0 . Here again k = 1 , but as said, this time, it is because the function f ( s ) depends only on one component of the vector s = ( s 1 , s 2 ) to begin with. To see this, observe that
f ( s ) = s D + ln R e s ρ ( z ) d z   = s 1 D 1 + s 2 D 2 + ln R e s 1 ρ 1 ( z ) + s 2 ρ 2 ( z ) d z   = s 1 D 1 + s 2 · 0 + ln R e s 1 z 2 + s 2 W ( | z | A ) d z   = s 1 D 1 + ln A A e s 1 z 2 d z   = s 1 D 1 + ln π s 1 · [ 1 2 Q ( A 2 s 1 ) ]   = s 1 D 1 + 1 2 ln π s 1 + ln [ 1 2 Q ( A 2 s 1 ) ] ,
where Q ( · ) is the well-known Q-function, defined as
Q ( t ) = t e x 2 / 2 d x 2 π .
Here, although formally there are k = 2 distortion constraints, the saddle-point integration is over one complex variable only, and so the redundancy is log n 2 n + O 1 n on top of the ordinary SLB, h ( U n ) n + Φ ( D ) .

5. Sliding-Window Distortion Constraints

Consider a situation where one wishes not only to control the ‘intensity’ of the reconstruction error signal, t = 1 n ρ ( z i ) , but also possibly to shape its ‘continuity’ or its ‘smoothness’ by imposing additional limitations, say, on the empirical autocorrelations of the error signal, z n = ( z 1 , z 2 , , z n ) , in order to suppress, for example, high frequencies, which might be disturbing for the human eye (in the case of images and video streams) or the human ear (in the case of audio signals). For instance, one might wish to constrain the error signal to obey the first lag autocorrelation constraint, 1 n t = 2 n z t z t 1 0.95 (and perhaps also further lags), in addition to the ordinary mean-square error constraint, say, 1 n t = 1 n z t 2 1 . More generally, consider the case where we impose k additive constraints pertaining to sliding-window functions of the form
t = m n ρ j ( z t m + 1 t ) n D j ,           j = 1 , 2 , , k ,
where m is a positive integer that designates the size of the sliding window. For m = 1 , we are back to ordinary additive distortion constraints, where the single-letter distortion function ρ j operates on single symbols separately. If the sizes of the sliding window are different for the k various constraints, we take m to be the largest one. Accordingly, in the above example where k = 2 and the constraints are 1 n t = 1 n z t 2 1 and 1 n t = 2 n z t z t 1 0.95 , we have ρ 1 ( z ) = z 2 , D 1 = 1 , ρ 2 ( z 1 , z 2 ) = z 1 · z 2 , and D 2 = 0.95 . In this case, the sliding-window size is m = 2 , and the minus signs in the correlation constraint are due to the reversal of the direction of the inequality in 1 n t = 2 n z t z t 1 0.95 , as opposed to the direction of inequalities in our distortion constraints in general.
How does the SLB extend to the case of sliding-window constraints of this type? In this section, we assume that m is fixed while n tends to infinity, and we focus merely on the main terms of the resulting SLB, disregarding the redundancy terms.
A straightforward extension of Lemma 1 and the subsequent derivation in Section 4, so as to apply to a set of k sliding-window distortion constraints with window size m, yields the lower bound
E { L [ ϕ ( U n ) ] } h ( U n ) inf β [ 0 , ) k log U n exp 2 β · t = m n ρ ( z t m + 1 t ) d z n + n β · D ,
where β = ( β 1 , , β k ) , D = ( D 1 , , D k ) , ρ ( z t m + 1 t ) = ( ρ 1 ( z t m + 1 t ) , , ρ k ( z t m + 1 t ) ) , and the dot operations are understood as inner products, as defined earlier.
The multi-dimensional integral
U n exp 2 β · t = m n ρ ( z t m + 1 t ) d z n = U n t = m n exp 2 β · ρ ( z t m + 1 t ) d z n
can be viewed as being obtained by iterated applications of a sliding-window integral operator whose kernel is given by K β ( z t m + 1 t ) = exp 2 β · ρ ( z t m + 1 t ) and the integration is over one component of z n at a time. Under certain regularity conditions, the value of this multi-dimensional integral grows exponentially with an exponential order of [ λ ( β ) ] n , where λ ( β ) is the spectral radius of the operator kernel, namely, the dominant eigenvalue, and then the asymptotic form of the SLB becomes
lim inf n E { L [ ϕ n ( U n ) ] } n h ¯ ( U ) inf β [ 0 , ) k [ log λ ( β ) + β D ] .
There are several equivalent formulas for calculating the spectral radius, λ ( β ) , of a sliding-window kernel K β ( · ) for a given β , for example, the Collatz–Wielandt formula [17,18] and the Donsker–Varadhan formula [19]. Both formulas are associated with calculus of variations. For details, see, e.g., Subsection 3.4 of [20]. In certain special cases, such as those that involve a symmetric kernel for m = 2 , K β ( z , z ) , the Rayleigh quotient formula
λ ( β ) = sup g U 2 g ( z ) K β ( z , z ) g ( z ) d z d z U g 2 ( z ) d z .
can also be used. Note that even in this relatively simple special case, the evaluation of λ ( β ) is a problem of calculus of variations and hence it is not trivial computationally. Clearly, in the finite-alphabet case, the operator kernel reduces to a finite-dimensional matrix, and then the spectral radius is simply the Perron–Frobenius eigenvalue (see, e.g., Theorem 8.2.11 of [21]).
As a simple example, consider the case k = m = 2 , ρ 1 ( z ) = z 2 , D 1 = D , ρ 2 ( z , z ) = z · z , and D 2 = θ , where D > 0 and θ [ 0 , 1 ) are given parameters. Then, following the analogous derivation of Example 4 in [20], we find that the resulting SLB is given by
lim inf n E { L [ ϕ n ( U n ) ] } n h ¯ ( U ) 1 2 log [ 2 π e D ( 1 θ 2 ) ] ,
or, in words, the rate penalty due to the additional autocorrelation constraint is 1 2 log 1 1 θ 2 on top of the SLB pertaining to the ordinary quadratic distortion constraint, h ( U n ) n 1 2 log ( 2 π e D ) .

6. Individual Sequences and Finite-State Encoders

We conclude the technical part of this article by deriving an individual-sequence counterpart of the SLB for finite-alphabet, deterministic sequences, which is based on a variation of Ziv and Lempel’s generalized Kraft inequality (see Lemma 2 of [13]).
Generally speaking, our model for lossy compression of individual sequences is based on the following simple structure. Each source block of length m, u m U m (m—positive integer) is first mapped by an arbitrary reproduction encoder (or vector quantizer) into a reproduction vector,
v m = q ( u m ) V m U m
and then the concatenation of the resulting m-vectors, { v m } , forming the sequence v 1 , v 2 , , is compressed losslessly by a finite-state encoder following the model of [13]. Note that in general, the adoption of this structure does not harm generality and optimality because the mapping between the reproduction vectors and the compressed bit-streams should be bijective (as otherwise additional distortion is injected in the compression phase), and so one may envision a scenario where the encoder holds a copy of the reproduction codebook and cascades it with lossless compression of the reproduction vectors.
To keep this article self-contained, we begin with some basic background on lossless compression of individual sequences by finite-state encoders and the 1978 Lempel–Ziv (LZ78) algorithm. Readers familiar with this background may safely skip Section 6.1 and move on directly to Section 6.2.

6.1. Background

Following the model of [13], consider a setting of lossless compression of v n on the basis of finite-state (FS) encoders. An FS encoder is defined by the set E = ( U , Y , Σ , f , g ) , where U is the finite alphabet of each symbol, v i , which is the same as the source alphabet and whose size is r; Y is a finite collection of binary, variable-length strings, which is allowed to consist of an empty string λ (whose length is zero); Σ is a set of s states of the encoder; f : Σ × U Y is the output function; and g : Σ × U Σ is the next-state function. Given an infinite input reproduction vector (obtained by concatenating infinitely many output vectors from the reproduction encoder), v = ( v 1 , v 2 , ) with v i U , i = 1 , 2 , , the FS encoder E produces an infinite output sequence, y = ( y 1 , y 2 , ) with y i Y , henceforth referred to as the compressed bit-stream, while passing through a sequence of states σ = ( σ 1 , σ 2 , ) with σ i Σ . The encoder is governed recursively by the equations
y i = f ( σ i , v i ) ,
σ i + 1 = g ( σ i , v i ) ,
for i = 1 , 2 , , with a fixed initial state σ 1 = σ Z . If at any step y i = λ , this is referred to as idling, as no output is generated, but only the state is kept updated in response to the input.
An encoder with s states, henceforth called an s-state encoder, is one for which | Σ | = s . For the sake of simplicity, we adopt a few notation conventions from [13]: Given a segment of input symbols v i j with i j and an initial state σ i , we use f ( σ i , v i j ) to denote the corresponding output segment y i j produced by E. Similarly, g ( σ i , v i j ) will denote the final state σ j + 1 after processing the inputs v i j , beginning from state σ i .
An FS encoder E is called information lossless (IL) if, given any initial state σ i Σ , any positive integer n, and any input string, v i i + n , the set ( σ i , f ( σ i , v i i + n ) , g ( σ i , v i i + n ) ) uniquely determines the corresponding input string v i i + n .
The incremental parsing process used by the LZ78 algorithm is a sequential procedure for processing a finite-alphabet input u n . At each step of this process, one determines the shortest string that has not yet occurred as a complete phrase in the current parsed set, with the possible exception of the last phrase, which might be incomplete. For example, applying this parsing method to the sequence
u 15 = 011010011000100
yields
0 , 1 , 10 , 100 , 11 , 00 , 01 , 00 .
Let us denote by c ( u n ) the total number of distinct phrases generated by this procedure (here, c ( u 15 ) = 8 ). In addition, let L Z ( u n ) represent the length in bits of the binary string produced by the LZ78 encoding of u n . By Theorem 2 in [13], the following inequality holds:
L Z ( u n ) [ c ( u n ) + 1 ] log { 2 r [ c ( u n ) + 1 ] }
which can easily be shown to be further upper bounded by
L Z ( u n ) c ( u n ) log c ( u n ) + n · ϵ 1 ( n ) ,
where ϵ 1 ( n ) tends to zero uniformly as n . In words, the LZ78 code length for v n is upper bounded by an expression whose leading term is c ( u n ) log c ( u n ) .
We shall refer to the quantity c ( u n ) log c ( u n ) as the unnormalized LZ complexity of u n to distinguish from the normalized LZ complexity defined as c ( u n ) log c ( u n ) n , which means the per-symbol LZ complexity.

6.2. SLB for Individual Sequences

The following lemma provides a variation of the generalized Kraft inequality of [13].
Lemma 5.
For every IL FS encoder with s states, every α > 1 , every β 0 , and every positive integer ℓ, which is an integer multiple of m,
σ Σ w l U l exp 2 α L [ f ( σ , q ( w l ) ) ] β ρ ( w l q ( w l ) ) s 2 z U 2 β ρ ( z ) l 2 α 1 1 ,
where  q ( w l ) q ( w 1 m , w m + 1 2 m , , w l m + 1 l ) = ( q ( w 1 m ) , q ( w m + 1 2 m ) , , q ( w l m + 1 l ) ) , the latter being defined as in (50) for vectors in  U l .
Proof. 
From the postulated IL property, it follows that given σ Σ , there cannot be more than s 2 k distinct vectors, { v l } , such that L [ f ( σ , v l ) ] = k for every positive integer k. Therefore,
σ Σ w l U l exp 2 α L [ f ( σ , q ( w l ) ) ] β ρ ( w l q ( w l ) ) = σ Σ k 1 { v l :   L [ f ( σ , v l ) ] = k } { w l :   q ( w l ) = v l } exp 2 α k β ρ ( w l q ( w l ) ) σ Σ k 1 s · 2 k · 2 α k z l U l 2 β ρ ( z l ) ) = s 2 · u U 2 β ρ ( z ) l · k = 1 2 ( α 1 ) k = s 2 z U 2 β ρ ( z ) l 2 α 1 1 .
This completes the proof of Lemma 5. □
Now, let divide n. For a given FS IL encoder E, a given u n U n , and the associated state sequence σ n Σ n (generated from u n by E using the next-state function g recursively), consider the joint empirical distribution
P ^ ( σ , w l ) = l n i = 0 n / l 1 I { σ i l + 1 = σ ,   u i l + 1 i l + l = w l } ,         σ Σ ,   w l U l
Then, according to Lemma 5,
  s 2 z U 2 β ρ ( z ) l 2 α 1 1 σ Σ w l U l exp 2 α L [ f ( σ , q ( w l ) ) ] β ρ ( w l q ( w l ) ) = σ Σ w l U l P ^ ( σ , w l ) exp 2 α L [ f ( σ , q ( w l ) ) ] β ρ ( w l q ( w l ) ) log P ^ ( σ , w l ) exp 2 α σ Σ w l U l P ^ ( σ , w l ) L [ f ( σ , q ( w l ) ) ] β w l U l P ^ ( w l ) ρ ( w l q ( w l ) ) + H ^ ( S , U l ) = exp 2 α · l n i = 0 n / l 1 L [ f ( σ i l + 1 , v i l + 1 i l + l ) ] β · l n i = 0 n / l 1 ρ ( u i l + 1 i l + l q ( u i l + 1 i l + l ) ) + H ^ ( S , U l ) = exp 2 α · l n t = 1 n L [ f ( σ t , v t ) ] β · l n t = 1 n ρ ( u t v t ) + H ^ ( S , U l ) ,
where P ^ ( w l ) = σ Σ P ^ ( σ , w l ) and H ^ ( S , U l ) is the joint entropy of the auxiliary random variables S Σ and U l U l , induced by the empirical joint distribution P ^ . It follows then that
    α · 1 n t = 1 n L [ f ( σ t , v t ) ] + β · 1 n t = 1 n ρ ( u t v t )   H ^ ( S , U l ) l log ( s 2 ) l log z U 2 β ρ ( z ) + log ( 2 α 1 1 ) l   H ^ ( U l ) l log z U 2 β ρ ( z ) 2 log s l + log ( 2 α 1 1 ) l   c ( u n ) log c ( u n ) n log z U 2 β ρ ( z ) Δ n ( l ) 2 log s l + log ( 2 α 1 1 ) l ,
where lim n Δ n ( l ) = 1 l , and the last step can be found in Equation (12) of [22] as well as references therein. By taking the limit of l (followed by the limit n ), the last three terms can be made arbitrarily small, whereas the first two terms serve as the individual-sequence counterpart of the right-hand side of Equation (20). Moving the distortion term to the right-hand side and optimizing over β , we obtain
α · 1 n t = 1 n L [ f ( σ t , v t ) ] c ( u n ) log c ( u n ) n inf β 0 β · 1 n t = 1 n ρ ( u t v t ) + log z U 2 β ρ ( z ) Δ n ( l ) 2 log s l + log ( 2 α 1 1 ) l = c ( u n ) log c ( u n ) n Φ 1 n t = 1 n ρ ( u t v t ) Δ n ( l ) 2 log s l + log ( 2 α 1 1 ) l .
Let L max = max σ , v L [ f ( σ , v ) ] . Selecting α = 1 + ζ l , where ζ l tends to zero at a sub-exponential rate (say, ζ l = 1 / l ), we have, on the one hand, the previous inequality and, on the other hand,
α · 1 n t = 1 n L [ f ( σ t , v t ) ] 1 n t = 1 n L [ f ( σ t , v t ) ] + ζ l · L max ,
and so our main result in this section is the following:
1 n t = 1 n L [ f ( σ t , v t ) ] c ( u n ) log c ( u n ) n Φ 1 n t = 1 n ρ ( u t v t )     Δ n ( l ) 2 log s l + log ( 2 ζ l 1 ) l ζ l · L max   c ( u n ) log c ( u n ) n Φ 1 n t = 1 n ρ ( u t v t )     Δ n ( l ) 2 log s l + log ( ζ l ln 2 ) l ζ l · L max ,
with the last four terms tending to zero in the limit of l followed by the limit n .
To conclude, the individual-sequence counterpart of the SLB for finite-alphabet source sequences is essentially of the same form as the classical SLB of the probabilistic setting, except that the normalized entropy term is replaced by the normalized LZ complexity of the sequence and the function Φ ( · ) is calculated at the point of the actual distortion, 1 n t = 1 n ρ ( u t v t ) .

7. Summary and Conclusions

In this article, we have derived several modified versions of the Kraft inequality for lossy compression, which allowed us to derive a few sharper forms and broader types of the classical SLB. This was performed in a unified framework that cover various classes of rate-distortion coding schemes and types of distortion constraints. Two classes of codes were considered: one-to-one codes and D-semifaithful codes (as well as codes that are both one-to-one and D-semifaithful). The other types of distortion constraints are sliding-window constraints. Finally, we also derived an individual-sequence counterpart of the Shannon lower bound, which relies on an extended Kraft inequality for IL FS encoders.
It is pointed out that some of the above-mentioned refinements and extensions required us to invoke more advanced analysis tools than those transitionally used in Shannon theory, such as the saddle-point method (in the context of D-semifaithful codes) and the spectral theory of linear operators (in the context of sliding-window distortion constraints).

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A

Proof of the Second Equality in Equation (4).
Let us define the density function,
g ( z ) = 2 β ρ ( z ) U 2 β ρ ( z ) d z .
Then,
inf β 0 log U 2 β ρ ( z ) d z + β D = inf β 0 sup f [ D ( f g ) ] + log U 2 β ρ ( z ) d z + β D = inf β 0 sup f U d z f ( z ) log 2 β ρ ( z ) f ( z ) + β D = inf β 0 sup f U d z f ( z ) log f ( z ) + β D U f ( z ) ρ ( z ) d z = ( a ) sup f inf β 0 U d z f ( z ) log f ( z ) + β D U f ( z ) ρ ( z ) d z = sup f inf β 0 h ( Z ) + β [ D E { ρ ( Z ) } ] = sup f E { ρ ( Z ) } > D h ( Z ) E { ρ ( Z ) } D = sup { Z :   E { ρ ( Z ) } D } h ( Z ) ,
where (a) is due to the fact that the objective function, U d z f ( z ) log f ( z ) + β [ D U f ( z ) ρ ( z ) d z ] , is concave in f and affine (and hence convex) in β . □

References

  1. Berger, T. Rate Distortion Theory—A Mathematical Basis for Data Compression; Prentice-Hall: Englewood Cliffs, NJ, USA, 1971. [Google Scholar]
  2. Gray, R.M. Source Coding Theory; Kluwer Academic Publishers: Norwell, MA, USA, 1990. [Google Scholar]
  3. Covet, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  4. Linder, T.; Zamir, R. On the asymptotic tightness of the Shannon lower bound. IEEE Trans. Inf. Theory 1994, 40, 2026–2031. [Google Scholar] [CrossRef]
  5. Koch, T. The Shannon lower bound is asymptotically tight. IEEE Trans. Inf. Theory 2016, 62, 6155–6161. [Google Scholar] [CrossRef]
  6. Kostina, V. Data compression with low distortion and finite blocklength. In Proceedings of the 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 29 September–2 October 2015; pp. 1127–1134. [Google Scholar]
  7. Kostina, V. When is Shannon’s lower bound tight at finite blocklength? In Proceedings of the 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 27–30 September 2016; pp. 982–989. [Google Scholar]
  8. Gastpar, M.; Sula, E. Shannon bounds for quadratic rate-distortion problems. IEEE J. Sel. Areas Inf. Theory 2024, 5, 597–608. [Google Scholar] [CrossRef]
  9. Campbell, L.L. Kraft inequality for decoding with respect to a fidelity criterion. IEEE Trans. Inf. Theory 1973, 19, 68–73. [Google Scholar] [CrossRef]
  10. Merhav, N. A comment on “A rate of convergence result for a universal D-semifaithful code”. IEEE Trans. Inf. Theory 1995, 41, 1200–1202. [Google Scholar] [CrossRef]
  11. Rissanen, J. Tight lower bounds for optimum code length. IEEE Trans. Inf. Theory 1982, 28, 348–349. [Google Scholar] [CrossRef]
  12. Zhang, Z.; Yang, E.-h.; Wei, V.K. The redundancy of source coding with a fidelity criterion. I. known statistics. IEEE Trans. Inf. Theory 1997, 43, 71–91. [Google Scholar] [CrossRef]
  13. Ziv, J.; Lempel, A. Compression of individual sequences via variable-rate coding. IEEE Trans. Inf. Theory 1978, 24, 530–536. [Google Scholar] [CrossRef]
  14. de Bruijn, N.G. Asymptotic Methods in Analysis, 2nd ed.; Dover Publications: New York, NY, USA, 1981. [Google Scholar]
  15. Merhav, N.; Weinberger, N. A toolbox for refined information-theoretic analyses with applications. Found. Trends Commun. Inf. Theory 2025, 22, 1–184. [Google Scholar] [CrossRef]
  16. Neuschel, T. Apéry polynomials and the multivariate saddle point method. Contr. Approx. 2014, 40, 487–507. [Google Scholar]
  17. Collatz, L. Einschlieβungssatz für die charakteristischen Zahlen von Matrizen. Math. Z. 1942, 48, 221–226. [Google Scholar]
  18. Wielandt, H. Unzerlegbare, nicht negative Matrizen. Math. Z. 1950, 52, 642–648. [Google Scholar] [CrossRef]
  19. Donsker, M.D.; Varadhan, S.R.S. Asymptotic evaluation of certain Markov process expectations for large time, IV. Commun. Pure Appl. Math. 1983, 36, 183–212. [Google Scholar] [CrossRef]
  20. Merhav, N.; Shamai, S. Volume-based lower bounds to the capacity of the Gaussian channel under pointwise additive input constraints. arXiv 2025, arXiv:2510.04095. [Google Scholar]
  21. Horn, R.A.; Johnson, C.R. Matrix Analysis; Cambridge University Press: New York, NY, USA, 1985. [Google Scholar]
  22. Merhav, N. Universal Slepian-Wolf coding for individual sequences. IEEE Trans. Inf. Theory 2025, 71, 783–796. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Merhav, N. Refinements and Generalizations of the Shannon Lower Bound via Extensions of the Kraft Inequality. Entropy 2026, 28, 76. https://doi.org/10.3390/e28010076

AMA Style

Merhav N. Refinements and Generalizations of the Shannon Lower Bound via Extensions of the Kraft Inequality. Entropy. 2026; 28(1):76. https://doi.org/10.3390/e28010076

Chicago/Turabian Style

Merhav, Neri. 2026. "Refinements and Generalizations of the Shannon Lower Bound via Extensions of the Kraft Inequality" Entropy 28, no. 1: 76. https://doi.org/10.3390/e28010076

APA Style

Merhav, N. (2026). Refinements and Generalizations of the Shannon Lower Bound via Extensions of the Kraft Inequality. Entropy, 28(1), 76. https://doi.org/10.3390/e28010076

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop