Next Article in Journal
Latent Abstractions in Generative Diffusion Models
Next Article in Special Issue
Classical Data in Quantum Machine Learning Algorithms: Amplitude Encoding and the Relation Between Entropy and Linguistic Ambiguity
Previous Article in Journal
Feature-Driven Semantic Communication for Efficient Image Transmission
Previous Article in Special Issue
Influence of Explanatory Variable Distributions on the Behavior of the Impurity Measures Used in Classification Tree Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Successive Refinement for Lossy Compression of Individual Sequences

by
Neri Merhav
The Viterbi Faculty of Electrical and Computer Engineering, Technion–Israel Institute of Technology, Technion City, Haifa 3200003, Israel
Entropy 2025, 27(4), 370; https://doi.org/10.3390/e27040370
Submission received: 23 February 2025 / Accepted: 28 March 2025 / Published: 31 March 2025
(This article belongs to the Collection Feature Papers in Information Theory)

Abstract

:
We consider the problem of successive-refinement coding for lossy compression of individual sequences, namely, compression in two stages, where in the first stage, a coarse description at a relatively low rate is sent from the encoder to the decoder, and in the second stage, an additional coding rate is allocated in order to refine the description and thereby improve the reproduction. Our main result is in establishing outer bounds (converse theorems) for the rate region where we limit the encoders to be finite-state machines in the spirit of Ziv and Lempel’s 1978 model. The matching achievability scheme is conceptually straightforward. We also consider the more general multiple description coding problem on a similar footing and propose achievability schemes that are analogous to the well-known El Gamal–Cover and the Zhang–Berger achievability schemes of memoryless sources and additive distortion measures.

1. Introduction

The notion of successive refinement of information refers to systems in which the reconstruction of the source occurs in multiple stages. In these systems, a single encoder encodes the source and communicates with either one decoder or multiple decoders in a step-by-step manner. At each stage, the encoder transmits a portion of the source information to the corresponding decoder, which also has access to all previous transmissions. Each decoder uses all available transmissions to reconstruct the source, possibly incorporating additional side information. The quality of the reconstruction at each stage (or by each decoder) is evaluated based on a predefined distortion measure. One of the motivations of this hierarchical structure is to allow scalability that meets the available channel resources of the various users who receive the compressed information or even to adapt to time-varying channel conditions of a single user. Several studies have addressed the successive refinement problem for probabilistic sources, most notably, memoryless sources, see, e.g., [1], where necessary and sufficient conditions for simultaneously achieving the rate-distortion function at all stages [2,3], where the rate-distortion region was fully characterized in general, and [4], where source coding error exponents were derived. In some later works, such as [5,6,7,8], successive refinement coding was also considered with the incorporation of side information. Successive-refinement coding is also an important special case of the so called multiple description coding, which in its simplest form, consists of two encoders that send two different individual descriptions of the source to two separate respective decoders (that do not cooperate with each other), and the compressed bit-streams pertaining to those descriptions are also combined and sent to yet another decoder, whose role is to produce a better reconstruction than both of those of the two individual decoders. The problem of fully characterizing the rate-distortion region of the multiple description coding problem is open in its general form, and there are certain inner and outer bounds to this region, see Chapter 13 of [9] for some details and references therein.
In this paper, we focus on successive refinement of information in the context of individual sequences, namely, deterministic source sequences, as opposed to the traditional setting of random sequences governed by a certain probabilistic mechanism. In that sense, this work can be viewed as an additional step in the development of multiuser information theory for individual sequences, following a series of earlier works of this flavor, initiated by Ziv and Lempel in [10,11,12,13], and continued by others in many articles, such as [14,15,16,17,18,19]. In particular, we consider the problem of successive-refinement coding for lossy compression of individual sequences in two stages, where in the first stage, a coarse description at a relatively low rate is sent from the encoder to the decoder, and in the second stage, an additional coding rate is allocated in order to refine the description and thereby improve the reproduction. Our main results are in establishing outer bounds (converse theorems) for the rate region with individual sequences, where we limit the encoders to be finite-state machines similarly as in [11,12,13] and others. The compatible achievability scheme is conceptually straightforward, and so, we believe that the deeper and more interesting part of the contribution of this work is in the converse theorems, namely, in the outer bounds. Our results are formulated and proved for two stages of coding, but their extension to any fixed number of stages is straightforward. These results can also be viewed as an extension of the fixed-distortion results of [14] to successive-refinement coding.
We also consider the more general multiple description coding problem and propose achievability schemes that are analogous to the well-known El Gamal–Cover [20] and the Zhang–Berger [21] achievability schemes of memoryless sources and additive distortion measures. There is a clear parallelism between the rate expressions that we obtain and those of [20,21], including those that are associated with the gaps between the outer bound and the corresponding inner bounds.
The outline of the remaining part of this paper is as follows. In Section 2, we establish notation conventions and formulate the problem setting. In Section 3, we provide some general background on the LZ algorithm as well as on its conditional form. Section 4 is devoted to the above-mentioned successive refinement outer bound, and finally, in Section 5, we address the more general multiple description problem.

2. Notation Conventions and Problem Formulation

Throughout the paper, random variables will be denoted by capital letters, specific values they may take will be denoted by the corresponding lower case letters and their alphabets will be denoted by calligraphic letters. Random vectors, their realizations, and their alphabets will be denoted, respectively, by capital letters, the corresponding lower case letters, and the corresponding calligraphic letters, all superscripted by their dimensions. More specifically, for a given positive integer, n, the source vector  ( x 1 , x 2 , , x n ) , with components,  x i i = 1 , 2 , , n , from a finite-alphabet,  X , will be denoted by  x n . The set of all such n-vectors will be denoted by  X n , which is the n–th order Cartesian power of the single-letter source alphabet,  X . Likewise, reproduction vectors of length n, such as  ( x ^ 1 , , x ^ n )  and  ( x ˜ 1 , , x ˜ n ) , with components,  x ^ i  and  x ˜ i i = 1 , , n , from finite-alphabets,  X ^  and  X ˜ , will be denoted by  x ^ n X ^ n  and  x ˜ n X ˜ n , respectively. An infinite source sequence  ( x 1 , x 2 , )  will be denoted by x. The cardinalities of  X X ^ , and  X ˜  will be denoted by  α β , and  γ , respectively. The value  α =  is allowed, to incorporate countably infinite and continuous source alphabets. By contrast,  β  and  γ  will always be finite.
For two positive integers i and j, where  i j , the notation  x i j  will be used to denote the substring  ( x i , x i + 1 , , x j ) . For  i = 1 , the subscript ‘1’ will be omitted, and so, the shorthand notation of  ( x 1 , x 2 , , x n )  will be  x n , as mentioned before. Similar conventions will apply to other sequences. Probability distributions will be denoted by the letter P with possible subscripts, depending on the context. The indicator function for an event  A  will be denoted by  I { A } , that is,  I { A } = 1  is  A  occurs, and  I { A } = 0 , if not. The logarithmic function,  log x , will be understood to be defined to the base 2. Logarithms to the base e will be denoted by ln. Let  d 1 : X n × X ^ n R +  and  d 2 : X n × X ˜ n R +  be two arbitrary distortion functions between source vectors,  x n , and corresponding reproduction vectors,  x ^ n  and  x ˜ n , respectively.
The successive refinement encoder model is as follows. It is composed of a cascade of two encoders: a reproduction encoder (namely, a vector quantizer) followed by a lossless encoder. The input to the reproduction encoder is the source sequence  x n  and the output is a pair of reproduction vectors,  ( x ^ n , x ˜ n ) X ^ n × X ˜ n  that obeys the distortion constraints,  d 1 ( x n , x ^ n ) n D 1  and  d 2 ( x n , x ˜ n ) n D 2 , where  D 1 0  and  D 2 0  are prescribed normalized distortion levels. There are no particular restrictions imposed on the distortion functions. We denote
B ( x n ) = { ( x ^ n , x ˜ n ) : d 1 ( x n , x ^ n ) n D 1 , d 2 ( x n , x ˜ n ) n D 2 } .
The mapping  X n X ^ n × X ˜ n , employed by the reproduction encoder, is arbitrary and not limited. The pair  ( x ^ n , x ˜ n )  serves as an input to the lossless encoder.
As for the lossless encoder, we follow the same modeling approach as in [11], but with a few adjustments to make it suitable to successive refinement. In particular, the lossless encoder is defined by a set
E = ( X ^ , X ˜ , U , V , S , Z , f 1 , f 2 , g 1 , g 2 ) ,
where  X ^  and  X ˜  are as before;  U  and  V  are two sets of variable-length binary strings, which both include the empty string  λ  of length zero;  S  and  Z  are two sets of states, each one containing q states;  f 1 : S × X ^ U  and  f 2 : Z × X ^ × X ˜ V  are the encoder output functions, and finally,  g 1 : S × X ^ S  and  g 2 : Z × X ^ × X ˜ Z  are two next-state functions. When the lossless encoder is fed by a sequence of pairs  ( x ^ 1 , x ˜ 1 ) , ( x ^ 2 , x ˜ 2 ) , , it outputs a corresponding sequence of pairs of binary strings,  ( u 1 , v 1 ) , ( u 2 , v 2 ) ,  according to the following recursive mechanism. For  t = 1 , 2 , :
u t = f 1 ( s t , x ^ t )
s t + 1 = g 1 ( s t , x ^ t )
v t = f 2 ( z t , x ^ t , x ˜ t )
z t + 1 = g 2 ( z t , x ^ t , x ˜ t ) ,
where the initial states,  s 1  and  z 1 , are assumed arbitrary fixed members of  S  and  Z , respectively. Similarly as in [11], we adopt the extended notation, of  f 1 ( s 1 , x ^ n )  for  u n g 1 ( s 1 , x ^ n )  for  s n + 1 , and similar notations associated with  f 2  and  g 2 .
An encoder E is said to be information lossless if for every positive integer k, the vector  ( s 1 , f 1 ( s 1 , x ^ k ) , g 1 ( s 1 , x ^ k ) )  uniquely determines  x ^ k  and likewise,
( s 1 , z 1 , f 1 ( s 1 , x ^ k ) , f 2 ( z 1 , x ^ k , x ˜ k ) , g 1 ( s 1 , x ˜ k ) , g 2 ( z 1 , x ^ k , x ˜ k ) )
uniquely determines  ( x ^ k , x ˜ k ) . Let  E ( q )  be the set of all information lossless encoders,  { E } , with  | S | q  and  | Z | q .
Given a lossless encoder E and a pair of inputs  ( x ^ n , x ˜ n ) , we define
[ ρ E ( x ^ n ) ] 1 = L ( u n ) n = L [ f 1 ( s 1 , x ^ n ) ] n
and
[ ρ E ( x ^ n , x ˜ n ) ] 2 = L ( u n ) + L ( v n ) n = L [ f 1 ( s 1 , x ^ n ) ] + L [ f 2 ( z 1 , x ^ n , x ˜ n ) ] n ,
where  L ( u n ) = i = 1 n l ( u i ) l ( u i )  being the length (in bits) of the binary string  u i , and similarly for  L ( v n ) . Recall that for the empty string,  λ , we define  l ( λ ) = 0 . The achievable rate region for  x n  that is associated with E is defined as
R E ( x n ) = ( x ^ n , x ˜ n ) B ( x n ) { ( R 1 , R 2 ) : R 1 [ ρ E ( x ^ n ) ] 1 , R 1 + R 2 [ ρ E ( x ^ n , x ˜ n ) ] 2 } ,
and the q-state achievable rate region for  x n  is defined as
R q ( x n ) = E E ( q ) R E ( x n ) .
The rationale behind the union operations in these definitions is that they are two-dimensional set-theoretic analogues of the minimization operations over the appropriate encoders  { E }  and reproduction vectors,  { x ^ n } , that appear for a single coding rate, like in [11,14], as can be noted from the simple relationship:
R : R min E { all   q state   encoders } min { x ^ : d ( x n , x ^ n ) n D } ρ E ( x ^ n ) = E { all   q state   encoders } { x ^ n : d ( x n , x ^ n ) n D } { R : R ρ E ( x ^ n ) }
for a given generic distortion function d and distortion level D.
For later use, we also define the joint empirical distribution of -blocks of  ( x ^ i + 1 i + , x ˜ i + 1 i + ) = 0 , 1 , , n / 1 , provided that divides n. Specifically, consider the empirical distribution,  P ^ = { P ^ ( x ^ , x ˜ ) , x ^ X ^ , x ˜ X ˜ } , of pairs of -vectors, defined as
P ^ ( x ^ , x ˜ ) = n i = 0 n / 1 I { x ^ i + 1 i + = x ^ , x ˜ i + 1 i + = x ˜ } , x ^ X ^ , x ˜ X ˜
Let  H ( X ^ , X ˜ )  denote the joint empirical entropy of an auxiliary pair of random -vectors,  ( X ^ , X ˜ ) , induced by  P ^ , that is,
H ( X ^ , X ˜ ) = ( x ^ , x ˜ ) X ^ × X ˜ P ^ ( x ^ , x ˜ ) log P ^ ( x ^ , x ˜ ) .
Accordingly,  H ( X ^ )  and  H ( X ˜ | X ^ )  will denote the corresponding marginal empirical entropy of  X ^  and the conditional empirical entropy of  X ˜  given  X ^ .
Our objective is to provide inner and outer bounds to the achievable rate region and show that they asymptotically coincide in the limit of large n followed by a limit of large q, in analogy to the asymptotic regime of [11].

3. Background

Before the exposition of the main results and their proofs, we revisit key terms and details related to the 1978 version of the LZ algorithm, also known as the LZ78 algorithm [11], which is the central building block in this work. The incremental parsing procedure of the LZ78 algorithm is a sequential parsing process applied to a finite-alphabet input vector,  x ^ n . According to this procedure, each new phrase is the shortest string not encountered before as a parsed phrase, except for the potential incompleteness of the last phrase. For instance, the incremental parsing of the vector  x ^ 15 = a b b a b a a b b a a a b a a  results in  a , b , b a , b a a , b b , a a , a b , a a . Let  c ( x ^ n )  denote the number of phrases in  x ^ n  resulting from the incremental parsing procedure (in the above example,  c ( x ^ 15 ) = 8 ). Furthermore, let  L Z ( x ^ n )  denote the length of the LZ78 binary compressed code for  x ^ n . According to [11], [Theorem 2] the following inequality holds:
L Z ( x ^ n ) [ c ( x ^ n ) + 1 ] log { 2 β [ c ( x ^ n ) + 1 ] } = c ( x ^ n ) log [ c ( x ^ n ) + 1 ] + c ( x ^ n ) log ( 2 β ) + log { 2 β [ c ( x ^ n ) + 1 ] } = c ( x ^ n ) log c ( x ^ n ) + c ( x ^ n ) log 1 + 1 c ( x ^ n ) + c ( x ^ n ) log ( 2 β ) + log { 2 β [ c ( x ^ k ) + 1 ] } c ( x ^ n ) log c ( x ^ n ) + log e + n ( log β ) log ( 2 β ) ( 1 ε n ) log n + log [ 2 β ( n + 1 ) ] = c ( x ^ n ) log c ( x ^ n ) + n · ϵ ( n ) ,
where we remind that  β  is the cardinality of  X ^ , and where both  ε n  and  ϵ ( n )  tends to zero as  n . In other words, the LZ code-length for  x ^ n  is upper bounded by an expression whose main term is  c ( x ^ n ) log c ( x ^ n ) . On the other hand,  c ( x ^ n ) log c ( x ^ n )  is also known to be the main term of a lower bound (see Theorem 1 of [11]) to the shortest code-length attainable by any information lossless finite-state encoder with no more than q states, provided that  log ( q 2 )  is very small compared to  log c ( x ^ n ) . In view of these facts, we henceforth refer to  c ( x ^ n ) log c ( x ^ n )  as the unnormalized LZ complexity of  x ^ n , whereas the normalized LZ complexity is defined as
ρ L Z ( x ^ n ) = c ( x ^ n ) log c ( x ^ n ) n .
A useful inequality that relates the empirical entropy of non-overlapping -blocks of  x ^ n  (where divides n) and  ρ L Z ( x ^ n )  (see, for example, Equation (26) of [22]), is the following:
H ( X ^ ) ρ L Z ( x n ) log [ 4 β 2 ] log β ( 1 ε n ) log n β 2 log [ 4 β 2 ] n 1 = ρ L Z ( x n ) δ n ( ) ,
It is obtained from the fact that the Shannon code for -blocks can be implemented using a finite-state encoder with no more than  β  states. Specifically, for a block code of length to be implemented by a finite-state machine, one defines the state at each time instant i to be the contents of the input, starting at the beginning of the current block (at time  · i / + 1 ) and ending at time  i 1 . The number of states for an input alphabet of size  β  is then  i = 0 1 β i = ( β 1 ) / ( β 1 ) < β . Therefore, the code-length of this Shannon code must comply with the lower bound of Theorem 1 in [11]. Note that  lim n δ n ( ) = 1 /  and so,  lim lim n δ n ( ) = 0 . Clearly, it is possible to let  = ( n )  increase with n slowly enough such that  δ n ( ( n ) ) 0  as  n , in particular,  ( n )  should be  o ( log n )  for that purpose.
In [22], the notion of the LZ complexity was extended to incorporate finite-state lossless compression in the presence of side information, namely, the conditional version of the LZ complexity. Given  x ^ n  and  x ˜ n , let us apply the incremental parsing procedure of the LZ algorithm to the sequence of pairs  ( ( x ^ 1 , x ˜ 1 ) , ( x ^ 2 , x ˜ 2 ) , , ( x ^ n , x ˜ n ) ) . As mentioned before, according to this procedure, all phrases are distinct with a possible exception of the last phrase, which might be incomplete. Let  c ( x ^ n , x ˜ n )  denote the number of distinct phrases. As an example (taken from [22]), let  n = 6  and consider the sequence pair  ( x ^ 6 , x ˜ 6 )  along with its joint incremental parsing as follows:
x ^ 6 = 0 | 1 | 0 1 | 0 1 | x ˜ 6 = 0 | 1 | 0 0 | 0 1 |
then  c ( x ^ 6 , x ˜ 6 ) = 4 . Let  c ( x ^ n )  denote the resulting number of distinct phrases of  x ^ k  (which may differ from  c ( x ^ n )  in individual parsing of  x ^ n  alone), and let  x ^ ( l )  denote the l-th distinct  x ^ –phrase,  l = 1 , 2 , , c ( x ^ n ) . In the above example,  c ( x ^ 6 ) = 3 . Denote by  c l ( x ˜ n | x ^ n )  the number of occurrences of  x ^ ( l )  in the parsing of  x ^ n , or equivalently, the number of distinct  x ˜ -phrases that jointly appear with  x ^ ( l ) . Clearly,  l = 1 c ( x ^ n ) c l ( x ˜ n | x ^ n ) = c ( x ^ n , x ˜ n ) . In the above example,  x ^ ( 1 ) = 0 x ^ ( 2 ) = 1 x ^ ( 3 ) = 01 c 1 ( x ˜ 6 | x ^ 6 ) = c 2 ( x ˜ 6 | x ^ 6 ) = 1 , and  c 3 ( x ˜ 6 | x ^ 6 ) = 2 . Now, the conditional LZ complexity of  x ˜ n  given  x ^ n  is defined as
ρ L Z ( x ˜ n | x ^ n ) = 1 n l = 1 c ( x ^ n ) c l ( x ˜ n | x ^ n ) log c l ( x ˜ n | x ^ n ) .
In [22] it was shown that  ρ L Z ( x ˜ n | x ^ n )  is the main term of the compression ratio achieved by the conditional version of the LZ algorithm for compressing  x ˜ n  in the presence of the side information  x ^ n , available to both encoder and decoder—see the compression scheme described in [22] (see also [15]), i.e., the length function,  L Z ( x ˜ n | x ^ n ) , of the coding scheme proposed therein is upper bounded (in parallel to (13)) by
L Z ( x ˜ n | x ^ n ) n ρ L Z ( x ˜ n | x ^ n ) + n ϵ ^ ( n ) ,
where  ϵ ^ ( n ) = O log ( log n ) log n  (see Equations (10) and (11) in [15]). On the other hand, analogously to [11] [Theorem 1], it was shown in [16], that  ρ L Z ( x ˜ n | x ^ n )  is also the main term of a lower bound to the compression ratio that can be achieved by any finite-state encoder with side information at both ends, provided that the number of states is not too large, similarly as described above for the unconditional version.
The inequality (15) also extends to the conditional case as follows (see [16]):
H ( X ˜ | X ^ ) ρ L Z ( x ˜ n | x ^ n ) δ n ( ) ,
where  δ n ( )  is the same as  δ n ( )  except that  β  therein is replaced by  ( β γ )  to accommodate the number of states associated with the conditional version of the aforementioned Shannon code applied to -blocks. By the same token, we also have
H ( X ^ , X ˜ ) ρ L Z ( x ^ n , x ˜ n ) δ n ( ) .

4. The Outer Bound for Successive Refinement

Our main result for finite-state encoders is the following.
Theorem 1.
For every  x n X n ,
R q ( x n ) R o ( x n ) = ( x ^ n , x ˜ n ) B ( x n ) R L Z ( x ^ n , x ˜ n ) ,
where
R L Z ( x ^ n , x ˜ n ) = { ( R 1 , R 2 ) : R 1 ρ L Z ( x ^ n ) Δ 1 ( q , n ) , R 1 + R 2 ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n | x ^ n ) Δ 2 ( q , n ) } ,
where  Δ 1 ( q , n )  and  Δ 2 ( q , n )  are defined as
Δ 1 ( q , n ) = log ( 4 q 2 ) log β ( 1 ϵ n ) log n + q 2 log ( 4 q 2 ) n
with  ϵ n 0  as  n , and
Δ 2 ( n , q ) = min { : d i v i d e s n } δ n ( ) + δ n ( ) + 1 log q 4 1 + log 1 + β γ q 4 .
Discussion 1.
Several comments are now in order.
1.
Since both  Δ 1 ( q , n )  and  Δ 2 ( q , n )  tend to zero as  n  for fixed q, the asymptotic achievability of  R o ( x n )  is conceptually straightforward: Given an internal point,  ( R 1 , R 2 ) R o ( x n ) , there must be at least one pair  ( x ^ n , x ˜ n ) B ( x n )  such that  ( R 1 , R 2 ) R L Z ( x ^ n , x ˜ n ) . Upon finding such a pair, proceed as follows: At the first stage, apply LZ78 compression to  x ^ n  at a coding rate of  R 1 = L Z ( x ^ n ) n  which is only slightly above  ρ L Z ( x ^ n )  for large n, as discussed in Section 3. At the second stage, apply conditional LZ compression of  x ˜ n  given  x ^ n  as side information at both ends, at an incremental coding rate of  R 2 = L Z ( x ˜ n | x ^ n ) n  which is close to  ρ L Z ( x ˜ n | x ^ n ) , as also explained in Section 3, and then the total rate,  R 1 + R 2 , is about  ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n | x ^ n ) . Similarly as in [11], there is still a certain gap between the achievability and the converse theorem because the achievability requires encoders whose number of states is not small compared to n, whereas the converse is significant when q is very small relative to n. As in [11], this gap can be closed in the asymptotic limit of large q by partitioning the sequence into non-overlapping blocks and starting over the LZ compression mechanism in each block separately. We will address this point in detail later on.
2.
Considering that the achievability is conceptually straightforward, as explained in item no. 1 above, the interesting and deeper result is the converse theorem. Since the second stage encoder receives both  x ^ n  and  x ˜ n  as inputs, it is immediate to lower bound the total coding rate, at the second stage, in terms of the joint compressibility of  ( x ^ n , x ˜ n ) , namely by  ρ L Z ( x ^ n , x ˜ n ) , but recall that the first-stage encoder must have already allocated a rate at least as large as  ρ L Z ( x ^ n ) , then in order to meet a lower bound of  ρ L Z ( x ^ n , x ˜ n )  on the total coding rate, the incremental rate,  R 2 , of the second stage must not exceed  ρ L Z ( x ^ n , x ˜ n ) ρ L Z ( x ^ n ) , and there is no apparent way to achieve such a coding rate, as far as the author can see. Nonetheless, since we can also lower bound the total rate of both stages by  ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n | x ^ n ) , then the achievability becomes obvious, as said. This point is not trivial because there is no chain rule that applies to the LZ complexities of arbitrary finite sequences. The proof that  ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n | x ^ n )  also serves as a lower bound (essentially), which requires a certain manipulation by using a generalized Kraft inequality and passing via empirical entropies, as can be seen in the proof.
3.
The choice of  x ^ n  exhibits a trade-off between the coding rate of the first stage and the incremental rate at the second stage because  x ^ n  is both compressed at the first stage and serves as side information at the second stage, so there might be a certain tension between selecting  x ^ n  for having small  ρ L Z ( x ^ n )  and selecting it for small  ρ L Z ( x ˜ n | x ^ n ) . Of course, an analogous tension exists also in successive refinement for memoryless sources [3]. The reproduction encoder must select  ( x ^ n , x ˜ n ) B ( x n )  that best compromises these criteria.
4.
The results extend straightforwardly to any finite number of stages, where at each stage one applies conditional LZ compression of the current reproduction given all previous reproductions.
Proof of Theorem 1.
We begin with the first stage. By definition, if  ( R 1 , R 2 ) R q ( x n ) , then there must exist an encoder  E E ( q )  and  ( x ^ n , x ˜ n ) B ( x n )  such that  R 1 [ ρ E ( x ^ n ) ] 1  and  R 1 + R 2 [ ρ E ( x ^ n , x ˜ n ) ] 2 . Now, according to Theorem 1 of [11]:
[ ρ E ( x ^ n ) ] 1 c ( x ^ n ) + q 2 n · log c ( x ^ n ) + q 2 4 q 2 + 2 q 2 n > c ( x ^ n ) + q 2 n · log [ c ( x ^ n ) + q 2 ] c ( x ^ n ) + q 2 n log ( 4 q 2 ) > c ( x ^ n ) log c ( x ^ n ) n c ( x ^ n ) log ( 4 q 2 ) n q 2 log ( 4 q 2 ) n c ( x ^ n ) log c ( x ^ n ) n log ( 4 q 2 ) log β ( 1 ϵ n ) log n q 2 log ( 4 q 2 ) n = ρ L Z ( x ^ n ) Δ 1 ( q , n ) ,
where  lim n ϵ n = 0  and the last inequality is an application of Equation (6) in [11]. Since  R 1 [ ρ E ( x ^ n ) ] 1 , it follows that
R 1 ρ L Z ( x ^ n ) Δ 1 ( q , n ) .
Moving on to the combined encoder of both stages, consider the following. According to Lemma 2 of [11] and due to the postulated information losslessness, the combined encoder, which has  q 2  states, must obey the following generalized Kraft inequality:
( x ^ , x ˜ ) X ^ × X ˜ exp 2 [ min s S L [ f 1 ( s , x ^ ) ] + min z Z L [ f 2 ( z , x ^ , x ˜ ) ] q 4 1 + log 1 + β γ q 4 .
This implies that the description length at the output of this encoder is lower bounded as follows.
n ( R 1 + R 2 ) n [ ρ E ( x ^ n , x ˜ n ) ] 2 = L ( u n ) + L ( v n ) = t = 1 n { L [ f 1 ( s t , x ^ t ) ] + L [ f 2 ( z t . x ^ t , x ˜ t ) ] } = m = 0 n / 1 j = 1 { L [ f 1 ( s m + j , x ^ m + j ) ] + L [ f 2 ( z m + j , x ^ m + j ) , x ˜ m + j ) ] } = m = 0 n / 1 { L [ f 1 ( s m + 1 , x ^ m + 1 m + ) ] + L [ f 2 ( z m + 1 , x ^ m + 1 m + , x ˜ m + 1 m + ) ] } m = 0 n / 1 min s S L [ f 1 ( s , x ^ m + 1 m + ) ] + min z Z L [ f 2 ( z , x ^ m + 1 m + , x ˜ m + 1 m + ) ] = n ( x ^ , x ˜ ) X ^ × X ˜ P ^ ( x ^ , x ˜ ) · min s S L [ f 1 ( s , x ^ ) ] + min z Z L [ f 2 ( z , x ^ , x ˜ ) ]
and so,
R 1 + R 2 1 ( x ^ , x ˜ ) X ^ × X ˜ P ^ ( x ^ , x ˜ ) · min s S L [ f 1 ( s , x ^ ) ] + min z Z L [ f 2 ( z , x ^ , x ˜ ) ] .
Now, by the generalized Kraft inequality above,
q 4 1 + log 1 + β γ q 4 ( x ^ , x ˜ ) X ^ × X ˜ exp 2 min s S L [ f 1 ( s , x ^ ) ] + min z Z L [ f 2 ( z , x ^ , x ˜ ) ] = ( x ^ , x ˜ ) X ^ × X ˜ P ^ ( x ^ , x ˜ ) · exp 2 min s S L [ f 1 ( s , x ^ ) + min z Z L [ f 2 ( z , x ^ , x ˜ ) ] log P ^ ( x ^ , x ˜ ) exp 2 ( x ^ , x ˜ ) X ^ × X ˜ P ^ ( x ^ , x ˜ ) · min s S L [ f 1 ( s , x ^ ) + [ min z Z L [ f 2 ( z , x ^ , x ˜ ) ] + H ( X ^ , X ˜ ) ,
where the last inequality follows from the convexity of the exponential function and Jensen’s inequality. This yields
log q 4 1 + log 1 + β γ q 4 H ( X ^ , X ˜ ) ( x ^ , x ˜ ) X ^ × X ˜ P ^ ( x ^ , x ^ ) · min s S L [ f 1 ( s , x ^ ) ] + min z Z L [ f 2 ( z , x ^ , x ˜ ) ] ,
implying that
R 1 + R 2 L ( u n ) + L ( v n ) n 1 ( x ^ , x ˜ ) X ^ × X ˜ P ^ ( x ^ , x ˜ ) · min s S L [ f 1 ( s , x ^ ) ] + min z Z L [ f 2 ( z , x ^ , x ˜ ) ] H ( X ^ , X ˜ ) 1 log q 4 1 + log 1 + β γ q 4 = H ( X ^ ) + H ( X ˜ | X ^ ) 1 log q 4 1 + log 1 + β γ q 4
Now, according to Equation (15),
H ( X ^ ) ρ L Z ( x ^ n ) δ n ( ) .
Similarly, according to Equation (19),
H ( X ˜ | X ^ ) ρ L Z ( x ˜ n | x ^ n ) δ n ( ) ,
and so,
R 1 + R 2 ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n | x ^ n ) δ n ( ) δ n ( ) 1 log q 4 1 + log 1 + β γ q 4 .
Maximizing this lower bound w.r.t. yields
R 1 + R 2 ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n | x ^ n ) Δ 2 ( n , q ) ,
where
Δ 2 ( n , q ) = min { : divides n } δ n ( ) + δ n ( ) + 1 log q 4 1 + log 1 + β γ q 4 .
This completes the proof of Theorem 1. □
Referring to the last part of comment no. 1 in the discussion that follows Theorem 1, we now address the gap in terms of the number of states. For an infinite source sequence  x = ( x 1 , x 2 , ) , we define the q-state achievable rate region for x as
R q ( x ) = m 1 n m R q ( x n ) ,
and finally, the finite-state achievable rate region for x is defined as
R ( x ) = q 1 R q ( x ) .
These definitions are two-dimensional counterparts of Equations (2)–(4) in [11], where the finite-state (lossless) compressibility of x is defined in several steps. In particular, the union over intersections in the definition of  R q ( x )  is the set-theoretic analogue of the limit superior operation, and the union operation in the definition of  R ( x )  is parallel to a limit of  q .
Let k be a positive integer that divides n, and consider the partition of  x ^ n  and  x ˜ n  into  n / k  blocks of length k, i.e.,  x ^ k t + 1 k t + k = ( x ^ k t + 1 , x ^ k t + 2 , , x ^ k t + k )  and  x ˜ k t + 1 k t + k = ( x ˜ k t + 1 , x ˜ k t + 2 , , x ˜ k t + k ) t = 0 , 1 , , n / k 1 . Next, define:
R k ( x ^ n , x ˜ n ) = { ( R 1 , R 2 ) : R 1 k n t = 0 n / k 1 ρ L Z ( x ^ k t + 1 k t + k ) Δ 1 ( q , k ) , R 1 + R 2 k n t = 0 n / k 1 [ ρ L Z ( x ^ k t + 1 k t + k ) + ρ L Z ( x ˜ k t + 1 k t + k | x ^ k t + 1 k t + k ) ] Δ 2 ( q , k ) } .
Then, similarly as in Theorem 1,
R q ( x n ) R o k ( x n ) = ( x ^ n , x ˜ n ) B ( x n ) R k ( x ^ n , x ˜ n ) ,
and so, for every positive integer N:
n N R q ( x n ) n N R o k ( x n ) ,
implying that
R q ( x ) = N 1 n N R q ( x n ) N 1 n N R o k ( x n ) = R o k ( x ) .
Since this holds for every positive integer k, then
R q ( x ) K 1 k K R o k ( x ) = R o ( x ) ,
and so,
R ( x ) R o ( x ) ,
which establishes an asymptotic version of the converse theorem.
As for the direct part, considering the fact that a block code of length k, operating on k-tuples of the two reconstruction vectors can be implemented by a finite-state machine with no more than  ( β γ ) k  states, we have
R ( β γ ) k ( x n ) R i k ( x n ) = ( x ^ n , x ˜ n ) B ( x n ) R + k ( x ^ n , x ˜ n ) ,
where
R + k ( x ^ n , x ˜ n ) = { ( R 1 , R 2 ) : R 1 k n t = 0 n / k 1 ρ L Z ( x ^ k t + 1 k t + k ) + O 1 log k , R 1 + R 2 k n t = 0 n / k 1 [ ρ L Z ( x ^ k t + 1 k t + k ) + ρ L Z ( x ˜ k t + 1 k t + k | x ^ k t + 1 k t + k ) ) ] + O log ( log k ) log k } .
R ( β γ ) k ( x ) = N 1 n N R ( β γ ) k ( x n ) N 1 n N R i k ( x n ) = R i k ( x ) K k R i K ( x )
and so,
R ( x ) = k 1 R ( β γ ) k ( x ) k 1 K k R i K ( x ) = R i ( x ) .
We have just proved the following theorem:
Theorem 2.
For every infinite individual sequence  x = ( x 1 , x 2 , ) ,
R o ( x ) R ( x ) R i ( x ) .
These inner and outer bounds are tight in the sense that the definitions of  R o ( x )  and  R i ( x )  are based on the same building blocks and the only difference is in terms that tend to zero as  k .

5. Multiple Description Coding

Consider next the configuration that is associated with the multiple description problem (see, e.g., Chapter 13 of [9]), where the source is an individual sequence and the encoders are modeled as finite-state machines. In particular, there are two q-state encoders and three decoders, which are defined as follows. Encoders 1 and 2 are fed by  x n  and produce two reconstructions,  x ^ n  and  x ˜ n , with distortions  d 1 ( x n , x ^ n ) n D 1  and  d 2 ( x n , x ˜ n ) n D 2 , respectively. Encoder 1 then compresses  x ^ n  losslessly and sends a compressed description to Decoder 1. Likewise, Encoder 2 does the same with  x ˜ n  and sends a compressed form to Decoder 2. There is no collaboration between Decoders 1 and 2. The third decoder, Decoder 0, receives both compressed descriptions and generates yet another reconstruction,  x ` n , with distortion  d 0 ( x n , x ` n ) n D 0 .
Using the same technique as in the proof of Theorem 1, it is easy to prove the following outer bound to the achievable rate region:
R o ( x n ) = ( x ^ n , x ˜ n , x ` n ) B ( x n ) R L Z ( x ^ n , x ˜ n , x ` n ) ,
where  B ( x n )  is redefined as
B ( x n ) = ( x ^ n , x ˜ n , x ` n ) : d 0 ( x n x ` n ) n D 0 , d 1 ( x n x ^ n ) n D 1 , d 2 ( x n x ˜ n ) n D 2 ,
and
R L Z ( x ^ n , x ˜ n , x ` n ) = { ( R 1 , R 2 ) : R 1 ρ L Z ( x ^ n ) Δ 1 ( q , n ) , R 2 ρ L Z ( x ˜ n ) Δ 1 ( q , n ) , R 1 + R 2 ρ L Z ( x ` n | x ^ n , x ˜ n ) + ρ L Z ( x ^ n , x ˜ n ) Δ 2 ( q , n ) } ,
with  Δ 1 ( q , n )  and  Δ 2 ( q , n )  being defined similarly as before. The sum-rate inequality is obtained by considering that the two encoders together compress losslessly the triple  ( x ^ n , x ˜ n , x ` n )  and so, the main term of the lower bound to  R 1 + R 2  is the joint empirical entropy of of  ( x ^ n , x ˜ n , x ` n ) , which can be decomposed as the sum of the joint empirical entropy of  ( x ^ n , x ˜ n )  and the conditional empirical entropy of  x ` n  given  ( x ^ n , x ˜ n ) , which in turn are essentially further lower bounded by  ρ L Z ( x ^ n , x ˜ n )  and  ρ L Z ( x ` n | x ^ n , x ˜ n ) , respectively.
We next present two inner bounds, where the first one is analogous to the El Gamal–Cover inner bound [20] and the second follows the same line of thought as that of the Zhang–Berger inner bound [21].
The former inner bound is given by
R i E G C ( x n ) = ( x ^ n , x ˜ n , x ` n ) B ( x n ) R i ( x ^ n , x ˜ n , x ` n ) ,
where
R i ( x ^ n , x ˜ n , x ` n ) = { ( R 1 , R 2 ) : R 1 ρ L Z ( x ^ n ) + ϵ ( n ) , R 2 ρ L Z ( x ˜ n ) + ϵ ( n ) , R 1 + R 2 ρ L Z ( x ` n | x ^ n , x ˜ n ) + ρ L Z ( x ^ n , x ˜ n ) + I ^ ( x ^ n ; x ˜ n ) + ϵ ( n ) + ϵ ^ ( n ) } ,
where  ϵ ( n )  and  ϵ ( n ) ^  are as in (13) and (18), respectively.
I ^ ( x ^ n ; x ˜ n ) = ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n ) ρ L Z ( x ^ n , x ˜ n ) .
The quantity  I ^ ( x ^ n ; x ˜ n )  plays a role of an empirical mutual information between  x ^ n  and  x ˜ n , which manifests the gap between the lower bounds to the sum-rate inequalities of the inner bound and the outer bound, analogously to the mutual information term of the El Gamal–Cover achievable region.
The achievability of the above inner bound is as follows. Given an internal point in  R i E G C ( x n ) , there must exist a reconstruction triple  ( x ^ n , x ˜ n , x ` n )  that meets the distortion constraints and the corresponding rate inequalities. The encoder applies individual LZ compression for both  x ^ n  and  x ˜ n  and sends the compressed versions, at rates  ρ L Z ( x ^ n )  and  ρ L Z ( x ˜ n )  (up to negligibly small terms for large n), to Decoder 1 and Decoder 2, respectively. It then applies conditional LZ compression of  x ` n  given  ( x ^ n , x ˜ n )  at rate  ρ L Z ( x ` n | x ^ n , x ˜ n )  (up to small terms), and splits this compressed bit stream between Decoders 1 and 2 without violating their rate inequalities. The rate sum is then essentially
ρ L Z ( x ` n | x ^ n , x ˜ n ) + ρ L Z ( x ^ n ) + ρ L Z ( x ˜ n ) = ρ L Z ( x ` n | x ^ n , x ˜ n ) + ρ L Z ( x ^ n , x ˜ n ) + I ^ ( x ^ n ; x ˜ n ) .
In the above description, we explained that the bit-stream associated with the conditional compression of  x ` n  given  ( x ^ n , x ˜ n )  is split between Decoders 1 and 2 without violating the rate inequalities. This is always possible because of the following simple fact: Given an internal point in the region  R = { ( R 1 , R 2 ) : R 1 > A , R 2 > B , R 1 + R 2 A + B + C } , there must exist  0 D C  such that  ( A + D , B + C D ) R . In particular, let  D = R 1 A 0 . Then,  R 2 = B + C D B  and  R 1 + R 2 = ( A + D ) + ( B + C D ) = A + B + C . In our case,  A = ρ L Z ( x ^ n ) B = ρ L Z ( x ˜ n ) , and  C = ρ L Z ( x ` n | x ^ n , x ˜ n ) .
The second achievability scheme, in the spirit of the Zhang–Berger scheme, is as follows. Here, in addition to  x ^ n x ˜ n  and  x ` n , we also generate an auxiliary finite-alphabet sequence,  u n . The encoder applies LZ compression to  u n  and conditional LZ compression of  x ^ n  given  u n  and sends both bit-streams to Decoder 1. At the same time, it also applies conditional LZ compression of  x ˜ n  given  u n  and sends the compressed forms of  u n  and  x ˜ n  to Decoder 2. Finally, the encoder applies conditional LZ compression of  x ` n  given  ( x ^ n , x ˜ n , u n )  and splits the compressed bit-stream between Decoders and Decoder 2 in a manner that meets the rate constraints. Thus,
R 1 ρ L Z ( u n ) + ρ L Z ( x ^ n | u n ) + α ρ L Z ( x ` n | x ^ n , x ˜ n , u n )
R 2 ρ L Z ( u n ) + ρ L Z ( x ˜ n | u n ) + ( 1 α ) ρ L Z ( x ` n | x ^ n , x ˜ n , u n ) ,
where  α [ 0 , 1 ] . Thus,
R 1 + R 2 2 · ρ L Z ( u n ) + ρ L Z ( x ^ n | u n ) + ρ L Z ( x ˜ n | u n ) + ρ L Z ( x ` n | x ^ n , x ˜ n , u n ) = 2 · ρ L Z ( u n ) + ρ L Z ( x ^ n , x ˜ n | u n ) + ρ L Z ( x ` n | x ^ n , x ˜ n , u n ) + I ^ ( x ^ n ; x ˜ n | u n ) ,
where
I ^ ( x ^ n ; x ˜ n | u n ) = ρ L Z ( x ^ n | u n ) + ρ L Z ( x ˜ n | u n ) ρ L Z ( x ^ n , x ˜ n | u n ) ,
is analogous to conditional mutual information. The first term in the rate sum is analogous to  2 I ( U ; X ) , the sum of the second and the third is analogous to  I ( X ; X ^ , X ˜ , X ` | U )  and the last term is analogous to  I ( X ^ ; X ˜ | U )  (see Theorem 13.4, p. 332 in [9]).

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Equitz, W.H.R.; Cover, T.M. Successive refinement of information. IEEE Trans. Inform. Theory 1991, 37, 269–275. [Google Scholar]
  2. Koshelev, V.N. On the divisibility of discrete sources with an additive single-letter distortion measure. Probl. Inf. Transm. (IPPI) 1994, 30, 27–43. [Google Scholar]
  3. Rimoldi, B. Successive refinement of information: Characterization of achievable rates. IEEE Trans. Inform. Theory 1994, 40, 253–259. [Google Scholar] [CrossRef]
  4. Kanlis, A.; Narayan, P. Error exponents for successive refinement by partitioning. IEEE Trans. Inform. Theory 1996, 42, 275–282. [Google Scholar] [CrossRef]
  5. Steinberg, Y.; Merhav, N. On successive refinement for the Wyner–Ziv problem. IEEE Trans. Inform. Theory 2004, 50, 1636–1654. [Google Scholar]
  6. Maor, A.; Merhav, N. On successive refinement with causal side information at the decoders. IEEE Trans. Inform. Theory 2008, 54, 332–343. [Google Scholar]
  7. Maor, A.; Merhav, N. On successive refinement for the Kaspi/Heegard–Berger problem. IEEE Trans. Inform. Theory 2010, 56, 3930–3945. [Google Scholar]
  8. Tian, C.; Diggavi, S.N. Multistage successive refinement for Wyner-Ziv source coding with degraded side informations. In Proceedings of the 2006 IEEE International Symposium on Information Theory (ISIT 2006), Seattle, WA, USA, 9–14 July 2006; pp. 1594–1598. [Google Scholar] [CrossRef]
  9. El Gamal, A.; Kim, Y.-H. Network Information Theory; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  10. Lempel, A.; Ziv, J. On the complexity of finite sequences. IEEE Trans. Inform. Theory 1976, 22, 75–81. [Google Scholar]
  11. Ziv, J.; Lempel, A. Compression of individual sequences via variable-rate coding. IEEE Trans. Inform. Theory 1978, 24, 530–536. [Google Scholar]
  12. Ziv, J. Distortion-rate theory for individual sequences. IEEE Trans. Inform. Theory 1980, 26, 137–143. [Google Scholar]
  13. Ziv, J. Fixed-rate encoding of individual sequences with side information. IEEE Trans. Inform. Theory 1984, 30, 348–352. [Google Scholar]
  14. Yang, E.-H.; Kieffer, J.C. Simple universal lossy data compression schemes derived from the Lempel-Ziv algorithm. IEEE Trans. Inform. Theory 1996, 42, 239–245. [Google Scholar] [CrossRef]
  15. Uyematsu, T.; Kuzuoka, S. Conditional Lempel-Ziv complexity and its application to source coding theorem with side information. IEICE Trans. Fundam. 2003, 86, 2615–2617. [Google Scholar]
  16. Merhav, N. Universal detection of messages via finite–state channels. IEEE Trans. Inform. Theory 2000, 46, 2242–2246. [Google Scholar]
  17. Merhav, N. A universal random coding ensemble for sample-wise lossy compression. Entropy 2023, 25, 1199. [Google Scholar] [CrossRef] [PubMed]
  18. Merhav, N. Lossy compression of individual sequences revisited: Fundamental limits of finite-state encoders. Entropy 2024, 26, 116. [Google Scholar] [CrossRef] [PubMed]
  19. Merhav, N. Universal Slepian-Wolf coding for individual sequences. IEEE Trans. Inform. Theory 2025, 71, 783–796. [Google Scholar]
  20. El Gamal, A.; Cover, T.M. Achievable rates for multiple descriptions. IEEE Trans. Inform. Theory 1982, 28, 851–857. [Google Scholar] [CrossRef]
  21. Zhang, Z.; Berger, T. New results in binary multiple descriptions. IEEE Trans. Inform. Theory 1987, 33, 502–521. [Google Scholar]
  22. Ziv, J. Universal decoding for finite-state channels. IEEE Trans. Inform. Theory 1985, 31, 453–460. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Merhav, N. Successive Refinement for Lossy Compression of Individual Sequences. Entropy 2025, 27, 370. https://doi.org/10.3390/e27040370

AMA Style

Merhav N. Successive Refinement for Lossy Compression of Individual Sequences. Entropy. 2025; 27(4):370. https://doi.org/10.3390/e27040370

Chicago/Turabian Style

Merhav, Neri. 2025. "Successive Refinement for Lossy Compression of Individual Sequences" Entropy 27, no. 4: 370. https://doi.org/10.3390/e27040370

APA Style

Merhav, N. (2025). Successive Refinement for Lossy Compression of Individual Sequences. Entropy, 27(4), 370. https://doi.org/10.3390/e27040370

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop