Next Article in Journal
The Method of Types for the AWGN Channel
Next Article in Special Issue
Extended Blahut–Arimoto Algorithm for Semantic Rate-Distortion Function
Previous Article in Journal
Kinetic Monte Carlo Modeling of the Spontaneous Deposition of Platinum on Au(111) Surfaces
Previous Article in Special Issue
Importance-Aware Resource Allocations for MIMO Semantic Communication
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Soft Classification in a Composite Source Model

Department of Electronic Engineering and Information Science, University of Science and Technology of China, Hefei 230026, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(6), 620; https://doi.org/10.3390/e27060620
Submission received: 28 April 2025 / Revised: 3 June 2025 / Accepted: 6 June 2025 / Published: 11 June 2025
(This article belongs to the Special Issue Semantic Information Theory)

Abstract

A composite source model consists of an intrinsic state and an extrinsic observation. The fundamental performance limit of reproducing the intrinsic state is characterized by the indirect rate–distortion function. In a remote classification application, a source encoder encodes the extrinsic observation (e.g., image) into bits, and a source decoder plays the role of a classifier that reproduces the intrinsic state (e.g., label of image). In this work, we characterize the general structure of the optimal transition probability distribution, achieving the indirect rate–distortion function. This optimal solution can be interpreted as a “soft classifier”, which generalizes the conventionally adopted “classify-then-compress” scheme. We then apply the soft classification to aid the lossy compression of the extrinsic observation of a composite source. This leads to a coding scheme that exploits the soft classifier to guide reproduction, outperforming existing coding schemes without classification or with hard classification.

1. Introduction

In [1,2], information transmission with additional noise is studied. In this setting, the encoder has access only to a noise-corrupted version of the original signal, while the goal is to minimize the expected distortion between the original signal and the final output. Under the mean-squared error (MSE) distortion criterion, the optimality of an estimate-then-compress scheme is proved. This indirect compression model shares similarities with the semantic communication problem, which has recently attracted much attention. A common model for studying semantic communication is the composite source model [3], which consists of an intrinsic state and an extrinsic observation. The intrinsic state corresponds to the semantic aspect of information, while the encoder observes only the extrinsic observation. In classification tasks, the states are discrete, and a scheme similar to the one mentioned above is the hard-classify-then-compress (HCTC) scheme, i.e., first performing optimal classification based on the observation, then compressing the classification result.
As will be shown in this paper, in general, however, this HCTC scheme is suboptimal due to the inherent “ambiguity” of the states. In a typical composite source, some observations correspond to multiple states with similar posterior probabilities. In this case, losslessly transmitting the classification result offers limited performance gains. The HCTC scheme, however, compresses the classification results indifferently, without accounting for the “ambiguity” of the observations. In [4,5], a soft classification scheme is investigated in a binary classification model. Unlike the HCTC scheme, the soft classification scheme directly compresses the observation and leverages the “ambiguity” of the source. In this paper, the study is further conducted and generalized to multi-class classification.
The value of the knowledge of semantic information in reconstructing observations has also received much attention. In most cases, a codebook precisely matched to the source statistics is hard to obtain. In contrast, the Gaussian codebook is well studied and commonly used. In this work, the role of the classification results obtained through the soft classification scheme in aiding the reconstruction of observations is investigated. We propose a classification-aided-compression (CAC) scheme and study its rate–distortion properties. An achievable upper bound proposed in [6], where the codebook precisely matched to the source statistics is available, is used for comparison. Our work compares the rate–distortion performance achievable using Gaussian codebooks, assisted by different classifiers when the matched codebook is unavailable, with the performance achievable when the matched codebook is available. Numerical results show the benefit of soft classification in improving the rate–distortion performance when only Gaussian codebooks are available, especially at high rates.

1.1. Related Works

In his landmark work [7,8], Shannon excluded the semantic aspects of information from the framework of his classical rate–distortion theory, asserting that the semantic information is irrelevant to the engineering problem. This viewpoint that communication systems should focus on symbol transmission has since become a foundational paradigm in source coding. However, in many practical scenarios, the encoder has only indirect access to the target signal (e.g., through observations corrupted by channel noise or measurement errors). Such indirect source coding problems have been studied in [1,2], but largely limited to simple cases involving AWGN.
To enhance the performance of communication systems in such scenarios, there has been ongoing consideration of how to leverage the semantic aspects of information in communication [9,10,11,12]. These works aim to establish a universal, task-independent definition of semantic information. In recent years, with the development of 5G and the growing interest in post-5G technologies, some researchers have taken the semantic aspects of information into account when designing communication systems [13,14,15,16], to enable new applications in emerging scenarios. Compared to the former works, in these studies, the semantic aspects of information are associated with specific goals. Our work aligns with the latter. Task-oriented compression has been investigated in several works, with quantization being the primary method [17,18,19,20]. In such works, classification [21,22], detection [23], and inferring a latent variable from compressed data [24] are common tasks.
A commonly used model for studying semantic communication is the composite source model [3], which consists of an intrinsic state and multiple sub-sources. Different values of the intrinsic state correspond to different sub-sources and output symbol distributions. The source coding problem for composite sources is studied in [25]. The rate–distortion problems of composite sources under various distortion metrics and constraints are investigated in [4,26], where the composite source model is also connected to the semantic aspects of information. Such rate–distortion problems involving an intrinsic state or noisy observations are referred to as “indirect rate–distortion problems” [1,2,3,27,28]. Two cases have been studied in [4]. In the first case, the state and the observation obey a joint Gaussian distribution, which is further explored in [27]. The second case involves binary classification, and the corresponding solution implies a soft classification scheme. In our work, this case is extended to multi-class classification settings.
The concept of semantic information and the composite source model has also drawn attention in real-world signal processing tasks such as image and speech processing. In certain machine vision tasks, only specific semantic features are required, and the bitrate needed to represent these features is often much lower than that required to encode the entire image. In [29], a scalable image compression scheme is proposed, which simultaneously satisfies the requirements of both human vision and machine vision tasks. The composite source model has also been used to characterize image and speech signals [30,31,32,33]. A widely used model in speech signal processing, the Hidden Markov Model (HMM), is essentially a type of composite source model [34]. In image classification tasks, the label of an image can be viewed as the intrinsic state, with the image generated by its corresponding sub-source. The classification task thus becomes equivalent to inferring the intrinsic state from the image [35]. Other tasks such as recognition [36] and anomaly detection [37] can be modeled in a similar manner. Recently developed generative models, such as the variational auto-encoder [38] and the diffusion model [39], are also related to the composite source model. Taking the variational auto-encoder as an example, different latent variables correspond to different output distributions, similar to how different intrinsic states correspond to different observation distributions in the composite source model. Beyond signal processing, composite sources have also been applied to other practical applications. In [40], a multi-source or composite source information fusion method is studied in the context of the failure mode and effects analysis (FEMA) problem, where assessments from different experts are treated as different sources. The model is also well-suited for capturing ambiguity in classification tasks, where observations do not correspond to unique labels but instead map to multiple possible states with different probabilities. In such scenarios, soft classification is employed [41,42,43].
Despite the numerous studies on semantic information and composite sources, several important issues remain to be addressed. For the composite source classification, studies such as [4,5] have only investigated the case of symmetric binary classification. Nevertheless, many practical tasks involve more general scenarios, such as multi-class classification. In these cases, the soft classification scheme that achieves the classification rate–distortion function is difficult to obtain. In [44], the authors propose a classification-based approach to reconstruct sparse sources, indicating the value of classification when the codebook matched to the source statistics is unavailable. However, only hard classification is considered therein. To the best of our knowledge, no existing work has considered the tradeoff between the classification rate and the effectiveness of classification in reducing the reconstruction rate under this kind of scenarios.

1.2. Contribution and Organization of Paper

Our main contributions include:
  • We characterize the classification rate–distortion function for general multi-class composite sources under classification distortion constraint only.
  • Based on this setting, we study two schemes: the HCTC scheme and the soft classification scheme. We convert the rate–distortion performance of the HCTC scheme into the rate–distortion function of a discrete source. We also identify sufficient conditions for the rate–distortion optimality of both schemes and analyze the rate–distortion properties of the soft classification scheme. Our analysis shows that, by allowing only a small additional distortion compared to the minimum achievable distortion, the upper bound of the rate of the soft classification scheme can decrease rapidly. Through numerical results, we compare the performance of the two schemes and show that each has different strengths and weaknesses under different scenarios.
  • When the reconstruction distortion is constrained, we derive an achievable upper bound of the reconstruction rate–distortion function and propose the CAC scheme using only Gaussian codebooks. Numerical results show that, with proper classification before compression, the rate–distortion performance of the CAC scheme can approach that of the scheme with a matched codebook. We also find that, under high-resolution conditions, the total bitrate of the CAC scheme can be minimized by separately optimizing the classifier and each sub-encoder.
The remainder of the paper is organized as follows.
Section 2 introduces and describes the composite source model and formulates the problem as an indirect rate–distortion problem. In Section 3, the classification rate–distortion function is characterized. We further study two achievable upper bounds of the rate–distortion function and, based on these, propose two classification schemes. Section 4 derives an upper bound of the rate–distortion function under reconstruction distortion constraint, and based on this, proposes the CAC scheme. Section 5 presents the numerical results. For cases with only the classification distortion constraint, we compare the performance of the two schemes. For cases with the reconstruction distortion constraint, we evaluate the performance of the proposed CAC scheme under classifiers of different levels. Finally, Section 6 concludes the paper.

2. Problem Formulation

The composite source consists of two parts: the intrinsic state S and the extrinsic observation X. Within the composite source, different states correspond to different sub-sources of observation X, each with a distinct distribution.
In the composite source coding, the encoder and decoder do not have access to the state S, while the observation X is available to the encoder [4]. The state can only be inferred from the extrinsic observation. Assume that the source has L states, with prior probabilities P 1 , , P L .
For state S, the conditional distribution of the observation X is denoted as P ( X | S ) . In the following work, we assume that the state S is a discrete label. Let X and S denote the alphabets of X and S, respectively. The conditional probability density function (pdf) corresponding to the t-th state is denoted as f t ( x ) . The pdf of the entire source is denoted by f ( x ) , given by,
f ( x ) = t = 1 L P t f t ( x ) .
Let the block length be denoted by n. When encoding a sequence of source observations X n , it is first mapped to an index W in the codebook. Here W takes value from { 1 , 2 , , 2 n R } , where R is the rate. On the decoder side, W is mapped to either the reconstruction of the states (classification result) S ^ or the reconstruction of the observation X ^ . The model for this coding problem is illustrated in Figure 1. In all of the schemes described in this work, we assume n + .
In different scenarios, the primary concern varies. In some cases, the emphasis is on classification distortion [21,22]. In some other cases, reconstruction distortion is the focus [44]. Consider two distortion measures: classification distortion measure d s and the reconstruction distortion measure d o . Let D s and D o denote the distortion constraints on classification and reconstruction, respectively. The rate–distortion problem can then be formulated as an optimization problem that minimizes the mutual information between X and S ^ (or X ^ ), subject to the corresponding distortion constraint [4]:
R ( D s ) = min I ( X ; S ^ ) , s . t . E { d ^ s ( X , S ^ ) } D s ,
or
R ( D o ) = min I ( X ; X ^ ) , s . t . E { d o ( X , X ^ ) } D o ,
where S , X , S ^ , and X ^ form a Markov chain S X S ^ / X ^ and d ^ s ( x , s ^ ) = E { d s ( S , s ^ ) | X = x } . Equation (2) describes the rate–distortion formulation for tasks concerned solely with classification distortion, while (3) corresponds to scenarios where reconstruction distortion is the main focus. In this work, we adopt the Hamming distortion metric to characterize classification distortion, i.e., d S ( s , s ^ ) = 1 ( s s ^ ) and use the MSE to measure the reconstruction distortion.
In the following sections, the rate–distortion function is referred to as the classification rate–distortion function when only classification distortion is constrained, and as the reconstruction rate–distortion function when only reconstruction distortion is constrained.

3. Rate–Distortion Analysis for Classification in Composite Sources

3.1. Classification Rate–Distortion Function

In this section, we consider the case where only the classification distortion is constrained. The corresponding optimization problem is formulated by (2).
Let the transition probability from X = x to S ^ = t be denoted as g t ( x ) , i.e., g t ( x ) = P ( S ^ = t | X = x ) . The expected distortion can be expressed as
E { d ^ S ( X , S ^ ) } = E { d S ( S , S ^ ) } = t = 1 L P t E { d S ( t , S ^ ) | S = t } = t = 1 L P t X f t ( x ) u = 1 L g u ( x ) 1 ( t u ) d x = 1 t = 1 L P t X f t ( x ) u = 1 L g u ( x ) 1 ( t = u ) d x = 1 X t = 1 L P t f t ( x ) g t ( x ) d x ,
while the rate (mutual information) is given by [45]
I ( X ; S ^ ) = H ( S ^ ) H ( S ^ | X ) .
The two terms H ( S ^ ) and H ( S ^ | X ) in the equation above can also be expressed with g t ( x ) . Therefore, the optimization problem can be solved by optimizing the transition probability g t ( x ) . As a result, the classification rate–distortion function can be characterized based on the optimization result.
Theorem 1. 
The classification rate–distortion function of a composite source is given by
R ( D s ) = t = 1 L X f ( x ) g t ( x ) log 2 g t ( x ) X f ( x ) g t ( x ) d x d x ,
where
g t ( x ) = P ( S ^ = t ) e λ P t f t ( x ) f ( x ) w = 1 L P ( S ^ = w ) e λ P w f w ( x ) f ( x ) ,
P ( S ^ = t ) = X f ( x ) g t ( x ) d x ,
and the parameter λ is chosen such that
1 X t = 1 L P t f t ( x ) g t ( x ) d x = D s .
Proof. 
See Appendix A. □
Remark 1. 
In Theorem 1, g t ( x ) represents the transition probability P ( S ^ = t | X = x ) . The theorem shows that, for a fixed X = x , the transition probability from X = x to S ^ = t satisfies
g t ( x ) P ( S ^ = t ) e λ P ( S = t | X = x ) .
The result aligns with the intuition that, for a fixed X = x , the transition probability g t ( x ) increases with the posterior probability P ( S = t | X = x ) .
The form of g t ( x ) implies a soft classification scheme. In this scheme, as λ increases, g t ( x ) approaches a step function, indicating that the classification becomes “harder”. At the same time, the rate increases and the distortion decreases. When λ + , we have
g t ( x ) 1 t = t ( x ) 0 o t h e r w i s e ,
for all x where the most probable state t ( x ) = arg max t P ( S = t X = x ) is unique. In this case, the classification reduces to hard classification. Thus, at the far left of the RD curve, the rate approaches the entropy of the hard classification results.
Conversely, as λ decreases towards 0, the functions g t ( x ) for t = [ 1 : L ] become smoother, making the classification more “random”. As a result, the distortion increases while the rate decreases. By adjusting λ, the functions g t ( x ) can be tuned to satisfy the distortion constraint.
Equation (7) also involves the marginal probability of S ^ , assigning larger weights to the states with higher marginal probabilities. This helps reduce the first term H ( S ^ ) in (5).

3.2. Two Classification Schemes

Next, two achievable upper bounds of the classification rate–distortion function are introduced and analyzed, along with the corresponding classification–compression schemes.

3.2.1. Hard-Classify-Then-Compress

The first scheme is called the hard-classify-then-compress (HCTC) scheme. In this scheme, a hard classification is first performed on the observation X, and the resulting classification outcome is denoted as S ¯ . This hard classification result S ¯ is then compressed into S ^ .
The procedure of the HCTC scheme is illustrated in Figure 2. As shown in the figure, the following Markov chain is formed:
S X S ¯ S ^ .
Define
d ˜ s ( s ¯ , s ^ ) = E S { d s ( S , s ^ ) | S ¯ = s ¯ } = E S { 1 ( S s ^ ) | S ¯ = s ¯ } = 1 E S { 1 ( S = s ^ ) | S ¯ = s ¯ } = 1 P ( S = s ^ | S ¯ = s ¯ ) .
Since the joint distribution of S and S ¯ is known, d ˜ ( s ¯ , s ^ ) can be calculated for each pair ( S ¯ , S ^ ) .
Theorem 2. 
The rate–distortion behavior of the HCTC scheme is equivalent to the rate–distortion function of the hard classification result S ¯ using the distortion metric defined in (13).
Proof. 
See Appendix B. □
Remark 2. 
Theorem 2 shows that the rate–distortion performance of the HCTC scheme can be reduced to the rate–distortion function of a discrete source, enabling numerical computation using the Blahut–Arimoto (BA) algorithm [46] and its variants such as [47].
Corollary 1. 
A sufficient condition for the optimality of the HCTC scheme is as follows: x 1 , x 2 X u ,   t = [ 1 : L ] , P ( S = t | X = x 1 ) = P ( S = t | X = x 2 ) , where X u = { x | P ( S = u | X = x ) P ( S = i | X = x ) , i = [ 1 : L ] } .
Proof. 
See Appendix C. □
Remark 3. 
The condition in Corollary 1 implies that all the observations X leading to the same hard classification result convey identical information about the intrinsic state S. Consequently, compressing the hard classification result S ¯ introduces no additional distortion compared to directly compressing the original observation X.

3.2.2. Symmetric Cases and Soft Classification

Based on Theorem 1, the classification rate–distortion function can be computed by alternatively optimizing g t ( x ) and P ( S ^ = t ) , t = [ 1 : L ] , in a manner similar to the BA algorithm [48].
When λ > 0 , if all P ( S ^ = t ) , t = 1 , 2 , are initialized as 1 L and remain unchanged after the first iteration, then in all subsequent iterations, we have
g t ( x ) ( i ) = P ( S ^ = t ) ( i 1 ) e λ P t f t ( x ) f ( x ) w = 1 L P ( S ^ = w ) ( i 1 ) e λ P w f w ( x ) f ( x ) P ( S ^ = t ) ( i ) = 1 L , t [ 1 : L ] ,
thus achieving convergence. For such composite sources, the optimal transition functions are given by
g t ( x ) = e λ P t f t ( x ) f ( x ) w = 1 L e λ P w f w ( x ) f ( x ) , t [ 1 : L ] .
Equation (15) shows that g t ( x ) takes the form of a softmax function with coefficient λ applied to the conditional probability P ( S = t | X = x ) . Unlike (7), this expression does not involve the marginal probability of S ^ . Consequently, g t ( x ) can be directly derived from the joint distribution of S and X using (15), without the need to solve any nonlinear equations. This significantly simplifies the computation.
Lemma 1. 
The optimal transition probability functions take the form of (15), if and only if
X f ( x ) e λ P t f t ( x ) f ( x ) w = 1 L P e λ P w f w ( x ) f ( x ) d x = 1 L , t [ 1 : L ] .
Proposition 1. 
One type of composite source that satisfies the condition in Lemma 1 is as follows (assume the source has L states):
  • The number of states is finite, and the observation space satisfies X R K .
  • The probability distributions of the sub-sources are distinct.
  • P t = 1 L , t [ 1 : L ] .
  • Let μ t denote the mean of the t-th sub-source. There exists a point μ 0 such that μ t μ 0 2 2 is equal for all t. For simplicity and without loss of generality, assume μ 0 = 0 in the following discussion.
  • For any two sub-sources indexed by a and b, there exists an orthonormal matrix H such that
    (1) 
    μ b = H μ a .
    (2) 
    t [ 1 : L ] , u [ 1 : L ] , s . t .
    f u ( H x ) = f t ( x ) , x X ,
    specifically,
    f b ( H x ) = f a ( x ) , x X
    (3) 
    If x X , H x X .
Proof. 
See Appendix D. □
Unfortunately, many practical scenarios do not satisfy the conditions in Proposition 1. Nevertheless, (15) can still be employed to obtain an achievable upper bound of the classification rate–distortion function in cases where the rate–distortion function itself is difficult to solve. Moreover, (15) induces a simplified soft classification scheme, in which the conditional probabilities P ( S ^ = t | X = x ) follow the transition functions g t ( x ) in (15). In the remainder of this paper, the soft classification scheme refers to this simplified version.
Assume a length-n sequence of observation X n (with n sufficiently large). The steps of the soft-classification are as follows:
  • Calculate the transition probabilities from observation X to the classification result S ^ using (20).
  • Calculate the marginal distribution P ( S ^ ) and the mutual information I ( X ; S ^ ) .
  • R > I ( X ; S ^ ) , randomly generate a codebook C containing 2 n R i.i.d. sequences S ^ n drawn according to P ( S ^ ) . Each sequence is a codeword, indexed by W { 1 , 2 , , 2 n R } .
  • When encoding, select the codeword that is distortion typical [45] with X n . If there is more than one such S ^ n , choose the one with the smallest index W . If no such codeword exists for a given X n , encode it using W = 1 .
  • At the decoder, recover the sequence S ^ n from the received index W using the codebook C . Due to the properties of the distortion-typical set, the rate–distortion pair ( R , D ) is achievable for any R > I ( X ; S ^ ) .
Theorem 3. 
An achievable upper bound on the classification rate–distortion function is given by
R ( D s ) I ( g ˜ t ) log L + X f ( x ) t = 1 L g ˜ t ( x ) log 2 g ˜ t ( x ) d x ,
where
g ˜ t ( x ) = e s P t f t ( x ) f ( x ) w = 1 L e s P w f w ( x ) f ( x ) ,
and s is set to satisfy
1 X t = 1 L P t f t ( x ) g ˜ t ( x ) d x = D s .
I ( g ˜ t ) = t = 1 L X f ( x ) g ˜ t ( x ) log 2 g ˜ t ( x ) f ( x ) g ˜ t ( x ) d x d x
is the rate required by this scheme. Its upper bound, denoted as R u b ( g ˜ t ) , is given in the last line of (19).
Proposition 2. 
The slope of the upper bound R u b ( g ˜ t ) is s log 2 e , and s varies from 0 to + , with the classification changing from “soft” to “hard”. As s + , x with unique t , the classification becomes the hard classification, corresponding to the optimal Bayesian classification. Conversely, When s = 0 , the classifier performs classification completely randomly.
Proof. 
See Appendix E. □
Remark 4. 
Denote the rate of the soft classifier as R s . Let R s ( s ) and D s ( s ) denote the rate and distortion, respectively, of the classifier parameterized by a fixed s. The corresponding upper bound on the rate is denoted as R u b ( s ) .
Proposition 2 can be expressed as
d R u b d D s = s log 2 e .
Consider the upper bound R u b . Let D s , m i n denote the minimum achievable distortion, attained by the Bayesian optimal classifier, (i.e., the hard classification). Now suppose a small additional distortion Δ D beyond D s , m i n is acceptable. In this scenario, since the classifier is almost a hard one, we have s + . When the distortion increases by Δ D , R u b can be reduced by δ R u b s log 2 e Δ D . This indicates that even a slight relaxation in the distortion constraint can significantly reduce the rate upper bound of the soft classification scheme.
Equation (20) gives the form of transition probability from the observation X to the classification result S ^ . Consider the following example of a four-class classification problem. Assume a composite source with state S and observation X. When S = t , X N ( μ t , I ) , where I is the identity matrix and μ 1 = ( 1 , 0 ) , μ 2 = ( 0 , 1 ) , μ 3 = ( 1 , 0 ) , μ 4 = ( 0 , 1 ) . Figure 3a–c illustrate g ˜ 1 ( x ) for s = 0 , 1 , and 100, respectively.
When s = 0 , the transition probability g ˜ 1 ( x ) is 0.25 for all X = x , indicating completely random classification. Comparing the cases where s = 1 and s = 100 , the latter case is very close to the hard classification since s is very large. As shown in Figure 3c, the transition probability function is almost 0-1 binary.
In the HCTC scheme, the classification results are compressed without considering the posterior probabilities P ( S = t | X = x ) for different x. Such compression can cause significant distortion since the classification results of X with “clear” states are also compressed. In contrast, in the soft classification scheme (Figure 3b), for X = x with “ambiguous” states (e.g., the point ( 0 , 0 ) ), H ( S ^ | X = x ) is significantly greater than 0. Consequently, the rate allocated to these classification results is much lower than that allocated to X with more certain states. Therefore, compression mainly occurs in these “ambiguous” regions. In these regions, the optimal classification already introduces non-negligible distortion. As a result, saving bits for these observations does not lead to significant additional distortion.
When the classification distortion constraint is tight, the rate that can be saved is limited. Therefore, the soft classification scheme, by concentrating bit savings primarily on X with “ambiguous” states, can achieve a smaller distortion than the HCTC scheme.
As shown in Corollary 1, when the “ambiguity” of observations with the same hard classification result is identical, the performance of the HCTC scheme is no worse than that of the soft classification scheme. When the prior probabilities of the states are different and the classification distortion constraint is loose, the performance of the HCTC scheme can even surpass that of the soft classification. In the soft classification scheme, as s 0 , g ˜ t ( x ) 1 L , t = [ 1 : L ] , x X . Thus, when R = 0 , D = 1 1 L , the classification becomes completely random, regardless of the prior distribution of the states. In contrast, the HCTC scheme classifies all observations into the class with the highest prior probability, resulting in a smaller distortion compared to the soft classification scheme.
It is also noteworthy that, unlike deep neural networks with the output of the softmax results, our soft classification scheme does not output the posterior distribution of the states. In practical applications, the observations are first compressed into an index W . Then the classification results are decoded based on W . Therefore, the soft classification scheme directly outputs a reconstruction of the state. Its difference from the hard classification is that, when the rate is constrained, by incorporating the source statistics and applying stronger compression to the “ambiguous” observations, the soft classification can reduce the expected distortion.

4. Classification-Aided Reconstruction of Composite Sources

In this section, we consider the case where the reconstruction distortion is constrained. The metric of the reconstruction distortion is the MSE distortion.
Consider an achievable upper bound of R ( D o ) . This problem can be formulated as the optimization problem (3). One approach is to perform direct lossy compression on the observation. Nevertheless, for most sources, the matched codebook is difficult to obtain. In contrast, for Gaussian sources, which are widely adopted in signal processing tasks, the codebook is easy to obtain. Therefore, in such cases, the codebook of a Gaussian source with the same variance (covariance matrix), can be used as a substitute. However, some sources have large variances, leading to poor rate–distortion performance when compressed directly. In [44], the magnitude classifying quantization (MCQ) scheme is proposed. In this scheme, the observation is first classified according to its amplitude. Then the observations are compressed by sub-encoders corresponding to the classification result. The study shows that classification changes the variance of the input to each sub-encoder. Hence when using Gaussian codebooks, although an additional rate is introduced for classification, the overall rate of the MCQ scheme can still be lower than that of direct compression in some cases.
In this section, the soft classification introduced in Section 3.2.2 is adopted, and the classification-aided-compression (CAC) scheme is proposed. Based on this scheme, an achievable upper bound of R ( D o ) is also derived.
Another upper bound of R ( D o ) is as follows:
R ( D o ) I ( X ; X ^ ) I ( X ; S ^ , X ^ ) = I ( X ; S ^ ) + I ( X ; X ^ | S ^ ) .
Inequality (24) implies the CAC scheme to encode the observation of an L-state composite source X under the reconstruction distortion D o . The first term represents classification, while the second term represents compression using the corresponding encoder according to the classification result.
The specific steps of the scheme are as follows:
  • Select a distortion level D and perform the soft classification. Denote the result as S ^ .
  • Encode X using the corresponding Gaussian encoder based on the soft classification result.
Figure 4 illustrates the steps of the CAC scheme.
An upper bound of R ( D o ) can be derived from the scheme described above.
Theorem 4. 
An achievable upper bound of R ( D o ) is
R ( D o ) min D [ D s , m i n , D s , m a x ] { R s ( D ) + u = 1 L P ( S ^ = u ) R X ˜ u , g } ,
where D s , m i n and D s , m a x are the minimum and maximum classification distortions that can be achieved by the soft classifier, R s ( D ) is the rate at the classification distortion D when using the soft classification scheme in Section 3.2.2, P ( S ^ = u ) is obtained from (8), and
R X ˜ u , g = 1 2 i = 1 K log 2 σ u i 2 D u i .
In (26), K denotes the dimension of X. σ u i 2 is the i-th eigenvalue of Σ u , where
Σ u = X f ( x ) g D , u ( x ) ( x γ u ) ( x γ u ) T d x X f ( x ) g D , u ( x ) d x ,
γ u = X x f ( x ) g D , u ( x ) d x X f ( x ) g D , u ( x ) d x ,
and g D , u ( x ) is the transition probability function of the soft classification scheme that satisfies (21) with the right-hand side replaced by D. For D u i , we have
D u i = min { α , σ u i 2 } ,
where α is chosen to satisfy
u = 1 L i = 1 K P ( S ^ = u ) D u i = D o .
Proof. 
See Appendix F. □
For the CAC scheme, in the compression step, X is compressed using a Gaussian encoder. Define
R u b ( D o ) R s ( D ) + u = 1 L P ( S ^ = u ) R X ˜ u , g ,
where D [ D s , m i n , D s , m a x ] . We have
R u b ( D o ) = R s ( D ) + 1 2 u = 1 L P ( S ^ = u ) i = 1 K log 2 σ u i 2 D u i = R s ( D ) + 1 2 u = 1 L P ( S ^ = u ) log 2 ( det ( Σ u ) ) 1 2 u = 1 L i = 1 K P ( S ^ = u ) log 2 D u i .
Consider two special cases where s = 0 and s + . When s = 0 , (20) becomes g ˜ t = 1 / L . Thus P ( S ^ = u ) = 1 / L , u = [ 1 : L ] . Hence u = [ 1 : L ] , the pdf of X given the condition that S ^ = u is
f ( x | S ^ = u ) = P ( X = x , S ^ = u ) P ( S ^ = u ) = f ( x ) 1 L 1 L = f ( x ) .
From (33), we can see that when s = 0 , the input to each sub-encoder obeys the same distribution as the composite source. It is also evident that, in this case, the rate required for the classification part is 0. Hence when s = 0 , the CAC scheme reduces to the direct compression. As s + , the classification becomes hard classification. In this case, the scheme is similar to the MCQ method proposed in [44], where a pre-compression classification based on amplitude is performed.
In the compression step, X is first subtracted by the mean of the corresponding sub-encoder input, resulting in a zero-mean vector for encoding. Then, orthogonal transformations are applied separately to the vectors to be encoded by each sub-encoder, ensuring their components are mutually uncorrelated. After the transformation, the distortion allocation D u i to each component of different classification results follows the classical reverse water-filling strategy. The only difference is that each D u i is weighted by its corresponding probability P ( S ^ = u ) .
When D o K σ u i 2 , u = [ 1 : L ] , i = [ 1 : K ] , according to (29), all D u i are equal. Hence, D u i = α and (32) can be rewritten as
R u b ( D o ) = R s ( D ) + 1 2 u = 1 L P ( S ^ = u ) log 2 ( det ( Σ u ) ) 1 2 u = 1 L i = 1 K P ( S ^ = u ) log 2 α = R s ( D ) + 1 2 u = 1 L P ( S ^ = u ) log 2 ( det ( Σ u ) ) K 2 log 2 α .
In (34), the third term depends solely on α . According to (30), we have
D u i = α = D o K ,
Hence the third term in (34) is solely determined by D o . Denote D T = K σ 2 , where σ 2 denotes the minimum eigenvalue among all Σ u across classifiers with s 0 (with s being the coefficient in (20) that controls the softness of the classifier). It can also be observed that in (34), the first two terms depend only on the classification operation and are independent of D o . Thus, when D o D T , the effects of the classifier and D o on R u b can be completely decoupled, which is convenient for further analysis. Therefore, the optimal classifier can be obtained by optimizing s to minimize the first two terms when D o D T .
In the high-resolution case, D o is very small, thus satisfying the condition above. Denote the first two terms in (34) as R C ( s ) , i.e., R C ( s ) = R s ( D ) + 1 2 u = 1 L P ( S ^ = u ) log 2 ( det ( Σ u ) ) . Let D o ( s , R ) denote the reconstruction distortion achieved by the CAC scheme when the coefficient of the classifier is s and the total rate is R. From (34), we have
D o ( s , R ) = K 2 2 K ( R C ( s ) R ) = K 2 2 K R C ( s ) 2 2 K R .
Hence in the high-resolution case, the achievable distortion of this scheme exhibits an exponential decay with respect to the rate R. By selecting a proper classifier, the coefficient of the function can be minimized, thus minimizing the reconstruction distortion.
As the reconstruction distortion constraint becomes more relaxed, the reverse water-filling scheme allows for completely omitting the bits used to reconstruct components corresponding to small eigenvalues in the covariance matrices of each classification result.

5. Numerical Result

In this section, we present and compare numerical results of the schemes developed in the previous sections.

5.1. Soft Classification and HCTC

This section presents and compares the performance of the schemes proposed in Section 3.2 under various settings. The procedures of the two schemes are detailed in Section 3.2.1 and Section 3.2.2, respectively.
First, we consider the Gaussian mixture (GM) source model with equal prior probabilities for all states. The distributions of the composite sources in the first seven settings are given as follows. For each setting, let μ t denote the mean of the t-th state, and Σ t its covariance matrix.
In the first two settings, the sources are 4-state GM sources. In both settings, μ 1 = ( 1 , 0 ) , μ 2 = ( 0 , 1 ) , μ 3 = ( 1 , 0 ) , μ 4 = ( 0 , 1 ) . In the first setting, Σ t = I 2 × 2 , t = 1 , 2 , 3 , 4 . In the second setting, Σ t = 2 I 2 × 2 , t = 1 , 2 , 3 , 4 .
In the next two settings, the sources remain 4-state GM sources, and the covariance matrices of all states are again identity matrices. In the third setting, the mean points of the states are μ 1 = ( 1 , 1 ) , μ 2 = ( 1 , 1 ) , μ 3 = ( 1 , 1 ) , μ 4 = ( 1 , 1 ) . In the fourth setting, the mean points of the states are μ 1 = ( 3 , 3 ) , μ 2 = ( 3 , 3 ) , μ 3 = ( 3 , 3 ) , μ 4 = ( 3 , 3 ) .
These four settings perfectly satisfy the conditions in Proposition 1. Hence in these settings, the rate–distortion function R ( D s ) attains the value of the upper bound R u b ( g ˜ t ) .
In the fifth setting, the number of states is increased to 9, with the mean points of the different states illustrated in Figure 5a. Furthermore, the number of states is further increased to 16 and 32 as shown in Figure 5b and Figure 5c, respectively. In these three settings, the covariance matrix of each state remains the identity matrix. The performance of the soft classification scheme and the HCTC scheme is compared under these seven settings as depicted in Figure 6.
As shown in Figure 6, the soft classification scheme consistently outperforms the HCTC scheme in all seven settings. In some cases, such as the sources depicted in Figure 6e–g, the conditions in Proposition 1 are not satisfied, indicating that the transition probability functions in (15) are not optimal. In Figure 6e,f, the performance of both schemes is compared with the rate–distortion function obtained from the constrained BA (CBA) algorithm [47]. In these cases, even with the suboptimal transition probability functions (20), the soft classification scheme still outperforms the HCTC scheme, and its performance is close to the rate–distortion function, especially when the distortion is near the minimum. As analyzed in Section 3.2.2, the soft classification scheme can better exploit the “ambiguity” in a composite source, thus reducing the rate without significantly increasing the distortion.
It is also noteworthy that in some sources, the gap between the two schemes is significant, while in some others, it is not. Figure 7 shows how the ratio of the bits required for the soft classification and the HCTC schemes changes as the extra classification distortion allowed beyond the Bayesian minimum classification distortion increases in different sources. In this figure, D s denotes the distortion constraint, and D s , m i n denotes the Bayesian minimum distortion. The classification distortion constraint D s takes the value from D s , m i n to 0.5 for all sources. The second setting is not shown in this figure since in this setting, the Bayesian minimum distortion is greater than 0.5.
As shown in Figure 7, the ratio of the bits required by the two schemes decreases dramatically as the distortion constraint D s increases slightly compared to the Bayesian minimum distortion in all sources, except for the fourth setting. In the fourth setting, the distances between the states are large relative to their variances, resulting in a very small probability of “ambiguity” for the state corresponding to the observation. As mentioned in Section 3.2.2, the soft classification scheme primarily exploits the “ambiguity” of the states. Hence in this setting, there is no significant difference between the two schemes.
Now, we compare the first, third, and fourth settings. All three settings have four states, and the covariance matrices of the states are identity matrices. As shown in Figure 7, in the sources where the states are closer to each other, the soft classification scheme achieves more significant bit savings. This still holds when comparing the fourth, fifth, and sixth settings, where the number of states varies. In these three settings, the outermost mean points are located at the same positions. Thus, as the number of states increases, the distances between the states decrease, and the performance gap between the two schemes becomes larger.
An interesting finding emerges from the comparison among the third, sixth, and seventh settings. Although the minimum distances between the mean points are the same across these three sources, the ratio of the bits required by the two schemes approaches one as the number of states increases. A straightforward explanation is that adding more states in this manner does not introduce significant additional “ambiguity” to the source, since, for most observations, only a few (at most 4) states have non-negligible posterior probabilities. Nevertheless, as the number of states increases, both schemes require more bits. Thus the relative gap between the two schemes narrows.
In some other cases, however, the degree of “ambiguity” is consistent across all parts of the source. In such scenarios, the advantages of soft classification become less evident. Consider a 4-class composite source where the observation X is also discrete when S = i ( i = [ 1 : 4 ] ) , P ( X = i ) = 0.7 , P ( X = j ) = 0.1 , j = [ 1 : 4 ] , j i .
Figure 8 shows the rate–distortion performance of the HCTC scheme and soft classification scheme in this discrete composite source. Figure 8a illustrates the cases where the prior probabilities of the states are all equal. The result shows that the performance of the two schemes is identical. Figure 8b shows the case where the prior probabilities of the states are unequal, with P ( S = 1 ) = 0.4 , P ( S = 2 ) = P ( S = 3 ) = P ( S = 4 ) = 0.2 . In this setting, the performance of the soft classification scheme is inferior to that of the HCTC scheme, which aligns with the analysis presented in Section 3.2.2.
In Figure 9, the performance of the soft classification scheme, the HCTC scheme, and the rate–distortion function (obtained via the CBA algorithm) is compared in a four-component composite source, where the sub-sources are the same as those in the first setting. The difference lies in the prior probabilities of the states: P ( S = 1 ) = 0.4 , P ( S = 2 ) = 0.2 , P ( S = 3 ) = 0.2 , P ( S = 4 ) = 0.2 . This setting clearly illustrates the strengths and weaknesses of the two schemes. When the distortion constraint is very loose, the HCTC scheme performs better since the soft classification tends to classify observations with an almost uniform distribution, while the HCTC scheme tends to classify the observations into the state with higher prior probability, leading to lower distortion. However, when the distortion constraint is tight, the soft classification outperforms the HCTC scheme by better saving bits in the “ambiguous” regions without significantly increasing the distortion.

5.2. Upper Bound of Reconstruction Rate–Distortion Function

In this part, the achievable upper bounds of R ( D o ) and their properties are shown in five cases.
The first setting is the same as the four-class Gaussian mixture source in Section 5.1. In the second setting, the distances between different sub-sources are increased, with μ 1 = ( 3 , 0 ) , μ 2 = ( 0 , 3 ) , μ 3 = ( 3 , 0 ) , μ 4 = ( 0 , 3 ) . The covariance matrices of all sub-sources remain identity matrices. The third setting considers a two-component composite source, where X is a scalar. Specifically, X N ( 0 , 1 ) when S = 1 , and X N ( 0 , 25 ) when S = 2 . The fourth setting is also a one-dimensional two-component GM source, where X N ( 1 , 1 ) when S = 1 , and X N ( 1 , 25 ) when S = 2 . The fifth setting involves a four-component two-dimensional GM source. When S = t , X N ( 0 , Σ t ) . We have
Σ 1 = 1 0 0 1 , Σ 2 = 1 0 0 25 , Σ 3 = 25 0 0 1 , Σ 4 = 5 0 0 5 .
In all of these settings, the distribution of the state S is uniform.
As mentioned in Section 4, the influence of D s and D o on R u b can be decoupled, and the optimal classifier can be obtained by minimizing R C ( s ) when the reconstruction distortion constraint satisfies D o D T . The function R C ( s ) in the first three settings is shown in Figure 10.
As s increases, the classification rate R s ( D ) also increases. Nevertheless, changing s also changes the conditional covariance matrices Σ u corresponding to each classification result. Thus, a tradeoff arises between the classification rate and the determinants of these covariance matrices.
In the first setting, the distributions of different states are close to each other, leading to a small determinant for the overall covariance matrix. Consequently, classification does not significantly reduce the determinants of the conditional covariance matrices. On the contrary, it introduces additional code length. Since the classification rate dominates in R C ( s ) , the value of R C ( s ) increases as the classification becomes harder.
In other cases where the determinant of the initial covariance matrix is large, such as in the second setting, applying a harder classification before compression can make the conditional distributions of the observation corresponding to each classification result more “concentrated”, thus reducing the determinants. Therefore, harder classification can help save bits as illustrated in Figure 10b.
In some cases, however, proper soft classification can help save bits as shown in Figure 10c. Specifically, in the third setting, it can be explained as follows. The weighted average of the input variances of the two sub-encoders, weighted by the probabilities of the classification results, remains unchanged after classification. Since the logarithmic function is concave, the corresponding weighted average of the log-variances of the sub-encoders after classification is smaller than that of the entire composite source, thereby reducing R C ( s ) . Nevertheless, when hard classification is applied, the weighted log-variance cannot be further reduced significantly, while the classification rate R s increases.
The curves of the achievable upper bounds R u b ( D o ) in the first two settings are shown in Figure 11. The performance is compared with the achievable upper bound R 1 ( D o ) of the rate–distortion function proposed in [6]. The bound R 1 ( D o ) characterizes the achievable rate–distortion pairs when a codebook matched to the source statistics is available and is asymptotically tight at high rates. In addition, the CAC scheme with a hard classifier is similar to the MCQ scheme in [44].
The results align with the R C ( s ) curve in Figure 10. At high rates, classifiers with smaller R C ( s ) values lead to lower distortion. In the first setting, direct compression performs better, while in the second setting, the CAC scheme with a hard classifier achieves better performance. In Figure 11b, the CAC scheme with a properly chosen classifier performs closely to the R 1 ( D o ) bound at high rates. This indicates that when the codebook matched to the source statistics is unavailable and only Gaussian codebooks can be used, classification before compression can reduce the variance and thus lower the rate in this setting. Moreover, the performance of this scheme can approach that of using a matched codebook.
For a fixed classifier, when D o > D T , a series of inflection points appear as D o increases. When D o > u = 1 L P ( S ^ = u ) t r ( Σ u ) , reconstructing every observation as the mean of the input to the corresponding sub-encoder satisfies the constraint. In this case, no coding is needed for reconstruction, and the overall code length equals the classification code length. Therefore, when D o is sufficiently large, a “softer” classifier can save bits. It can also be observed from Figure 11 that in these two settings, the “harder” the classifier, the smaller the values of D o corresponding to the turning points. This is because in both cases, a “harder” classifier results in smaller eigenvalues and trace values of the covariance matrices.
Figure 11b further shows that for some composite sources, the optimal classifier varies with the reconstruction distortion constraint. Denote the optimal s as s . At high rates, s = 10 , whereas when D o = 3.2 , s = 4 .
Next, the high-rate scenario is studied in the last three settings.
Figure 12 compares the distortion–rate performance of three schemes—direct compression, CAC with a soft classifier, and CAC with a hard classifier—along with D 1 ( R ) , the distortion–rate function corresponding to R 1 ( D o ) , under high rate conditions in the last three settings. It can be observed that, with a properly chosen soft classifier, the CAC scheme significantly outperforms the other two schemes in all settings. In the third setting, the CAC scheme with a soft classifier can achieve approximately a 15% reduction in distortion at the same rate compared to both direct compression and the CAC scheme with a hard classifier. In the fourth and fifth settings, the distortion reductions are approximately 21% and 25%, respectively. These numerical results are consistent with the analysis in Section 4, which shows that, under high-rate conditions and with a fixed classifier, the achievable distortion of the three schemes decays exponentially with the rate R. Under this condition, when the rate is the same for all three schemes, the ratio of the distortions remains constant, and this constant does not vary with the rate. Therefore, optimizing s in the classifier effectively minimizes the coefficient of the exponential function in (36), thereby minimizing the distortion. The comparison between the performance of the CAC scheme with a soft classifier and the D 1 ( R ) curve further demonstrates that, with a properly selected classifier and Gaussian codebooks, the CAC scheme can approach the rate–distortion performance of the scheme using codebook matched to the source statistics.

6. Conclusions

We have generalized the classification problem in [4,5] into the multi-class classification scenario. In the case where only the classification distortion is constrained, we have characterized the rate–distortion function and proposed two classification schemes based on two upper bounds. We have also studied the scenario where the reconstruction distortion is constrained and evaluated the performance of the CAC scheme. In our future work, we plan to explore methods for implementing soft classification or compression methods that asymptotically approach the information-theoretic performance limits. In addition, investigating data-driven methods for estimating the rate–distortion behaviors of composite sources under both constraints is also an interesting direction for future research.

Author Contributions

Conceptualization, W.Z.; Methodology, Y.C., W.Z. and J.L.; Software, Y.C. and J.L.; Validation, Y.C.; Formal Analysis, Y.C.; Investigation, Y.C. and W.Z.; Resources, W.Z.; Data Curation, Y.C. and J.L.; Writing—Original Draft Preparation, Y.C.; Writing—Review and Editing, W.Z. and Y.C.; Visualization, Y.C.; Supervision, W.Z.; Project Administration, W.Z.; Funding Acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported in part by the National Natural Science Foundation of China under Grant 62231022.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MSEMean-Squared Error
HCTCHard-Classify-Then-Compress
CACClassification-Aided-Compression
AWGNAdditive White Gaussian Noise
HMMHidden Markov Model
MCQMagnitude Classifying Quantization
GMGaussian Mixture

Appendix A. Proof of Theorem 1

When encoding X, the state S cannot be observed. So the classification process can be simplified as a coding problem from X to S ^ . For X = x and S ^ = t , the distortion d ^ s ( x , t ) is
d ^ s ( x , t ) = E { d s ( S , t ) | X = x } = 1 P t | x ,
where P t | x = P ( S = t | X = x ) . For convenience, such a shorthand will be used in the rest of this paper.
Hence the indirect rate–distortion problem is transformed into a direct one. The optimal transition function g t ( x ) can be derived from Section 4.2 of [3].
g t ( x ) = P ( S ^ = t ) e λ ( 1 P t | x ) w = 1 L P ( S ^ = w ) e λ ( 1 P w | x ) = P ( S ^ = t ) e λ P t | x w = 1 L P ( S ^ = w ) e λ P w | x = P ( S ^ = t ) e λ P t f t ( x ) f ( x ) w = 1 L P ( S ^ = w ) e λ P w f w ( x ) f ( x ) ,
where P ( S ^ = t ) is the marginal probability of S ^ = t :
P ( S ^ = t ) = X f ( x ) g t ( x ) d x .
The parameter λ is chosen to satisfy the constraint on classification distortion:
1 X t = 1 L P t f t ( x ) g t ( x ) d x = D s .
With the transition probability function, the rate (mutual information) can be obtained:
R ( D s ) = t = 1 L X f ( x ) g t ( x ) log 2 g t ( x ) X f ( x ) g t ( x ) d x d x .

Appendix B. Proof of Theorem 2

From the chain rule of entropy, we have
H ( S ¯ , S ^ | X ) = H ( S ¯ | X ) + H ( S ^ | S ¯ , X ) = H ( S ^ | X ) + H ( S ¯ | X , S ^ ) .
In the HCTC scheme, S ¯ is a function of X, which indicates that H ( S ¯ | X ) = 0 . Also we have H ( S ¯ | X , S ^ ) H ( S ¯ | X ) = 0 . Hence H ( S ¯ | X , S ^ ) = 0 .
From (12), we also have H ( S ^ | S ¯ , X ) = H ( S ^ | S ¯ ) . Hence, we have H ( S ^ | S ¯ ) = H ( S ^ | X ) . Therefore
I ( X ; S ^ ) = H ( S ^ ) H ( S ^ | X ) = H ( S ^ ) H ( S ^ | S ¯ ) = I ( S ^ ; S ¯ ) .
For the distortion, the classification distortion constraint is E { d s ( S , S ^ ) } D s . Since
E S ¯ , S ^ { d ˜ s ( S ¯ , S ^ ) } = E S ¯ , S ^ { E S { d s ( S , S ^ ) | S ¯ } } = E { d s ( S , S ^ ) } ,
the distortion constraint is equivalent to
E S ¯ , S ^ { d ˜ s ( S ¯ , S ^ ) } D s .
Hence the rate–distortion property of the HCTC scheme can be transformed into the rate–distortion problem of compressing the hard classification result S ¯ .

Appendix C. Proof of Corollary 1

For x X u , denote P t | x = P t | u .
P ( S = t | S ¯ = u ) = P ( S = t , S ¯ = u ) P ( S ¯ = u ) = P t X u f t ( x ) d x X u f ( x ) d x = X u f ( x ) P t | x d x X u f ( x ) d x = P t | u .
According to Theorem 2, we have
d ( S ^ = t , S ¯ = u ) = 1 P t | u .
As the rate–distortion performance of the HCTC scheme is the rate–distortion function of S ¯ , the transition probabilities from S ¯ to S ^ are the same as the conditional probabilities in the rate–distortion function of discrete source S ¯ . Denote the conditional probabilities P ( S ^ = t | S ¯ = u ) as g t ( u ) , and it satisfies [3]
g t ( u ) = P ( S ^ = t ) e λ P t | u w = 1 L P ( S ^ = w ) e λ P w | u P ( S ^ = t ) = u = 1 L P ( S ¯ = u ) g t ( u ) .
Now consider the transition probability functions g t ( x ) that achieves the rate–distortion function R ( D s ) . From Theorem 1, it satisfies
g t ( x ) = P ( S ^ = t ) e λ P t | x w = 1 L P ( S ^ = w ) e λ P w | x
P ( S ^ = t ) = X f ( x ) g t ( x ) d x .
Since x 1 , x 2 X u , the posterior probability distributions of the states are equal, we have g t ( x 1 ) = g t ( x 2 ) , t = [ 1 : L ] . Hence for x X u , g t ( x ) is a constant for each t = 1 , , L . Denote the constant as Q t u . (A1) can be rewritten as
Q t u = P ( S ^ = t ) e λ P t | u w = 1 L P ( S ^ = w ) e λ P w | u
(A2) can be rewritten as
P ( S ^ = t ) = X f ( x ) g t ( x ) d x = u = 1 L X u f ( x ) Q t u d x = u = 1 L Q t u X u f ( x ) d x = u = 1 L P ( S ¯ = u ) Q t u
So for x X u , g t ( x ) = g t ( u ) satisfies the conditions of achieving the rate–distortion function R ( D s ) . Hence the HCTC scheme has the optimal rate–distortion performance under this condition.

Appendix D. Proof of Proposition 1

In this section, we assume that μ 0 = 0 for simplicity without loss of generality.
First, assume that the sub-source t 1 and t 2 both satisfy
f t 1 ( x ) = f u ( H x ) f t 2 ( x ) = f u ( H x ) .
This contradicts the second condition. Hence the orthonormal transformation transformation in Proposition 1 is one-to-one.
For the pdf of the entire source f ( x ) , we have
f ( H x ) = 1 L u = 1 L f u ( H x ) = 1 L t = 1 L f t ( x ) = f ( x ) .
The second equation stands because of the one-to-one property of the transformation.
For the prior probability of classification result b, we have
P ( S ^ = b ) = X f ( x ) e λ P ( S = b | X = x ) w = 1 L e λ P ( S = w | X = x ) d x = X f ( H x ) e λ P ( S = b | X = H x ) w = 1 L e λ P ( S = w | X = H x ) d H x = X f ( x ) e λ P ( S = b | X = H x ) w = 1 L e λ P ( S = w | X = H x ) d H x = X f ( x ) e λ P ( S = b | X = H x ) w = 1 L e λ P ( S = w | X = H x ) d x .
The third equation stands because of (A4). The fourth equation stands because H is an orthonormal matrix. Further, we have
P ( S = b | X = H x ) = f b ( H x ) L f ( H x ) = f a ( x ) L f ( x ) = P ( S = a | X = x ) .
Since the transformation by left-multiplying with H is one-to-one between the states, we can obtain w = 1 L e λ P ( S = w | X = H x ) = w = 1 L e λ P ( S = w | X = x ) .
Hence for states a and b, we have P ( S ^ = a ) = P ( S ^ = b ) . Since a and b are arbitrary, the prior probabilities of the classification results S ^ are equal to 1 / L .

Appendix E. Proof of Proposition 2

Similar to [3], Proposition 2 can be proved as follows:
D = X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x ( 1 P t | x ) d x = 1 X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x P t | x d x ,
R u b = log 2 L + X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x log 2 s s P t | x w = 1 L e s P w | x d x = log 2 e ( ln L + X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x s P t | x d x + X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x ln 1 w = 1 L e s P w | x d x ) = log 2 e ( s + s X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x P t | x d x + ln e s + ln L + X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x ln 1 w = 1 L e s P w | x d x ) = log 2 e ( s D + X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x ln L e s w = 1 L e s P w | x d x ) = log 2 e ( s D + X f ( x ) ln L e s w = 1 L e s P w | x d x ) .
Denote η ( x ) = L e s w = 1 L e s P w | x . Then we have
R u b = log 2 e ( s D + X f ( x ) ln η ( x ) d x ) ,
d R u b d D = s log 2 e + log 2 e ( D d s d D + X f ( x ) 1 η ( x ) d η ( x ) d D d x ) = s log 2 e + log 2 e ( D d s d D + X f ( x ) η ( x ) d η ( x ) d s d s d D d x ) = s log 2 e + log 2 e ( X f ( x ) η ( x ) d η ( x ) d s d x D ) d s d D .
For the marginal probability P ( S ^ = t ) , we have
P ( S ^ = t ) = X e s P t | x w = 1 L e s P w | x f ( x ) d x .
As t = 1 L P ( S ^ = t ) = 1 , we have
t = 1 L P ( S ^ = t ) = X t = 1 L f ( x ) e s P t | x w = 1 L e s P w | x d x = 1 L X t = 1 L f ( x ) e s P t | x L e s e s w = 1 L e s P w | x d x = 1 L X t = 1 L f ( x ) e s ( 1 P t | x ) η ( x ) d x = 1 .
Hence the derivative of the second-to-last line of (A10) with respect to s should be 0. So we have
1 L X t = 1 L f ( x ) [ e s ( 1 P t | x ) ( P t | x 1 ) η ( x ) + e s ( 1 P t | x ) d η ( x ) d s ] d x = 0 ,
X t = 1 L f ( x ) e s P t | x [ ( P t | x 1 ) η ( x ) + d η ( x ) d s ] d x = 0 ,
X t = 1 L f ( x ) ( 1 P t | x ) e s P t | x w = 1 L e s P w | x L e s d x + X f ( x ) t = 1 L e s P t | x d η ( x ) d s d x = 0 .
From (A5), the first term on the left-hand side of the above equation is L e s D . For the second term, from the definition of η ( x ) , we have t = 1 L e s P t | x = L e s η ( x ) . Hence we have
L e s D + L e s X f ( x ) η ( x ) d η ( x ) d s d x = 0 .
So
X f ( x ) η ( x ) d η ( x ) d s d x D = 0 .
Combined with (A8), we have
d R d D = s log 2 e .

Appendix F. Proof of Theorem 4

Consider the case of encoding n observations ( n + ), where the reconstruction distortion is constrained. From the scheme described in Section 4, the rate of classification and reconstruction can be considered separately. By selecting a classification distortion D, the classification results can be described by n R s ( D ) bits with the soft classification scheme described in Section 3.2.2. Hence an achievable classification rate is R s ( D ) .
Then, according to the results of the soft classification, the observations are encoded with the corresponding sub-encoders. Since
I ( X ; X ^ | S ^ ) = u = 1 L P ( S ^ = u ) I ( X ; X ^ | S ^ = u ) ,
the average rate of each reconstruction encoder can be considered separately.
Consider the u-th sub-encoder. Denote the conditional expectation of the observations conditioned on the classification result u as γ u . We have
γ u = E { X | S ^ = u } = X x f ( x ) g D , u ( x ) d x X f ( x ) g D , u ( x ) d x .
Denote the conditioned covariance matrix conditioned on classification result u as Σ u . We have
Σ u = E { ( X γ u ) ( X γ u ) T | S ^ = u } = X ( X γ u ) ( X γ u ) T f ( x ) g D , u ( x ) d x X f ( x ) g D , u ( x ) d x .
Denote X ˜ u = X γ u . We use the u-th sub-encoder to encode X ˜ u . An achievable joint distribution between X ˜ u and X ^ is shown in Figure A1.
Figure A1. An achievable joint distribution between X ˜ u and X ^ .
Figure A1. An achievable joint distribution between X ˜ u and X ^ .
Entropy 27 00620 g0a1
First we diagonalize Σ u as Σ u = H u Σ Y u H u T , where H u is an orthonormal matrix. Assume X ˜ u = H u Y u . Σ Y u is the covariance matrix of Y u . Hence the components of Y u are independently conditioned on S ^ . Denote the dimension of X as K. Denote the i-th component of Y u as Y u i , with the corresponding variance denoted by σ u i 2 . Then assume Y ^ u i = σ u i 2 D u i σ u i 2 ( Y u i + Z u i ) , where Z u i N ( 0 , D u i σ u i 2 σ u i 2 D u i ) . u [ 1 : L ] , i [ 1 : N ] , Z u i is independent of X. Finally, the reconstruction X ^ = H u Y ^ u + γ u .
First we consider the distortion
E { X X ^ 2 2 | S ^ = u } = E { X ˜ u H u Y ^ u 2 2 | S ^ = u } = E { H u Y u H u Y ^ u 2 2 | S ^ = u } = E { Y u Y ^ u 2 2 | S ^ = u } = i = 1 K E { ( Y u i Y ^ u i ) 2 | S ^ = u } .
Under the joint distribution described in Figure A1, we have
E { ( Y u i Y ^ u i ) 2 | S ^ = u } = E { ( Y u i σ u i 2 D u i σ u i 2 ( Y u i + Z u i ) ) 2 | S ^ = u } = E { ( D u i σ u i 2 Y u i σ u i 2 D u i σ u i 2 Z u i ) 2 | S ^ = u } = ( D u i σ u i 2 ) 2 E { Y u i 2 | S ^ = u } + ( σ u i 2 D u i σ u i 2 ) 2 E { Z u i 2 | S ^ = u } = D u i .
Hence the overall expected distortion is
E { X X ^ 2 2 } = u = 1 L i = 1 K P ( S ^ = u ) D u i .
Then consider the rate (mutual information) of the u-th sub-encoder
I ( X ; X ^ | S ^ = u ) = h ( X ^ | S ^ = u ) h ( X ^ | X , S ^ = u ) .
For the first term in (A17),
h ( X ^ | S ^ = u ) = h ( H u Y ^ u | S ^ = u ) = h ( Y ^ u | S ^ = u ) = i = 1 K h ( Y ^ u i | S ^ = u ) .
It is obvious that the variance of Y ^ u i is σ u i 2 D u i . Under variance constraint, Gaussian source has the maximum differential entropy [45]. Hence we have
h ( Y ^ u i | S ^ = u ) 1 2 log 2 ( 2 π e ( σ u i 2 D u i ) ) .
For the second term in (A17),
h ( X ^ | X , S ^ = u ) = h ( H u Y ^ u | X , S ^ = u ) = h ( Y ^ u | X , S ^ = u ) = i = 1 K h ( Y ^ u i | X , S ^ = u ) ,
h ( Y ^ u i | X , S ^ = u ) = h ( σ u i 2 D u i σ u i 2 ( Y u i + Z u i ) | X , S ^ = u ) = h ( σ u i 2 D u i σ u i 2 Z u i | S ^ = u ) = 1 2 log 2 ( 2 π e ( σ u i 2 D u i ) D u i σ u i 2 ) .
Combined with (A18), we have
I ( X ; X ^ | S ^ = u ) 1 2 i = 1 K log 2 σ u i 2 D u i .
Assume D is an L × K matrix, with the element in the u-th row and i-th column taking the value of D u i . Denote R X ˜ u , g = 1 2 i = 1 K log 2 σ u i 2 D u i , which gives the rate of the u-th sub-encoder with the joint distribution described in Figure A1.
From (A19), we can see that for all D and D that satisfy u = 1 L i = 1 K P ( S ^ = u ) D u i D o , and D D s , the classification and reconstruction distortion constraints pair ( D , D o ) can be achieved by the overall rate R s ( D ) + u = 1 L P ( S ^ = u ) R X ˜ u , g .
When the classification distortion D is fixed, minimizing the overall rate is equivalent to minimizing the reconstruction rate. Hence we have the following optimization problem:
min D u i , u = [ 1 : L ] , i = [ 1 : K ] u = 1 L i = 1 K P ( S ^ = u ) log 2 σ u i 2 D u i s . t . u = 1 L i = 1 K P ( S ^ = u ) D u i D o D u i σ u i 2 , u = [ 1 : L ] , i = [ 1 : K ]
Hence we have the following Lagrange function:
J ( D ) = u = 1 L i = 1 K P ( S ^ = u ) log 2 σ u i 2 D u i + μ u = 1 L i = 1 K P ( S ^ = u ) D u i ,
where μ is the Lagrange multiplier. By setting the partial derivative of J with respect to D u i to 0, we obtain
D u i = log 2 e μ .
It can be seen from (A22) that under the optimal condition, every D u i is equal to a same constant. Denote the constant as α . We find that the solution to this problem is quite similar to the reverse water-filling method. So combining with the constraint D u i σ u i 2 , u = [ 1 : L ] , i = [ 1 : K ] , the optimal solution to minimize the reconstruction rate is
D u i = min { α , σ u i 2 } ,
where α is chosen to satisfy
u = 1 L i = 1 K P ( S ^ = u ) D u i = D o .

References

  1. Dobrushin, R.; Tsybakov, B. Information transmission with additional noise. IRE Trans. Inf. Theory 1962, 8, 293–304. [Google Scholar] [CrossRef]
  2. Wolf, J.; Ziv, J. Transmission of noisy information to a noisy receiver with minimum distortion. IEEE Trans. Inf. Theory 1970, 16, 406–411. [Google Scholar] [CrossRef]
  3. Berger, T. Rate Distortion Theory; Prentice-Hall: Englewood Cliffs, NJ, USA, 1971. [Google Scholar]
  4. Liu, J.; Zhang, W.; Poor, H.V. A Rate-Distortion Framework for Characterizing Semantic Information. In Proceedings of the 2021 IEEE International Symposium on Information Theory (ISIT), Melbourne, Australia, 12–20 July 2021; pp. 2894–2899. [Google Scholar] [CrossRef]
  5. Wang, Y.; Guo, T.; Bai, B.; Han, W. The Estimation-Compression Separation in Semantic Communication Systems. In Proceedings of the 2022 IEEE Information Theory Workshop (ITW), Mumbai, India, 1–9 November 2022; pp. 315–320. [Google Scholar]
  6. Gerrish, A.; Schultheiss, P. Information rates of non-Gaussian processes. IEEE Trans. Inf. Theory 1964, 10, 265–271. [Google Scholar] [CrossRef]
  7. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  8. Shannon, C.E. Coding theorems for a discrete source with a fidelity criterion. IRE Nat. Conv. Rec 1959, 4, 1. [Google Scholar]
  9. Bar-Hillel, Y.; Carnap, R. Semantic information. Br. J. Philos. Sci. 1953, 4, 147–157. [Google Scholar] [CrossRef]
  10. Floridi, L. Outline of a Theory of Strongly Semantic Information. Minds Mach. 2004, 14, 197–221. [Google Scholar] [CrossRef]
  11. Bao, J.; Basu, P.; Dean, M.; Partridge, C.; Swami, A.; Leland, W.; Hendler, J.A. Towards a theory of semantic communication. In Proceedings of the 2011 IEEE Network Science Workshop, West Point, NY, USA, 22–24 June 2011; pp. 110–117. [Google Scholar] [CrossRef]
  12. Juba, B. Universal Semantic Communication; Springer: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  13. Popovski, P.; Simeone, O.; Boccardi, F.; Gunduz, D.; Sahin, O. Semantic-Effectiveness Filtering and Control for Post-5G Wireless Connectivity. arXiv 2019, arXiv:1907.02441. [Google Scholar] [CrossRef]
  14. Kountouris, M.; Pappas, N. Semantics-Empowered Communication for Networked Intelligent Systems. IEEE Commun. Mag. 2021, 59, 96–102. [Google Scholar] [CrossRef]
  15. Seo, H.; Kang, Y.; Bennis, M.; Choi, W. Bayesian Inverse Contextual Reasoning for Heterogeneous Semantics- Native Communication. IEEE Trans. Commun. 2024, 72, 830–844. [Google Scholar] [CrossRef]
  16. Liu, X.; Sun, Y.; Wang, Z.; You, L.; Pan, H.; Wang, F.; Cui, S. User Centric Semantic Communications. arXiv 2024, arXiv:2411.03127. [Google Scholar]
  17. Shlezinger, N.; Eldar, Y.C.; Rodrigues, M.R.D. Hardware-Limited Task-Based Quantization. IEEE Trans. Signal Process. 2019, 67, 5223–5238. [Google Scholar] [CrossRef]
  18. Ordentlich, O.; Polyanskiy, Y. Optimal Quantization for Matrix Multiplication. arXiv 2024, arXiv:2410.13780. [Google Scholar]
  19. Shlezinger, N.; van Sloun, R.J.G.; Huijben, I.A.M.; Tsintsadze, G.; Eldar, Y.C. Learning Task-Based Analog-to-Digital Conversion for MIMO Receivers. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 9125–9129. [Google Scholar] [CrossRef]
  20. Shlezinger, N.; Amar, A.; Luijten, B.; van Sloun, R.J.G.; Eldar, Y.C. Deep Task-Based Analog-to-Digital Conversion. IEEE Trans. Signal Process. 2022, 70, 6021–6034. [Google Scholar] [CrossRef]
  21. Lexa, M.A.; Johnson, D.H. Joint Optimization of Distributed Broadcast Quantization Systems for Classification. In Proceedings of the 2007 Data Compression Conference (DCC’07), Snowbird, UT, USA, 27–29 March 2007; pp. 363–374. [Google Scholar] [CrossRef]
  22. Dogahe, B.M.; Murthi, M.N. Quantization for classification accuracy in high-rate quantizers. In Proceedings of the 2011 Digital Signal Processing and Signal Processing Education Meeting (DSP/SPE), Sedona, AZ, USA, 4–7 January 2011; pp. 277–282. [Google Scholar] [CrossRef]
  23. Poor, H. Fine quantization in signal detection and estimation. IEEE Trans. Inf. Theory 1988, 34, 960–972. [Google Scholar] [CrossRef]
  24. Kipnis, A.; Rini, S.; Goldsmith, A.J. The Rate-Distortion Risk in Estimation From Compressed Data. IEEE Trans. Inf. Theory 2021, 67, 2910–2924. [Google Scholar] [CrossRef]
  25. Fontana, R. Universal codes for a class of composite sources (Corresp.). IEEE Trans. Inf. Theory 1980, 26, 480–482. [Google Scholar] [CrossRef]
  26. Liu, J.; Poor, H.V.; Song, I.; Zhang, W. A Rate-Distortion Analysis for Composite Sources Under Subsource-Dependent Fidelity Criteria. arXiv 2024, arXiv:2405.11818. [Google Scholar] [CrossRef]
  27. Liu, J.; Shao, S.; Zhang, W.; Poor, H.V. An Indirect Rate-Distortion Characterization for Semantic Sources: General Model and the Case of Gaussian Observation. IEEE Trans. Commun. 2022, 70, 5946–5959. [Google Scholar] [CrossRef]
  28. Witsenhausen, H. Indirect rate distortion problems. IEEE Trans. Inf. Theory 1980, 26, 518–521. [Google Scholar] [CrossRef]
  29. Liu, K.; Liu, D.; Li, L.; Ning, Y.; Li, H. Semantics-to-Signal Scalable Image Compression with Learned Revertible Representations. Int. J. Comput. Vis. 2021, 129, 2605–2621. [Google Scholar] [CrossRef]
  30. Rabiner, L.R.; Schafer, R.W. Introduction to Digital Speech Processing. Found. Trends® Signal Process. 2007, 1, 1–194. [Google Scholar] [CrossRef]
  31. Kalveram, H.; Meissner, P. Itakura-saito clustering and rate distortion functions for a composite source model of speech. Signal Process. 1989, 18, 195–216. [Google Scholar] [CrossRef]
  32. Gibson, J.D.; Hu, J. Rate Distortion Bounds for Voice and Video. Found. Trends® Commun. Inf. Theory 2014, 10, 379–514. [Google Scholar] [CrossRef]
  33. Gibson, J. Rate distortion functions and rate distortion function lower bounds for real-world sources. Entropy 2017, 19, 604. [Google Scholar] [CrossRef]
  34. Rabiner, L. A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef]
  35. Lu, D.; Weng, Q. A survey of image classification methods and techniques for improving classification performance. Int. J. Remote Sens. 2007, 28, 823–870. [Google Scholar] [CrossRef]
  36. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  37. Chandola, V.; Banerjee, A.; Kumar, V. Anomaly detection: A survey. ACM Comput. Surv. (CSUR) 2009, 41, 1–58. [Google Scholar] [CrossRef]
  38. Kingma, D.P. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  39. Croitoru, F.A.; Hondru, V.; Ionescu, R.T.; Shah, M. Diffusion models in vision: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10850–10869. [Google Scholar] [CrossRef] [PubMed]
  40. Tang, Y.; Fei, Z.; Huang, L.; Zhang, W.; Zhao, B.; Guan, H.; Huang, Y. Failure Mode and Effects Analysis Method on the Air System of an Aircraft Turbofan Engine in Multi-Criteria Open Group Decision-Making Environment. Cybern. Syst. 2025, 1–32. [Google Scholar] [CrossRef]
  41. Popkov, Y.S.; Volkovich, Z.; Dubnov, Y.A.; Avros, R.; Ravve, E. Entropy “2”-Soft Classification of Objects. Entropy 2017, 19, 178. [Google Scholar] [CrossRef]
  42. Sharma, R.; Garg, P.; Dwivedi, R. A literature survey for fuzzy based soft classification techniques and uncertainty estimation. In Proceedings of the 2016 International Conference System Modeling & Advancement in Research Trends (SMART), Moradabad, India, 25–27 November 2016; pp. 71–75. [Google Scholar] [CrossRef]
  43. Khatami, R.; Mountrakis, G.; Stehman, S.V. Predicting individual pixel error in remote sensing soft classification. Remote Sens. Environ. 2017, 199, 401–414. [Google Scholar] [CrossRef]
  44. Weidmann, C.; Vetterli, M. Rate Distortion Behavior of Sparse Sources. IEEE Trans. Inf. Theory 2012, 58, 4969–4992. [Google Scholar] [CrossRef]
  45. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2005. [Google Scholar]
  46. Blahut, R. Computation of channel capacity and rate-distortion functions. IEEE Trans. Inf. Theory 1972, 18, 460–473. [Google Scholar] [CrossRef]
  47. Chen, L.; Wu, S.; Ye, W.; Wu, H.; Zhang, W.; Wu, H.; Bai, B. A constrained BA algorithm for rate-distortion and distortion-rate functions. arXiv 2023, arXiv:2305.02650. [Google Scholar]
  48. Yeung, R.W. A First Course in Information Theory; Springer Science & Business Media: New York, NY, USA, 2012. [Google Scholar]
Figure 1. System model.
Figure 1. System model.
Entropy 27 00620 g001
Figure 2. The procedure of the HCTC scheme.
Figure 2. The procedure of the HCTC scheme.
Entropy 27 00620 g002
Figure 3. g ˜ 1 ( x ) when s is set to 0, 1 and 100. (a) g ˜ 1 ( x ) when s = 0 , (b) g ˜ 1 ( x ) when s = 1 , (c) g ˜ 1 ( x ) when s = 100 .
Figure 3. g ˜ 1 ( x ) when s is set to 0, 1 and 100. (a) g ˜ 1 ( x ) when s = 0 , (b) g ˜ 1 ( x ) when s = 1 , (c) g ˜ 1 ( x ) when s = 100 .
Entropy 27 00620 g003
Figure 4. The procedure of the CAC scheme.
Figure 4. The procedure of the CAC scheme.
Entropy 27 00620 g004
Figure 5. The mean points distribution of the fifth, sixth and seventh settings. (a) The mean points distribution of the fifth setting. (b) The mean points distribution of the sixth setting. (c) The mean points distribution of the seventh setting.
Figure 5. The mean points distribution of the fifth, sixth and seventh settings. (a) The mean points distribution of the fifth setting. (b) The mean points distribution of the sixth setting. (c) The mean points distribution of the seventh setting.
Entropy 27 00620 g005
Figure 6. The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the 4-state, 16-state, and 32-state composite sources. (a) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the first setting. (b) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the second setting. (c) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the third setting. (d) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the fourth setting. (e) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the fifth setting. (f) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for sixth setting. (g) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for seventh setting.
Figure 6. The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the 4-state, 16-state, and 32-state composite sources. (a) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the first setting. (b) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the second setting. (c) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the third setting. (d) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the fourth setting. (e) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for the fifth setting. (f) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for sixth setting. (g) The classification rate–distortion performance of the soft classification scheme versus the HCTC scheme for seventh setting.
Entropy 27 00620 g006
Figure 7. The change in the ratio of the bits required for the soft classification and the HCTC as the extra classification distortion constraint beyond the Bayesian minimum classification distortion increases in different sources.
Figure 7. The change in the ratio of the bits required for the soft classification and the HCTC as the extra classification distortion constraint beyond the Bayesian minimum classification distortion increases in different sources.
Entropy 27 00620 g007
Figure 8. The rate–distortion performance of the composite sources with discrete observation. (a) The rate–distortion performance of the composite source with discrete observation with equal prior probabilities of states, (b) The rate–distortion performance of the composite source with discrete observation with unequal prior probabilities of states.
Figure 8. The rate–distortion performance of the composite sources with discrete observation. (a) The rate–distortion performance of the composite source with discrete observation with equal prior probabilities of states, (b) The rate–distortion performance of the composite source with discrete observation with unequal prior probabilities of states.
Entropy 27 00620 g008
Figure 9. The rate–distortion performance of soft classification scheme and HCTC scheme in the composite source where the prior distribution of states in the first setting is changed to be non-uniform.
Figure 9. The rate–distortion performance of soft classification scheme and HCTC scheme in the composite source where the prior distribution of states in the first setting is changed to be non-uniform.
Entropy 27 00620 g009
Figure 10. The plot of R C ( s ) in different settings. (a) R C ( s ) in the first setting, (b) R C ( s ) in the second setting, and (c) R C ( s ) in the third setting.
Figure 10. The plot of R C ( s ) in different settings. (a) R C ( s ) in the first setting, (b) R C ( s ) in the second setting, and (c) R C ( s ) in the third setting.
Entropy 27 00620 g010
Figure 11. The plot of R u b ( + , D o ) in different settings. (a) R u b ( + , D o ) in the first setting, (b) R u b ( + , D o ) in the second setting.
Figure 11. The plot of R u b ( + , D o ) in different settings. (a) R u b ( + , D o ) in the first setting, (b) R u b ( + , D o ) in the second setting.
Entropy 27 00620 g011
Figure 12. The distortion–rate performance of different schemes in the 3rd, 4th, and 5th settings. (a) The distortion–rate performance of different schemes in the third setting. (b) The distortion–rate performance of different schemes in the fourth setting. (c) The distortion–rate performance of different schemes in the fifth setting.
Figure 12. The distortion–rate performance of different schemes in the 3rd, 4th, and 5th settings. (a) The distortion–rate performance of different schemes in the third setting. (b) The distortion–rate performance of different schemes in the fourth setting. (c) The distortion–rate performance of different schemes in the fifth setting.
Entropy 27 00620 g012
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, Y.; Liu, J.; Zhang, W. Soft Classification in a Composite Source Model. Entropy 2025, 27, 620. https://doi.org/10.3390/e27060620

AMA Style

Cao Y, Liu J, Zhang W. Soft Classification in a Composite Source Model. Entropy. 2025; 27(6):620. https://doi.org/10.3390/e27060620

Chicago/Turabian Style

Cao, Yuefeng, Jiakun Liu, and Wenyi Zhang. 2025. "Soft Classification in a Composite Source Model" Entropy 27, no. 6: 620. https://doi.org/10.3390/e27060620

APA Style

Cao, Y., Liu, J., & Zhang, W. (2025). Soft Classification in a Composite Source Model. Entropy, 27(6), 620. https://doi.org/10.3390/e27060620

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop