Next Article in Journal
Successive Refinement for Lossy Compression of Individual Sequences
Next Article in Special Issue
Rate–Distortion–Perception Trade-Off in Information Theory, Generative Models, and Intelligent Communications
Previous Article in Journal
Action Recognition with 3D Residual Attention and Cross Entropy
Previous Article in Special Issue
Broadcast Channel Cooperative Gain: An Operational Interpretation of Partial Information Decomposition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature-Driven Semantic Communication for Efficient Image Transmission

1
School of Mathematics and Statistics, Henan University of Science and Technology, Luoyang 471000, China
2
Intelligent System Science and Technology Innovation Center, Longmen Laboratory, Luoyang 471000, China
3
Engineering Research Center of Intelligent Swarm Systems, Ministry of Education, Zhengzhou University, Zhengzhou 450001, China
4
School of Computer, Henan University of Engineering, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(4), 369; https://doi.org/10.3390/e27040369
Submission received: 27 February 2025 / Revised: 21 March 2025 / Accepted: 28 March 2025 / Published: 31 March 2025
(This article belongs to the Special Issue Semantic Information Theory)

Abstract

:
Semantic communication is an emerging approach that enhances transmission efficiency by conveying the semantic content of information more effectively. It has garnered significant attention in recent years. However, existing semantic communication systems for image transmission typically adopt direct transmission of features or uniformly compress features before transmission. They have not yet considered the differential impact of features on image recovery at the receiver end and the issue of bandwidth limitations during actual transmission. This paper shows that non-uniform processing of features leads to better image recovery under bandwidth constraints compared to uniform processing. Based on this, we propose a semantic communication system for image transmission, which introduces non-uniform quantization techniques. In the feature transmission stage, the system performs varying levels of quantization based on the differences in feature performance at the receiver, thereby reducing the bandwidth requirement. Inspired by quantitative quantization techniques, we design a non-uniform quantization algorithm capable of dynamic bit allocation. This algorithm, under bandwidth constraints, dynamically adjusts the quantization precision of features based on their contribution to the completion of tasks at the receiver end, ensuring the quality and accuracy of the transmitted data even under limited bandwidth conditions. Experimental results show that the proposed system reduces bandwidth usage while ensuring image reconstruction quality.

1. Introduction

The global deployment and application of 5G technology have brought existing wireless communication systems closer to the Shannon limit in terms of performance [1]. This makes it difficult to further increase the channel capacity, and they cannot meet the massive and high-frequency data transmission demands under limited bandwidth conditions [2]. The development of artificial intelligence technology has provided new pathways for overcoming the limitations of traditional communication, driving the development of a communication paradigm aimed at transmitting semantic information–semantic communication. Compared to traditional communication, semantic communication emphasizes the transmission of information’s semantic content, thereby improving communication efficiency and task execution accuracy [3].
Current research mainly focuses on semantic communication for data types such as text, speech, and images. For example, Wang et al. [4] designed a semantic communication system for speech transcription based on the masked autoencoder; Xie et al. [5] combined visual and semantic data to design a semantic communication system for visual question answering. Hu et al. [6,7] used variational autoencoders to design a semantic communication system for image classification; Pan et al. [8] utilized semantic features from vehicle perception images to design an image segmentation semantic communication system for vehicle driving decision-making. However, these studies mainly focus on using the transmitted data to complete tasks at the receiver end, and they do not adequately address the issues related to data recovery and reconstruction, especially considering the high dimensionality and noise sensitivity of image data.
As a result, some researchers have focused on semantic communication systems for image transmission. E. Bourtsoulatze et al. [9] proposed an image transmission system based on convolutional neural networks (CNNs), which achieved better image recovery quality than traditional image transmission methods. K. Tonchev et al. [10] designed a semantic communication system based on graph convolutional networks (GCNs) to achieve the transmission and recovery of 3D video data. In addition to CNN-based communication systems, K. Yang et al. [11] used the self-attention mechanism and global modeling capability of transformers to design an image transmission system based on the Swin Transformer architecture, which showed superior image recovery performance compared to traditional communication systems and some semantic communication systems based on CNN architectures.
Although the image reconstruction results from the aforementioned semantic communication systems are satisfactory, they all aim to improve the quality of information recovery at the receiver end, with little consideration of the differential contribution of semantic information to the receiver.
This limitation has led to further consideration of how to optimize the transmission of semantic information, particularly how to apply differentiated encoding based on the contribution of different semantic information to the receiver. To address this, Lin et al. [12] proposed a semantic importance-based encoding algorithm that aggregates image channel features to obtain the importance weights of feature maps for task results, prioritizing the transmission of high-importance semantic feature maps to reduce redundant information in the transmission process. Considering the limited bandwidth issue in practical transmission, Huang et al. [13] divided the image into a base layer and an enhancement layer, prioritizing the transmission of key information at low bit rates to reduce transmission resource consumption. These approaches bear some resemblance to adaptive quantization methods in prior work. For instance, Wu et al. [14] proposed an adaptive quantization scheme based on semantic segmentation, dividing images into regions of interest (ROIs) and regions of non-interest (RONIs) with distinct encoding to optimize transmission efficiency. Gupta et al. [15] developed a method that dynamically adjusts quantization levels for neural network weights based on their statistical distribution, targeting memory reduction. Han et al. [16] introduced a technique combining pruning and quantization, adaptively assigning fixed bit-widths to weights based on magnitude for computational efficiency. Lei et al. [17] designed an ROI-based adaptive quantization for image compression, adjusting bit allocation to prioritize predefined regions critical for tasks like classification. However, these methods have notable limitations, that is, Wu et al. and Lei et al. relied on predefined ROI distinctions, restricting global optimization and adaptability to dynamic conditions; Gupta et al. and Han et al. applied uniform quantization across layers or parameters, neglecting feature-specific importance and bandwidth constraints during transmission.
In contrast, this paper proposes a feature-driven semantic communication framework that advances beyond these methods with non-uniform quantization based on feature contribution factors. Unlike Wu et al. [14] and Lei et al.’s [17] region-specific approaches or Gupta et al. [15] and Han et al.’s [16] uniform quantization, our method introduces a feature contribution factor to dynamically allocate bits across all image features according to their reconstruction importance. It optimizes global image recovery rather than predefined regions and adapts in real-time to bandwidth constraints, ensuring both efficiency and high-quality reconstruction under limited resources. The main contributions are as follows:
  • We designed a feature-driven semantic communication framework that uses non-uniform quantization methods based on the contribution factor of image features, optimizing bandwidth usage while ensuring image quality.
  • This paper designs a non-uniform quantization algorithm for transmitted features, which is based on fixed-point quantization technology. The algorithm dynamically adjusts the quantization precision by combining the contribution factors of features and the transmission bandwidth requirements, maximizing data transmission efficiency and image quality during the transmission process.
  • The performance of the proposed system was validated under additive white Gaussian noise (AWGN), Rayleigh fading channels, and a composite channel model. This composite model combines Rician fading, multipath effects, and time-varying noise. The simulation results demonstrate the improvements in image transmission efficiency and quality of the proposed system compared to traditional uniform quantization. These improvements are particularly evident under limited bandwidth conditions across all tested channel environments.
The remainder of this paper is organized as follows: Section 2 describes the quantization techniques and the development of semantic communication. Section 3 presents the problem formulation, along with the system model and the design of the non-uniform quantization algorithm. Section 4 analyzes the existing simulation results. The conclusion is provided in Section 5, and Section 6 outlines potential directions for future work.

2. Related Works

2.1. Quantitative Methods

Quantization techniques were initially used for the digital processing of analog signals [18]. In particular, during the analog-to-digital conversion (ADC) process, continuous analog signals are discretized into a finite number of digital values, achieving a digital representation of the signal [19]. Quantization methods are divided into uniform and non-uniform quantization based on the difference in quantization intervals. Uniform quantization divides the data range into multiple equal-sized intervals for quantization, while non-uniform quantization performs quantization with unequal intervals based on the statistical properties of the data to improve quantization efficiency. In the field of image processing, quantization methods are widely applied in image compression, especially in early image compression standards such as JPEG [20], which uses uniform quantization techniques to quantize the coefficients after discrete cosine transform (DCT), adjusting the balance between compression ratio and image quality.
With the rise of deep learning and the increasing complexity of neural network models, quantization techniques have been redefined as an important method for neural network compression. Lin et al. [21] demonstrated the potential of quantization as a regularization technique by applying different bit quantization methods to multiple neural networks. Jacob et al. [22] proposed a pure integer quantization method, where both the weight and activation values were uniformly quantized to the INT8 type, further optimizing network performance. Notably, Liu et al. [23] introduced the binary neural network (BNN) quantization method, where all network weight and activation values are quantized to +1 and −1, significantly simplifying the model structure and reducing storage requirements. Inspired by this method, subsequent research has proposed various improvements and optimization methods [24].
In recent years, to improve the flexibility and applicability of quantization methods, mixed-precision quantization based on parameter differences has garnered significant attention [25]. This approach, which we refer to as a method of non-uniform quantization, allocates different bit precisions to parameters in different layers.
Chen et al. [26] designed a deep neural network model compression algorithm based on a three-value weight network, where full-precision weights were quantized into four states, enhancing the flexibility of weight quantization. Zhen et al. [27] introduced Hessian-aware quantization (HAWQ), which analyzes the curvature information of network layer outputs to establish a quantization sensitivity evaluation model, allocating the optimal bit width for each layer. Similarly, Zhang et al. [28] proposed a mixed-precision quantization method for depth-separable convolutional networks, assigning different bit widths to each layer based on the weight distribution and importance of each layer to reduce model size.
Uhlich et al. [29] proposed a method where the dynamic range and step size of the quantizer were parameterized as differentiable variables. By dynamically updating these parameters during training, they achieved the optimal bit width for mixed-precision quantization. In the ImageNet task, this method achieved less than 0.5% accuracy loss while compressing the ResNet-18 model by 4.24 times. Wu et al. [30] proposed the differentiable neural architecture search (DNAS), where a supernetwork with different bit-width candidate operations is constructed. By relaxing discrete bit-width choices to continuous probability distributions using Gumbel-Softmax, they assigned lower precision to redundant layers and retained high precision for sensitive layers, achieving model compression.
Wang et al. [31] proposed the APQ framework, which uses an evolutionary algorithm to search for mixed-precision strategies on a pre-trained “Once-For-All” network. By estimating latency and energy consumption for each layer at different bit-widths, the framework quickly computes model accuracy and resource usage, enabling pruning and mixed-precision quantization. Van et al. [32] introduced a Bayesian quantization framework that decomposes weights into multiple sets of power-of-two bit-width residuals. Using a variational inference approach, they jointly optimized bit-width and pruning gates, achieving mixed-precision quantization.
Additionally, some works have utilized reinforcement learning methods to dynamically select the appropriate quantization bit-width for each layer of neural networks. This helps maintain model inference accuracy while reducing storage and computational resource consumption. For example, Wang et al. [33] used the deep deterministic policy gradient (DDPG) to optimize per-layer bit-widths, incorporating hardware feedback (e.g., delay, energy consumption) to achieve multi-objective constraint optimization. In MobileNet-V2, this method achieved 8.03x model compression with only a 1.14% accuracy loss, considering FPGA resource constraints. Elthakeb [34] et al. proposed an LSTM-based reinforcement learning framework with an asymmetric reward function to balance accuracy and compression, prioritizing high precision for key layers and avoiding model collapse due to low precision.
Liu et al. [35] introduced a hierarchical reinforcement learning architecture (HLC and LLC controllers), jointly optimizing network architecture and per-core mixed precision. Using hardware model predictions, they reduced real deployment costs. This method, when applied to ResNet-50, reduced latency by 39% and energy consumption by 33%, compared to traditional methods. In summary, the development of non-uniform quantization techniques has greatly improved the efficiency of neural network models, enabling them to maintain high accuracy while reducing storage and computational demands.

2.2. Semantic Communication

As early as 1953, Weaver introduced the concept of semantics [36]. It focuses on transmitting semantic information rather than raw data [37,38]. Early communication theory mainly focused on the reliability of channel transmission. It paid limited attention to the semantics of the transmitted information [39]. However, with evolving communication needs [40] and advances in deep learning technology [41], semantic communication systems have gained increasing attention. Traditional communication systems focus on the accurate transmission of data. They do not consider the semantic relevance of the transmitted information. In contrast, using semantic encoding modules to extract and compress semantic features can improve communication efficiency and optimize the process [42,43,44,45]. Yet, many existing semantic communication systems face challenges in information transmission. These include high model complexity [46], large computational resource demands [47], and limited transmission efficiency.
To address these issues, some researchers have explored importance-based encoding. This approach aims to prioritize and encode information features based on their importance. It seeks to enhance transmission efficiency and reduce redundant data. These methods typically rank features and prioritize the transmission of higher-ranked ones. Zhou et al. [48] used feature importance and robustness as priority metrics. They prioritized the transmission of top-ranked features to ensure key image features are sent even under poor channel conditions. Gao et al. [49] defined semantic value to measure the importance of semantic information. They selected top-ranked semantic triples for transmission. Ma et al. [50] applied generative artificial intelligence (GenAI) to divide visual data into semantic segments. They ranked these segments by task relevance and transmitted only those with higher priority. Although these methods work in some cases, they generally face issues. These include high computational overhead, poor real-time performance, and complexity in dynamically assessing feature importance. In bandwidth-limited or resource-constrained environments, efficiently transmitting key features without adding computational burden remains a critical challenge. To address this, we propose a semantic communication framework based on feature contribution factors and non-uniform quantization. Compared to existing methods, our framework avoids complex feature ranking and predictive computation. It adjusts quantization precision according to each feature’s contribution to image recovery quality. Our method can adapt to the varying impacts of features on image reconstruction.

3. Feature-Driven Semantic Communication System

In this section, we discuss our proposed method in detail. First, we provide a theoretical analysis to demonstrate that, under resource-constrained conditions, applying non-uniform quantization to features, as opposed to uniform quantization, leads to better image reconstruction quality. Next, we introduce the semantic communication system model that we have designed. Finally, we elaborate on the non-uniform quantization method for features that we developed and explore its application in image reconstruction.

3.1. Problem Formulation

In communication systems, to reduce transmission bandwidth, quantization is applied to the extracted features before transmission. This process discretizes continuous-valued features, reducing data size while maintaining essential information. In semantic communication systems for image transmission, it is commonly assumed that all features contribute equally to image reconstruction at the receiver. Under this assumption, uniform quantization is applied to all features, treating them as equally important. However, in practice, different features have varying levels of significance. Some features play a more critical role in enhancing image quality, while others contribute less. Ignoring the non-uniform impact of features and applying uniform quantization can lead to the following issues:
  • Information loss: Important features may suffer excessive quantization errors, resulting in significant degradation of image quality.
  • Inefficient resource allocation: Less important features may receive unnecessary bit allocation, leading to inefficient transmission.
To address these issues, this paper proposes a feature-aware non-uniform quantization method that adapts quantization parameters based on the varying contributions of features. We introduce the contribution factor, c i , to measure the importance of each feature in the image reconstruction. It is defined as follows:
c i = r ( x i ) i = 1 n r ( x i ) ,
where c i represents the contribution factor of feature x i , n denotes the total number of features, and r ( x i ) quantifies the impact of feature x i on image reconstruction quality. Our objective is to design an adaptive quantization algorithm that optimizes transmission efficiency while maintaining high image recovery quality. In the next section, we analyze the advantages of non-uniform quantization in reducing quantization error.

3.2. Analysis of the Advantages of Non-Uniform Quantization in Image Recovery

Through analysis, we demonstrate the advantages of non-uniform quantization over uniform quantization in image recovery quality under a fixed bit budget.

3.2.1. Assumptions

To facilitate the theoretical analysis, we make the following assumptions:
Assumption 1 (feature distribution): The extracted features X = { x 1 , x 2 , , x N } are assumed to have a smooth probability distribution within their dynamic ranges. While deep neural networks may produce feature distributions that are approximately Gaussian or slightly non-uniform, we assume that within each feature’s dynamic range, the values can be approximated as locally uniform. This assumption can simplify the quantization error model and allow for an analytical derivation of the mean squared error (MSE).
Assumption 2 (quantization as the sole source of distortion): We assume that quantization is the sole source of distortion in the system, excluding factors such as channel noise or encoding errors.
Assumption 3 (bit budget constraint): We assume that the total available bits for quantization, denoted by R, is the sum of the bit allocations b i for each feature x i , as follows:
R = i = 1 N b i , where b i Z .
Additionally, the total bit budget is constrained by the maximum channel capacity R max , such that we have the following:
R R max .
Assumption 4 (mean squared error as the quality metric): We use mean squared error (MSE) as the metric for evaluating image reconstruction quality, defined as follows:
M S E = 1 N i = 1 N ( x i x ^ i ) 2 ,
where x i represents the original image and x ^ i denotes the reconstructed image.
Assumption 5 (quantization error model): We assume that the quantization error for each feature x i follows a uniform distribution within its respective quantization interval:
ε i = x i x ^ i Uniform Δ i 2 , Δ i 2 ,
where Δ i is the quantization step size.
With these assumptions, we have the following theorem.

3.2.2. Theorem and Proof

Theorem 1. 
Under Assumptions 1–5, the optimal bit allocation for minimizing the Mean Squared Error (MSE) in image recovery follows a non-uniform quantization scheme. The number of bits b i assigned to each feature x i is proportional to the logarithm of its contribution factor c i , given by the following:
b i * = 1 2 log 2 2 c i · C · ln 2 λ .
Furthermore, non-uniform quantization results in a lower overall MSE compared to uniform quantization:
M S E non-uniform M S E uniform .
Proof. 
Step 1: Formulating the MSE Expression According to Assumption 5, the quantization error ε i follows a uniform distribution in its quantization interval, leading to a quantization noise variance:
E [ ε i 2 ] = Δ i 2 12 .
By Assumption 4, the total MSE is given by the following:
M S E = i = 1 N c i · Δ i 2 12 .
Substituting the quantization step size, we have the following:
Δ i = x i , max x i , min 2 b i 1 ,
we obtain the following:
M S E = i = 1 N c i · ( x i , max x i , min ) 2 12 ( 2 2 b i 1 ) .
Step 2: optimization problem formulation. The objective is to minimize the total MSE subject to the bit budget constraint given in Assumption 3:
i = 1 N b i = R .
Using the Lagrange multiplier method, we define the Lagrangian function as follows:
L = i = 1 N c i · ( x i , max x i , min ) 2 12 ( 2 2 b i 1 ) + λ i = 1 N b i R .
Following the standard derivation using the Lagrange multiplier method, the optimal bit allocation is obtained as follows:
b i * = 1 2 log 2 2 c i · C · ln 2 λ .
Step 3: Performance Comparison Applying the arithmetic-geometric mean (AM-GM) inequality, we have the following:
1 N i = 1 N a i i = 1 N a i 1 N ,
we establish the following:
M S E non-uniform M S E uniform .
This completes the proof.  □
The above theoretical analysis establishes that the proposed non-uniform quantization strategy effectively reduces MSE compared to uniform quantization. Equation (16) indicates that the number of bits assigned to each feature depends on its relative importance, leading to a more efficient representation of the image data. This result ensures reduced storage and transmission bandwidth requirements while maintaining image recovery quality.

3.3. Overall Architecture

Based on the above theory, this paper designs a feature-contribution-factor-based non-uniform quantization semantic communication system, which reduces data storage and transmission bandwidth requirements while ensuring image recovery quality. The semantic communication system considered in this paper consists of two main parts, namely, the transmitter and the receiver. The system flow is shown in Figure 1.
The transmitter extracts features from the input image, x, through the encoder, E, and obtains the encoded feature vector, s; the transmission process is represented as follows:
s = E ( · ; θ 1 ) ,
where θ 1 represents the parameter set of the encoder. Based on the contribution factors of the features, the representation of the feature s is non-uniformly quantized to obtain the compressed feature s = q ( · ; θ 2 ) , where θ 2 is the parameter set used for non-uniform quantization. Before transmitting the compressed feature through the wireless channel, power normalization is performed to ensure that the signal satisfies the specified average power constraint. In this paper, we consider a general fading channel model, where the transmission process is represented as follows:
y = h s + n ,
where h denotes the channel state information (CSI) vector, representing the channel characteristics, ⊙ denotes element-wise multiplication, and n is the noise vector, with each component independently sampled from a Gaussian distribution. The noise distribution is characterized by noise power σ n 2 , and I k is the identity matrix. The receiver performs the corresponding dequantization on the received feature vector y to obtain y = q ( · ; θ 3 ) . The decoder D is then used to reconstruct the dequantized feature into the image x ^ , which can be expressed as follows:
x ^ = D ( · ; θ 4 ) ,
where θ 3 represents the parameter set for dequantization, and θ 4 represents the parameter set of the decoder.

3.4. Non-Uniform Quantization Algorithm Based on Feature Contribution Factors

This section designs a non-uniform quantization algorithm based on feature contribution factors. The algorithm can assign differential quantization precision to image features based on their contribution, maintaining the integrity of key information while reducing the data size, lowering transmission cost, improving transmission efficiency, and facilitating compatibility with digital communication systems. This foundation optimizes bandwidth utilization for efficient data transmission in practical communication environments. The core idea of the algorithm is to first classify features into different categories based on their contribution factors. Then, the total bit count is allocated to each category of features, followed by the allocation of bits to the symbol, integer, and fractional parts, and calculating the quantization step sizes for each part. As depicted in Figure 2, the n features after semantic encoding are divided into m feature categories based on their contribution factors, denoted as D 1 , D 2 , , D m . Each category contains t 1 , t 2 , , t m features, and their sum equals n. For each feature category D j , the total number of bits allocated for the feature quantization representation is denoted as b k , where k represents the feature category, and quantization is performed according to a specific bit allocation formula.
First, based on the feature contribution factors, the sender classifies the features into priority categories before transmission. The larger the contribution factor, the higher the priority of the corresponding feature. These features are divided into m categories according to their priority. For each feature category D j , the total number of bits allocated for feature quantization representation is denoted as b k , where k represents the feature category. The bit allocation formula is defined as follows:
b k = i = 1 k c i k = 1 M i = 1 k c i · B
where B is the total bit budget of the system, and c i represents the contribution factor of feature x i , typically determined by the task objective or reconstruction quality. In an ideal scenario, these factors exhibit certain patterns. However, in practical applications, the complexity of task objectives and reconstruction quality makes it difficult to directly obtain the contribution factors. In this paper, the gradient values are used as a measure of the feature contribution factors. Next, for each feature category, based on the allocated bit count b k , the bits are divided into symbol bits b k 0 , integer bits b k 1 , and fractional bits b k 2 , with the constraint that b k 0 = 1 . The quantization step size formulas for the integer and fractional parts are as follows:
Δ k 1 = | x k | max | x k | min 2 b k 1 1 , Δ k 2 = | x k | max | x k | min 2 b k 2 1
For each feature x k i in the category, the corresponding quantization step size is used to quantize the feature, and the integer and fractional parts are combined to yield the final quantized feature value x ˜ k i . The expression is as follows:
x ˜ k i = Δ k 1 × | x k i | | x k | min Δ k + 1 2 · L ( b k ) + Δ k 2 × | x k i | | x k | min Δ k + 1 2
where L ( · ) represents the left shift operation, which is used to combine the integer and fractional parts according to their allocated bit precision. Algorithm 1 summarizes the entire workflow of the non-uniform quantization algorithm.
Algorithm 1 Contribution factor-based non-uniform quantization.
1:
Input: Feature set X, contribution factors C, total bit budget B, number of categories m.
2:
Output: Quantized feature set X ˜ .
3:
Step 1: Feature classification
4:
for each feature x i in X do
5:
    Assign contribution factor c i based on task or reconstruction quality.
6:
    Classify features into m categories based on c i .
7:
end for
8:
Step 2: Bit allocation
9:
for each category D j  do
10:
    Allocate bits b k to each category based on the contribution factors:
b k = i = 1 k c i k = 1 M i = 1 k c i · B .
11:
end for
12:
Step 3: Bit splitting
13:
for each category D j  do
14:
    Split b k into symbol bits b k 0 = 1 , integer bits b k 1 , and fractional bits b k 2 ,
15:
    Calculate the quantization step sizes Δ k 1 and Δ k 2 :
Δ k 1 = | x k | max | x k | min 2 b k 1 1 , Δ k 2 = | x k | max | x k | min 2 b k 2 1 .
16:
end for
17:
Step 4: Quantization
18:
for each feature x k i in D j  do
19:
    Quantize the feature using Δ k 1 and Δ k 2 :
x k i ˜ = Δ k 1 · | x k i | | x k | min Δ k 1 + 0.5 · L ( b k 2 ) + Δ k 2 · | x k i | | x k | min Δ k 2 + 0.5 .
20:
end for
21:
Step 5: Output
22:
Return quantized feature set X ˜ .
At the receiver end, the received quantized values need to be dequantized. First, the received quantized feature values are separated into integer and fractional parts according to the given integer bits b k 1 and fractional bits b k 2 . Then, based on the quantization step sizes for the integer and fractional parts, they are dequantized. The restored integer and fractional parts are combined to form the final dequantized feature value x ^ k i . The detailed dequantization algorithm is shown in Algorithm 2.
Algorithm 2 Dequantization algorithm for contribution factor-based non-uniform quantization.
1:
Input: Quantized feature set X ˜ , feature categories D 1 , D 2 , , D m , contribution factors C, total bit budget B, quantization step sizes Δ k 1 , Δ k 2 .
2:
Output: Restored feature set X.
3:
Step 1: Bit splitting and inverse bit allocation
4:
for each feature category D j  do
5:
    Split each category’s total bit allocation b k into symbol bits b k 0 , integer bits b k 1 , and fractional bits b k 2 ,
6:
    Input: Inverse quantization step sizes Δ k 1 and Δ k 2 from the sender.
7:
end for
8:
Step 2: Dequantization of each feature
9:
for each quantized feature x k i ˜ in D j  do
10:
    Extract the integer part and fractional part from x k i ˜ using the left shift operation L ( b k 2 ) :
x ^ int = x k i ˜ L ( b k 2 ) ,
x ^ f r a c = x ˜ k i mod L ( b k 2 ) .
11:
    Restore the original value for integer and fractional parts:
x int = | x k | min + x ^ int · Δ k 1 ,
x frac = | x k | min + x ^ f r a c · Δ k 2 .
12:
    Combine the integer and fractional parts:
x k i = x int + x frac .
13:
end for
14:
Step 3: Output
15:
Return the restored feature set X = { x 1 , x 2 , , x n } .
To preserve the gradient information during the quantization process, this paper adopts the straight-through estimator (STE). Using the STE technique, during backpropagation, an approximate gradient is used to replace the non-differentiable part of the quantization process, ensuring that the model can be properly optimized.

4. Experiments and Discussion of Results

In this section, we first introduce the experimental setup, including the dataset, comparison methods, and experimental details. Finally, the experimental results are presented and discussed.

4.1. Experiment Settings

Experimental environment: The experiments were conducted on a Windows 11 64-bit operating system. The computer was equipped with an Intel i9-12700H processor, an NVIDIA GeForce RTX4060 graphics card, and 8 GB of RAM. The project environment was based on the PyTorch framework and implemented in Python 3.9.
Datasets: The experiments were conducted using the CIFAR10 dataset, consisting of 10 categories of color images, with each category containing 6,000 images, for a total of 60,000 images. In addition, we evaluated our approach on the CIFAR100 dataset and Tiny ImageNet. CIFAR100 comprises 100 categories, with 600 images per category, totaling 60,000 images, offering greater diversity in image classes compared to CIFAR10. Tiny ImageNet consists of 200 categories, with 500 images per category, totaling 100,000 images, providing increased variation in both content and complexity while remaining computationally feasible within the scope of our experimental setup. During training, all images from the datasets were randomly cropped into 256 × 256 patches to ensure consistency across evaluations.
Evaluation metrics: To evaluate image reconstruction quality, we analyzed it from two perspectives, namely, image fidelity and perceptual quality, using the following three key metrics: PSNR, MS-SSIM, and LPIPS.
On the one hand, PSNR and MS-SSIM served as metrics for assessing image fidelity, quantifying the consistency between the reconstructed image and the original image by measuring pixel-wise differences and structural similarity. For PSNR, we used mean squared error (MSE) as the loss function to optimize the model. MSE is defined as the squared difference between the pixel values of the original and reconstructed images, and PSNR is derived by applying a logarithmic transformation to map MSE to a signal-to-noise ratio, expressed in decibels (dB). For MS-SSIM, we used the multi-scale structural similarity (MS-SSIM) index as the loss function to guide model training. In the result analysis, we report the absolute value of MS-SSIM to ensure consistent comparison and analysis with other metrics such as PSNR.
On the other hand, to evaluate the perceptual quality of the generated images, we introduced learned perceptual image patch similarity (LPIPS), a deep learning-based metric. Specifically, for both the original and generated images, we used a pre-trained VGG model to extract multi-level deep features. The perceptual difference between the images was then quantified by calculating the weighted L 2 distance in the feature space. In this study, we used LPIPS only as an evaluation metric to assess the perceptual similarity in image reconstruction quality.
Training U: We used an end-to-end approach to train the model as a whole. The training process was divided into two stages. In the first stage, we trained the model to effectively extract feature representations of the image. Here, we used its variant, Swin Transformer, as the semantic encoding module for the image. The parameter settings are the same as those proposed in [22], namely [ N 1 ; N 2 ] = [ 2 ; 4 ] , [ C 1 ; C 2 ] = [ 128 ; 256 ] , with a window size of 2. In the second stage, we loaded the parameters trained in the first stage and applied the method designed in this paper to process the semantic encoded image features. To reduce computational complexity, we set the feature category m = 3 . We explored the effects of different bit allocation strategies on the reconstructed image quality when the total bit count B was fixed. We use the Adam optimizer with a learning rate of 1 × 10 4 and a batch size of 128.

4.2. Results and Analysis

Through comparative experiments, the performance of the proposed model under different channel conditions and compression ratios was validated. Figure 3 depicts a comparison of image recovery quality on the AWGN channel without considering bandwidth limitations during transmission. “Origin” represents the PSNR performance when no quantization is applied to the features. “Uniform_10” indicates the PSNR value when 10-bit quantization is applied to all features. “Hierarchical_10” represents the PSNR value when all features are divided into three feature levels, and 10-bit quantization is applied. The bit allocation for the three feature levels, based on their contribution factors, is set as follows: 5, 3, and 2. The division of the symbol bits, integer bits, and fractional bits is as follows: 5:131, 3:111, 2:110.
As shown in Figure 4, after quantizing the features, the image recovery quality at the receiver is noticeably affected compared to the case without quantization. As the available bandwidth increases (i.e., as the bit count per feature increases), the impact of quantization gradually diminishes. In both the AWGN and Rayleigh channels, at low signal-to-noise ratios (SNRs), the image recovery quality for both uniform and non-uniform quantization methods is similar. However, in the Rayleigh channel, the non-uniform quantization method slightly underperforms. At high SNR values, the image recovery quality still lags behind that of the unquantized case, but this gap diminishes, possibly due to the image recovery quality falling within an acceptable range.
To comprehensively evaluate the performance of the proposed model, we also tested and trained our model using the MS-SSIM metric. MS-SSIM is a commonly used image quality assessment metric that performs a weighted calculation of brightness, contrast, and structural similarity across multiple scales, providing a more accurate reflection of the perceptual quality of the image. The MS-SSIM performance of the model under different conditions is illustrated in Figure 4.
As presented in Figure 4, with the increase in bandwidth, the MS-SSIM values of different quantization methods gradually increase. The model proposed in this paper performs better in the AWGN channel compared to the Rayleigh channel.
To evaluate the perceptual quality of the image reconstruction, we used the LPIPS metric, which assesses the perceptual similarity between the reconstructed and original images. The results are shown in the figures below.
Figure 5 shows the LPIPS values for image reconstruction of the CIFAR10 dataset under both AWGN and Rayleigh channels using different quantization methods. As the SNR increases, the LPIPS values decrease for all methods, indicating better perceptual quality of the reconstructed images. Both uniform and non-uniform quantization methods have higher LPIPS values compared to the unquantized method, as quantization inevitably affects image quality. However, the LPIPS values for all three methods are below the threshold for high-quality image recovery (0.05). This indicates that the proposed methods in this paper maintain the perceptual quality of the reconstructed images within an acceptable range, despite a slight decrease in quality.
Figure 6 presents the reconstruction results for the CIFAR10 dataset over an AWGN channel with 8-bit transmission. The experimental findings reveal that, under the constraint of limited transmission bits, the proposed non-uniform quantization method outperforms the uniform quantization method across the PSNR, MS-SSIM, and LPIPS metrics.
Specifically, in the PSNR metric (Figure 6a), non-uniform quantization shows significant improvement, particularly at higher SNR values. In terms of the MS-SSIM metric (Figure 6b), the non-uniform quantization method consistently maintains higher values compared to the uniform quantization method, indicating its superior ability to preserve structural details. Additionally, the LPIPS metric (Figure 6c) demonstrates that non-uniform quantization results in fewer perceptual differences from the original image, emphasizing its effectiveness in preserving perceptual quality.
To assess the robustness of our method in real wireless environments, we designed a composite channel model with multipath fading and time-varying noise. This model combines Rician fading ( K = 3 ), which includes a dominant line-of-sight component, and time-varying noise modulated as a sinusoidal waveform ( sin ( t / 10 ) , where t [ 0 , 100 ] ) to simulate dynamic signal-to-noise ratio (SNR) changes. This model reflects more realistic conditions in daily wireless communication compared to simplified AWGN and Rayleigh models. We conducted a small-scale experiment using this real channel model to evaluate the performance of our method. In the experiment, the CIFAR-100 and Tiny ImageNet datasets were tested at SNRs of 4 dB and 10 dB, with SBR values of 1, 4, 7, 10, and 13, with a bit budget of 8 bits.
The results (shown in Figure 7 and Figure 8) show that at low SNR, such as SNR = 1 dB, our method achieves a minimum PSNR of 30 dB, an MS-SSIM of 0.97, and an LPIPS value below 0.2, which indicates good image reconstruction quality. At high SNR, while our method achieves an MS-SSIM of 0.98 and an LPIPS value below 0.014, the PSNR reaches only 33 dB. This PSNR value, although acceptable, is still relatively low, which may result in some loss of fine details and image quality. Therefore, there is still room for improvement in terms of preserving detailed structures and achieving higher-quality image restoration.
Table 1 shows the image reconstruction quality of the Tiny ImageNet dataset under AWGN, Rayleigh, and composite channels with 8-bit transmission.
The results indicate that while there is a decrease in performance in the composite channel compared to the AWGN and Rayleigh channels, the image quality remains within an acceptable range. Specifically, the PSNR value in the composite channel (30.10 dB) is lower than in both the AWGN (33.12 dB) and Rayleigh (32.07 dB) channels, reflecting a slight reduction in reconstruction quality. However, the LPIPS value for the composite channel (0.0133) is very close to that of the AWGN channel (0.0135), suggesting that the perceptual quality of the reconstructed image is still maintained at an acceptable level.

4.3. Computational Overhead Analysis

To assess the computational feasibility of our non-uniform quantization scheme, which incorporates feature classification and bit-allocation logic, we conducted a profiling analysis and compared it to uniform quantization. This evaluation was conducted on the Tiny ImageNet dataset using test images under an AWGN channel (SNR = 10 dB, 8-bit transmission), consistent with the experimental setup used in previous sections. We measured the training time per epoch (epoch = 128), inference time per image, and peak memory usage for both methods.
For training, we profiled the full pipeline, including feature extraction, gradient-based contribution factor computation, bit allocation, and quantization. For inference, we measured the end-to-end process from the input image to the reconstructed output. The results are presented in Table 2. Compared to uniform quantization, our non-uniform approach increases training time by 20% (from 30 minutes to 36 minutes per epoch). Inference time rose by 12% (from 0.175 seconds to 0.195 seconds per image). Peak memory usage increased by 7.6% (from 6.5 GB to 7.0 GB). Despite these increases, our approach remains practical for resource-limited systems, especially when considering the gains in bandwidth efficiency under fixed bit budgets.

4.4. Ablation Studies on Feature Weighting

To evaluate the effectiveness of our gradient-based weighting approach for determining contribution factors, we conducted an ablation study, comparing it with two simpler heuristics, namely, the largest magnitude and the standard deviation of feature values. The experiment was performed on the Tiny ImageNet datasets, using both an AWGN ideal channel (SNR = 10 dB, 8-bit transmission) and the designed composite channel. Reconstruction quality was evaluated using PSNR and LPIPS. As shown in Table 3, our gradient-based method achieved a PSNR of 34.52 dB and an LPIPS of 0.028, outperforming the largest magnitude heuristic (PSNR = 30.25dB, LPIPS = 0.046) and the standard deviation heuristic (PSNR = 32.12 dB, LPIPS = 0.054). Similarly, under the composite channel, our method also outperformed the heuristics in both PSNR and LPIPS, as shown in Table 4. These results demonstrate that our gradient-based approach effectively captures contextual dependencies, leading to better accuracy and perceptual quality compared to the simpler heuristics in both channels.

5. Conclusions

In this paper, we propose a semantic communication system for image transmission, which utilizes a non-uniform quantization method and feature contribution factors to achieve efficient transmission under bandwidth constraints. The core of the system is a non-uniform quantization algorithm that dynamically allocates bit rates based on the contribution of each feature. This approach enhances both image transmission quality and compression performance. Additionally, we present an implementation of this non-uniform quantization system for image transmission, based on the Swin Transformer. To validate the feature importance measurement based on contribution factors, we conducted ablation experiments comparing our method with other approaches. Under limited bandwidth and varying signal-to-noise ratio (SNR) conditions, the experimental results demonstrate that our method outperforms both unquantized and uniform quantization approaches. The proposed quantization framework effectively balances image reconstruction quality and compression efficiency, providing reliable support for semantic communication systems.

6. Future Work

While our current framework focuses on single-user image transmission, the non-uniform quantization approach based on feature contribution factors offers promising extensions to broader scenarios. Below, we outline several directions for future research, which are based on potential extensions of our approach to multi-user systems, multimodal data, and dynamic network environments.
In multi-user scenarios, our method could be extended through a joint optimization strategy. By considering individual user requirements (such as image quality preferences) and network conditions (such as available bandwidth), we can dynamically allocate bits and bandwidth across users. This would ensure fair resource distribution while maintaining transmission efficiency, and could potentially incorporate game-theoretic or fairness-based optimization techniques.
For multimodal data, such as combined image and audio inputs, our approach can evolve by incorporating multimodal feature selection. By jointly assessing the contribution factors of features across modalities, we can optimize quantization bits to prioritize critical information (such as key visual details or audio cues). This would enable efficient cross-modal transmission, particularly suited for tasks like multimedia streaming or multimodal perception.
In dynamic network environments, real-time adaptive bit allocation could be optimized through the integration of reinforcement learning. This would enable the system to adjust quantization thresholds dynamically based on fluctuating channel conditions (e.g., signal-to-noise ratio under AWGN or Rayleigh fading), ensuring a balance between image quality and bandwidth usage. Such an approach would enhance robustness and adaptability in practical deployments.
Another important extension in future work will focus on dynamic bandwidth variations. We plan to explore how fluctuations in available bandwidth can affect the performance of our system. By incorporating real-time bandwidth adaptation, we can further optimize the system’s ability to maintain image quality and transmission efficiency under conditions of varying available bandwidth, simulating more realistic scenarios encountered in mobile or dynamic environments.
These extensions will broaden the applicability of our feature-driven semantic communication framework, addressing diverse real-world challenges. Future work will further explore these directions through both theoretical modeling and experimental validation. In addition, beyond current datasets (CIFAR-10, CIFAR-100, Tiny ImageNet), we also plan to validate our method on larger, more complex datasets like ImageNet and COCO, which include high-resolution images and diverse scenes, to further test its scalability in real-world semantic communication scenarios.

Author Contributions

Conceptualization, methodology, formal analysis, and investigation: J.Z. and Y.Z.; writing—original draft preparation: A.C. and A.L.; writing—review and editing: J.Z., H.X. and B.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Major Science and Technology Projects of Longmen Laboratory (grant no. 231100220400, 231100220300); the Key Scientific Research Project in Colleges and Universities of Henan Province of China (grant no. 21A510003); the Key Science and Technology Research Project of Henan Province of China (grant no. 222102210053); and the Open Found of the Engineering Research Center of Intelligent Swarm Systems, Ministry of Education (grant no. ZZU-CISS-2024003).

Institutional Review Board Statement

Not applicable

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. You, X.H. Shannon theory and future 6G’s technique potentials. Sci. Sin. Inf. 2020, 50, 1377–1394. (In Chinese) [Google Scholar] [CrossRef]
  2. Attaran, M. The impact of 5G on the evolution of intelligent automation and industry digitization. J. Ambient Intell. Hum. Comput. 2023, 14, 5977–5993. [Google Scholar] [CrossRef]
  3. Zhang, Y.M.; Wang, F.Y.; Xu, W.J.; Liu, C.Y. Intelligence Semantic Communications: A New Paradigm for Networked. In Proceedings of the 2022 IEEE 32nd International Workshop on Machine Learning for Signal Processing (MLSP), Xi’an, China, 22–25 August 2022. [Google Scholar]
  4. Weng, Z.; Qin, Z. Robust Semantic Communications for Speech Transmission. arXiv 2024, arXiv:2403.05187. [Google Scholar]
  5. Xie, H.; Qin, Z.; Li, G.Y. Task-Oriented Multi-User Semantic Communications for VQA. IEEE Wirel. Commun. Lett. 2022, 11, 553–557. [Google Scholar] [CrossRef]
  6. Hu, Q.; Zhang, G.; Qin, Z.; Cai, Y.; Yu, G.; Li, G.Y. Robust Semantic Communications Against Semantic Noise. In Proceedings of the 2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall), London, UK, 26–29 September 2022. [Google Scholar]
  7. Hu, Q.; Zhang, G.; Qin, Z.; Cai, Y.; Yu, G.; Li, G.Y. Robust Semantic Communications With Masked VQ-VAE Enabled Codebook. IEEE Trans. Wirel. Commun. 2023, 22, 8707–8722. [Google Scholar] [CrossRef]
  8. Xu, W.; Zhang, Y.; Wang, F.; Qin, Z.; Liu, C.; Zhang, P. Semantic Communication for the Internet of Vehicles:A Multiuser Cooperative Approach. IEEE Veh. Technol. Mag. 2023, 18, 100–109. [Google Scholar] [CrossRef]
  9. Tonchev, K.; Bozhilov, I.; Manolova, A. Semantic Communication System for 3D Video. In Proceedings of the 2023 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT and NCON), Phuket, Thailand, 22–25 March 2023. [Google Scholar]
  10. Bourtsoulatze, E.; Burth Kurka, D.; Gündüz, D. Deep Joint Source-Channel Coding for Wireless Image Transmission. IEEE Trans. Cogn. Commun. Netw. 2019, 5, 567–579. [Google Scholar] [CrossRef]
  11. Yang, K.; Wang, S.; Dai, J.; Tan, K.; Niu, K.; Zhang, P. WITT: A Wireless Image Transmission Transformer for Semantic Communications. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023. [Google Scholar]
  12. Lin, S.; Liu, F.; Guo, C.; Tong, W. Research on Semantic Coding Algorithm Based on Semantic Importance. J. Beijing Univ. Posts Telecommun. 2024, 47, 10–16. [Google Scholar]
  13. Huang, D.; Tao, X.; Gao, F.; Lu, J. Deep Learning-Based Image Semantic Coding for Semantic Communications. In Proceedings of the 2021 IEEE Global Communications Conference (GLOBECOM), Madrid, Spain, 7–11 December 2021. [Google Scholar]
  14. Wu, J.L.; Wu, C.; Lin, Y.F.; Yoshinaga, T.; Zhong, L.; Chen, X.F.; Ji, Y.S. Semantic segmentation-based semantic communication system for image transmission. Digit. Commun. Netw. 2024, 10, 519–527. [Google Scholar] [CrossRef]
  15. Gupta, K.; Asthana, A. Reducing the Side-Effects of Oscillations in Training of Quantized YOLO Networks. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Steamboat Springs, CO, USA, 24–26 March 2014; Volume 11, pp. 2440–2449. [Google Scholar]
  16. Han, S.; Mao, H.; Dally, W.J. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv 2016, arXiv:1510.00149v5. [Google Scholar]
  17. Lei, L.; Hong, X.; Shi, J.; Su, M.; Lin, C.; Xia, W. Quantization-Based Adaptive Deep Image Compression Using Semantic Information. IEEE Access 2023, 11, 118061–118077. [Google Scholar]
  18. Tsumura, K.; Maciejowski, J. Optimal Quantization of Signals for System Identification. In Proceedings of the 2003 European Control Conference (ECC), Cambridge, UK, 1–4 September 2003. [Google Scholar]
  19. Xia, X.; Sun, G.; Zhang, K.; Wu, S.; Wang, T.; Xia, L.; Liu, S. NanoSats/CubeSats ADCS Survey. In Proceedings of the 2017 29th Chinese Control And Decision Conference (CCDC), Chongqing, China, 28–30 May 2017. [Google Scholar]
  20. Wallace, G.K. The JPEG still picture compression standard. IEEE Trans. Consum. Electron. 1992, 38, xviii–xxxiv. [Google Scholar]
  21. Lin, J.; Tang, J.; Tang, H.; Yang, S.; Chen, W.-M.; Wang, W.-C.; Xiao, G.; Dang, X.; Gan, C.; Han, S. AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration. arXiv 2024, arXiv:2306.00978. [Google Scholar]
  22. Jacob, B.; Kligys, S.; Chen, B.; Zhu, M.; Tang, M.; Howard, A.; Adam, H.; Kalenichenko, D. Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  23. Liu, Z.; Wu, B.; Luo, W.; Yang, X.; Liu, W.; Cheng, K.-T. Bi-Real Net: Enhancing the Performance of 1-bit CNNs with Improved Representational Capability and Advanced Training Algorithm. arXiv 2018, arXiv:1808.00278. [Google Scholar]
  24. Srikotr, T.; Mano, K. Sub-band Vector Quantized Variational AutoEncoder for Spectral Envelope Quantization. In Proceedings of the TENCON 2019—2019 IEEE Region 10 Conference (TENCON), Kochi, India, 17–20 October 2019. [Google Scholar]
  25. Sun, J.H. Mixed-precision quantization technology of convolutional neural networks. Inf. Technol. 2020, 6, 66–69. [Google Scholar]
  26. Chen, Y.; Cai, X.; Liang, X.; Wang, M. Compression algorithm for weights quantized deep neural network models. J. Xidian Univ. 2019, 46, 132–138. [Google Scholar]
  27. Dong, Z.; Yao, Z.; Gholami, A.; Mahoney, M.; Keutzer, K. HAWQ: Hessian AWare Quantization of Neural Networks With Mixed-Precision. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  28. Zhang, C.; Zuo, G.; Zheng, Z.; Zhang, W.; Rao, Y.; Jiang, Z. Quantizing Separable Convolution of MobileNets with Mixed Precision. J. Electron. Imaging 2024, 33, 013013. [Google Scholar] [CrossRef]
  29. Uhlich, S.; Mauch, L.; Kazuki, Y.; Fabien, C.; Javier, A.G.; Stephen, T.; Thomas, K.; Akita, N. Differentiable Quantization of Deep Neural Networks. arXiv 2019, arXiv:1905.11452v1. [Google Scholar]
  30. Wu, B.C.; Wang, Y.H.; Zhang, P.Z.; Tian, Y.D.; Vajda, P.; Kurt, K. Mixed Precision Quantization of ConvNets via Differentiable Neural Architecture Search. arXiv 2018, arXiv:1812.00090. [Google Scholar]
  31. Wang, T.Z.; Wang, K.; Cai, H.; Lin, J.; Liu, Z.J.; Han, S. APQ: Joint Search for Network Architecture, Pruning and Quantization Policy. arXiv 2020, arXiv:2006.08509. [Google Scholar]
  32. van Baalen, M.; Louizos, C.; Nagel, M.; Amjad, R.A.; Wang, Y.; Blankevoort, T.; Welling, M. Bayesian Bits: Unifying Quantization and Pruning. arXiv 2020, arXiv:2005.07093. [Google Scholar]
  33. Wang, K.; Liu, Z.J.; Lin, Y.J.; Han, S. HAQ: Hardware-Aware Automated Quantization. arXiv 2018, arXiv:1811.08886. [Google Scholar]
  34. Elthakeb, A.T.; Pilligundla, P.; Mireshghallah, F.; Yazdanbakhsh, A.; Esmaeilzadeh, H. ReLeQ: A Reinforcement Learning Approach for Deep Quantization. arXiv 2018, arXiv:1811.01704. [Google Scholar]
  35. Lou, Q.; Guo, F.; Liu, L.T.; Kim, M.J.; Lei, J. AutoQ: Automated Kernel-Wise Neural Network Quantization. arXiv 2019, arXiv:1902.05960. [Google Scholar]
  36. Weaver, W. Recent contributions to the mathematical theory of communication. In ETC: A Review of General Semantics; Institute of General Semantics: Fort Worth, TX, USA, 1953; pp. 261–281. [Google Scholar]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  38. Yang, W.; Du, H.; Liew, Z.Q.; Lim, W.Y.B.; Xiong, Z.; Niyato, D.; Chi, X.; Shen, X.; Miao, C. Semantic Communications for Future Internet: Fundamentals, Applications, and Challenges. IEEE Commun. Surv. Tutor. 2023, 25, 213–250. [Google Scholar]
  39. Rothstein, J. Review of “Science and Information Theory” by Leon Brillouin. Science 1956, 124, 492–493. [Google Scholar]
  40. Sudan, M. Communication Amid Uncertainty. In Proceedings of the 2012 IEEE Information Theory Workshop, Lausanne, Switzerland, 3–7 September 2012. [Google Scholar]
  41. O’Shea, T.J.; Karra, K.; Clancy, T.C. Learning to Communicate:Channel Auto-Encoders, Domain Specific Regularizers, and Attention. In Proceedings of the 2016 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Limassol, Cyprus, 12–14 December 2016. [Google Scholar]
  42. Shi, G.; Xiao, Y.; Li, Y.; Xie, X. From Semantic Communication to Semantic-Aware Networking: Model, Architecture, and Open Problems. IEEE Commun. Mag. 2021, 59, 44–50. [Google Scholar]
  43. Feng, Y.; Xu, J.; Liang, C.; Yu, G.; Hu, L.; Yuan, T. Decoupling Source and Semantic Encoding: An Implementation Study. Electronics 2023, 12, 2755. [Google Scholar] [CrossRef]
  44. Xie, H.; Qin, Z. A Lite Distributed Semantic Communication System for Internet of Things. IEEE J. Sel. Areas Commun. 2021, 39, 142–153. [Google Scholar]
  45. Xu, X.; Xiong, H.; Wang, Y.; Che, Y.; Han, S.; Wang, B.; Zhang, P. Knowledge-Enhanced Semantic Communication System with OFDM Transmissions. Sci. China Inf. Sci. 2023, 66, 172302. [Google Scholar]
  46. Zhang, W.; Zhang, H.; Ma, H.; Shao, H.; Wang, N.; Leung, V.C.M. Predictive and Adaptive Deep Coding for Wireless Image Transmission in Semantic Communication. IEEE Trans. Wirel. Commun. 2023, 22, 5486–5501. [Google Scholar]
  47. Qi, Z.; Feng, Y.; Qin, Z. Adaptive Sampling and Joint Semantic-Channel Coding under Dynamic Channel Environment. arXiv 2025, arXiv:2502.072362025. [Google Scholar]
  48. Zhou, K.Q.; Zhang, G.Y.; Cai, Y.L.; Hu, Q.Y.; Yu, G.D. FAST Feature Arrangement for SemanticTransmission. arXiv 2019, arXiv:2305.03274. [Google Scholar]
  49. Gao, S.; Qin, X.Q.; Chen, L.; Han, K.F.; Zhang, P. Importance of Semantic Information Based on Semantic Value. IEEE Trans. Commun. 2024, 72, 5443–5457. [Google Scholar]
  50. Ma, Y.; Xu, C.M.; Liu, Z.Y.; Zhang, S.Q.; Tafazolli, R. Importance-Aware Source-Channel Coding for Multi-ModalTask-Oriented Semantic Communication. arXiv 2025, arXiv:2502.16194. [Google Scholar]
Figure 1. The overall architecture diagram of the proposed non-uniform quantization semantic communication system.
Figure 1. The overall architecture diagram of the proposed non-uniform quantization semantic communication system.
Entropy 27 00369 g001
Figure 2. The overall architecture of the proposed non-uniform quantization.
Figure 2. The overall architecture of the proposed non-uniform quantization.
Entropy 27 00369 g002
Figure 3. (a,b) Describes the image reconstruction PSNR values and MS-SSIM for the CIFAR10 dataset under the AWGN channel, with features transmitted using 10-bit uniform quantization, 6-bit uniform quantization, and no feature compression. (c,d) Show the PSNR values for image reconstruction of the CIFAR10 dataset under the Rayleigh channel.
Figure 3. (a,b) Describes the image reconstruction PSNR values and MS-SSIM for the CIFAR10 dataset under the AWGN channel, with features transmitted using 10-bit uniform quantization, 6-bit uniform quantization, and no feature compression. (c,d) Show the PSNR values for image reconstruction of the CIFAR10 dataset under the Rayleigh channel.
Entropy 27 00369 g003aEntropy 27 00369 g003b
Figure 4. (a,b) Describe the image reconstruction MS-SSIM values for the CIFAR10 dataset under the AWGN channel, with features transmitted using 10-bit uniform quantization, 6-bit uniform quantization, and no feature compression. (c,d) Describe the MS-SSIM values for image reconstruction of the CIFAR10 dataset under the Rayleigh channel.
Figure 4. (a,b) Describe the image reconstruction MS-SSIM values for the CIFAR10 dataset under the AWGN channel, with features transmitted using 10-bit uniform quantization, 6-bit uniform quantization, and no feature compression. (c,d) Describe the MS-SSIM values for image reconstruction of the CIFAR10 dataset under the Rayleigh channel.
Entropy 27 00369 g004aEntropy 27 00369 g004b
Figure 5. (a) Shows the LPIPS values for image reconstruction of the CIFAR10 dataset under the AWGN channel with no bandwidth limitations, comparing three methods: 10-bit uniform quantization for each feature, 10-bit non-uniform quantization, and no feature quantization. (b) Presents the LPIPS values for the CIFAR10 dataset under the AWGN channel with no bandwidth limitations, comparing the same three methods: 6-bit uniform quantization, 6-bit non-uniform quantization, and no feature quantization. (c) Describes the LPIPS values for image reconstruction of the CIFAR10 dataset under the Rayleigh channel with no bandwidth limitations, comparing 10-bit uniform quantization for each feature, 10-bit non-uniform quantization, and no feature quantization. (d) Shows the LPIPS values for the CIFAR10 dataset under the Rayleigh channel with no bandwidth limitations, comparing 6-bit uniform quantization, 6-bit non-uniform quantization, and no feature quantization.
Figure 5. (a) Shows the LPIPS values for image reconstruction of the CIFAR10 dataset under the AWGN channel with no bandwidth limitations, comparing three methods: 10-bit uniform quantization for each feature, 10-bit non-uniform quantization, and no feature quantization. (b) Presents the LPIPS values for the CIFAR10 dataset under the AWGN channel with no bandwidth limitations, comparing the same three methods: 6-bit uniform quantization, 6-bit non-uniform quantization, and no feature quantization. (c) Describes the LPIPS values for image reconstruction of the CIFAR10 dataset under the Rayleigh channel with no bandwidth limitations, comparing 10-bit uniform quantization for each feature, 10-bit non-uniform quantization, and no feature quantization. (d) Shows the LPIPS values for the CIFAR10 dataset under the Rayleigh channel with no bandwidth limitations, comparing 6-bit uniform quantization, 6-bit non-uniform quantization, and no feature quantization.
Entropy 27 00369 g005
Figure 6. (a) Describes the relationship between PSNR and SNR under different quantization methods (Uniform and Non_uniform) in an AWGN channel with an 8-bit transmission limit. (b) Describes the relationship between MS-SSIM and SNR under different quantization methods (Uniform and Non_uniform) in an AWGN channel with an 8-bit transmission limit. (c) Shows the effect of 8-bit transmission on CIFAR10 images in an AWGN channel in terms of LPIPS.
Figure 6. (a) Describes the relationship between PSNR and SNR under different quantization methods (Uniform and Non_uniform) in an AWGN channel with an 8-bit transmission limit. (b) Describes the relationship between MS-SSIM and SNR under different quantization methods (Uniform and Non_uniform) in an AWGN channel with an 8-bit transmission limit. (c) Shows the effect of 8-bit transmission on CIFAR10 images in an AWGN channel in terms of LPIPS.
Entropy 27 00369 g006
Figure 7. (a) Describes the relationship between PSNR and SNR for the CIFAR100 dataset under the composite channel with 8-bit transmission. (b) Describes the relationship between MS-SSIM and SNR for the CIFAR100 dataset under the composite channel with 8-bit transmission. (c) Describes the relationship between PSNR and SNR for the Tiny ImageNet dataset under the composite channel with 8-bit transmission. (d) Describes the relationship between MS-SSIM and SNR for the Tiny ImageNet dataset under the composite channel with 8-bit transmission.
Figure 7. (a) Describes the relationship between PSNR and SNR for the CIFAR100 dataset under the composite channel with 8-bit transmission. (b) Describes the relationship between MS-SSIM and SNR for the CIFAR100 dataset under the composite channel with 8-bit transmission. (c) Describes the relationship between PSNR and SNR for the Tiny ImageNet dataset under the composite channel with 8-bit transmission. (d) Describes the relationship between MS-SSIM and SNR for the Tiny ImageNet dataset under the composite channel with 8-bit transmission.
Entropy 27 00369 g007aEntropy 27 00369 g007b
Figure 8. (a) Describes the relationship between LPIPS for the CIFAR100 dataset under the composite channel with 8-bit transmission. (b) Describes the relationship between LPIPS for the Tiny ImageNet dataset under the composite channel with 8-bit transmission.
Figure 8. (a) Describes the relationship between LPIPS for the CIFAR100 dataset under the composite channel with 8-bit transmission. (b) Describes the relationship between LPIPS for the Tiny ImageNet dataset under the composite channel with 8-bit transmission.
Entropy 27 00369 g008
Table 1. Performance across channels (SNR = 4 dB).
Table 1. Performance across channels (SNR = 4 dB).
ChannelPSNR (dB)LPIPSMS-SSIM (dB)
AWGN33.120.01350.982
Rayleigh32.070.0210.984
Composite30.100.01330.981
Table 2. Computational overhead comparison (Tiny ImageNet, SNR = 10 dB).
Table 2. Computational overhead comparison (Tiny ImageNet, SNR = 10 dB).
MethodTraining Time (min/epoch)Inference Time (s/image)Peak Memory (GB)
Uniform Quantization300.1756.5
Non-Uniform Quantization (Ours)360.1957.0
Table 3. Ablation study results (SNR = 10 dB, Tiny ImageNet, AWGN channel).
Table 3. Ablation study results (SNR = 10 dB, Tiny ImageNet, AWGN channel).
Weighting MethodPSNR (dB)LPIPSMS-SSIMInference Time (s)
Gradient-based (Ours)34.520.0280.9750.195
Largest magnitude30.250.0460.9250.143
Standard deviation32.120.0540.9300.162
Table 4. Ablation study results (SNR = 10 dB, Tiny ImageNet, composite channel).
Table 4. Ablation study results (SNR = 10 dB, Tiny ImageNet, composite channel).
Weighting MethodPSNR (dB)LPIPSMS-SSIMInference Time (s)
Gradient-based (Ours)32.450.0350.9600.215
Largest magnitude28.550.0620.9100.175
Standard deviation30.470.0700.9200.190
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Zhang, Y.; Ji, B.; Chen, A.; Liu, A.; Xu, H. Feature-Driven Semantic Communication for Efficient Image Transmission. Entropy 2025, 27, 369. https://doi.org/10.3390/e27040369

AMA Style

Zhang J, Zhang Y, Ji B, Chen A, Liu A, Xu H. Feature-Driven Semantic Communication for Efficient Image Transmission. Entropy. 2025; 27(4):369. https://doi.org/10.3390/e27040369

Chicago/Turabian Style

Zhang, Ji, Ying Zhang, Baofeng Ji, Anmin Chen, Aoxue Liu, and Hengzhou Xu. 2025. "Feature-Driven Semantic Communication for Efficient Image Transmission" Entropy 27, no. 4: 369. https://doi.org/10.3390/e27040369

APA Style

Zhang, J., Zhang, Y., Ji, B., Chen, A., Liu, A., & Xu, H. (2025). Feature-Driven Semantic Communication for Efficient Image Transmission. Entropy, 27(4), 369. https://doi.org/10.3390/e27040369

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop