Next Article in Journal
Estimation of Wind Turbine Heights with Shadows Using Gaofen-2 Satellite Imagery
Next Article in Special Issue
Toward Adversarial Robustness Network Intrusion Detection Based on Multi-Model Ensemble Approach
Previous Article in Journal
Hyperspectral Target Tracking via Spatial–Spectral Attention Weight Variance Gradient and Depth Contrast Enhancement
Previous Article in Special Issue
Lightweight ECC-Based Self-Healing Federated Learning Framework for Secure IIoT Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Steganalysis Network for Weak Steganographic Signal Extraction and Enhancement

School of Computer, Electronics and Information, Guangxi University, Nanning 530000, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(4), 1329; https://doi.org/10.3390/s26041329
Submission received: 9 January 2026 / Revised: 5 February 2026 / Accepted: 13 February 2026 / Published: 19 February 2026
(This article belongs to the Special Issue Advances and Challenges in Sensor Security Systems)

Abstract

The purpose of digital image steganalysis is to identify the signal embedded in the natural image by steganography. In the spatial domain, this embedded signal only modifies the image value of the natural image by ± 1 , and this modification is weak. However, most of the existing convolutional neural networks use popular components to design or optimize the network structure, without deeply exploring the network’s ability to recognize such weak modifications. In order to deal with this problem, we propose a novel preprocessing structure, the learnable filter constrained by high-pass prior (LFCHP), to improve the network’s ability to extract weak embedded signals in the preprocessing stage, as well as a second-order signal auxiliary branch (SSAB) to reduce the suppression of weak embedded signals during convolution stacking, and a new pooling method, SoftPool, to reduce the loss of weak embedded signals during downsampling. Combining these three structures, we propose a steganalysis network, WSERNet, for weak steganographic signal extraction and enhancement. Experiments conducted under identical conditions demonstrate that the proposed method achieves an accuracy improvement of 1.08–2.96% over state-of-the-art spatial-domain steganalysis algorithms across three steganographic schemes at four embedding rates, and exhibits excellent generalization capabilities across different steganography techniques.

1. Introduction

Image steganography is an algorithm that hides information within the redundancy of an image, making it imperceptible to the human eye [1,2,3]. In contrast, image steganalysis is a technique used to detect whether embedded information exists within an image, thus preventing the dissemination of illicit information [4].
Unlike encrypted communication methods [5], existing adaptive image steganography algorithms embed secret information into complex regions of images that are difficult to detect [6,7]. Traditional non-deep learning image steganalysis algorithms usually require hand-designed high-pass filters to extract embedding residuals [8,9,10], but hand-designed filters have limitations [11,12,13,14]. Different from traditional image steganalysis algorithms, image steganalysis algorithms based on deep learning not only reduce dependence on domain knowledge, but also achieve simultaneous optimization of feature extractors and feature classifiers, so more and more image steganalysis algorithms based on deep learning are emerging [15,16,17].
In image steganalysis algorithms, the preprocessing layer is capable of computing noise residuals in complex regions of images, thereby enhancing network performance and accelerating network convergence [18]. Therefore, the preprocessing layer is often employed as a core component in image steganalysis [16,17,19]. However, existing preprocessing layers have their limitations. A typical preprocessing layer design involves using a set of filters from SRM as fixed-parameter convolutional kernels. However, constrained by the limitations of manual design, fixed convolutional kernel parameters are often suboptimal [20]. To mitigate the drawbacks of fixed convolutional kernels, researchers [21,22] have proposed initializing fixed kernels as parameters and optimizing them during network training. However, trainable parameter initialization of filter kernels tends to get trapped in local optima during training, resulting in subpar training outcomes. In order to break free from the dependency and limitations of manual priors, researchers have developed methods [20,23] for completely random initialization. These methods can automatically search for optimal parameters, avoiding the potential harm of manually designed filter kernels to the overall network. However, achieving the ideal scenario of automatically searching for optimal parameters without introducing prior knowledge may result in the network not effectively focusing on the complex areas of image secret message embedding.
Furthermore, in order to ensure the concealment of information in image steganography, the embedded signal only modifies the pixel value of the natural image by ± 1 , which is very weak. However, both convolution and pooling operations inherently suppress weak signals [24,25], thus potentially weakening the features of these faint steganographic signals during the successive convolutional layers [22,26]. Additionally, in steganalysis, the most commonly used pooling methods are average pooling and global average pooling applied after convolutional layers. However, average pooling assigns equal weights to elements within the kernel neighborhood of feature maps, failing to reinforce the importance of key element values in the feature maps and potentially weakening the features of embedded signals during pooling. Although max pooling may lose significant information and is less commonly used in steganalysis, it retains representative results [27] and should be utilized.
To address the limitations of the aforementioned steganalysis methods, aiming to enhance preprocessing performance and mitigate the adverse effects of convolution and pooling on weak signals, we propose a steganalysis network called WSERNet. Specifically, the contributions of our proposed method are summarized as follows:
  • We design a learnable preprocessing structure with high-pass prior constraints. This structure initializes the parameters of the filtering kernel randomly while introducing high-pass prior knowledge through the addition of a high-pass constraint branch. This allows the new network to automatically search for optimal parameters during training while focusing on the complex areas of secret message embedding in images through high-pass prior constraints.
  • We introduce a novel auxiliary branch for extracting second-order signals to capture different-order signals, effectively enhancing the features of weak steganographic embedding signals. By utilizing the second-order signal extraction auxiliary branch at multiple levels, we can strengthen the intensity of weak signals, thereby alleviating the suppression of weak signals by convolution. Moreover, these second-order signal auxiliary branches can be removed during inference to avoid complex computations and save time.
  • We employ the SoftPool pooling method [27] and extend it to a new global soft pooling (GSP) method. SoftPool pooling can utilize all elements in the kernel neighborhood while assigning greater weights to representative elements, thereby preventing excessive weakening of features of weak embedding signals during pooling. GSP can diversify the compression of features extracted by convolutional layers, better capturing information from feature maps.
  • Experimental results demonstrate that WSERNet outperforms the state-of-the-art spatial domain image steganalysis networks in terms of performance and exhibits superior generalization performance across different steganography techniques.
The remaining part of this paper is organized as follows: In Section 2, we review closely related work in spatial domain steganalysis. In Section 3, we provide detailed explanations of the proposed WSERNet. Section 4 validates the effectiveness of the proposed method through extensive experiments. Finally, Section 5 presents the conclusion of this paper and outlines future work.

2. Related Works

Tan et al. [15] first applied deep learning methods to image steganalysis. Their network consisted of four convolutional layers, with the first layer parameters multiplied by the KV filter and trained using stacked convolutional autoencoders. Qian et al. [16] proposed the Gaussian-Neuron CNN (GNCNN), which combines a preprocessing layer with convolutional layers. Unlike Tan et al., GNCNN directly employs fixed KV filters as the preprocessing layer and then stacks five convolutional layers. To adapt to modifications in pixel values of stego images, GNCNN introduces a new activation function, the Gaussian activation function, in the convolutional layers to replace ReLU. Xu et al. [17] also introduced Xu-Net, utilizing KV filter kernels and five convolutional layers, but with more detailed designs on the convolutional layers. They applied TanH activation functions in the first and second convolutional layers to constrain activation values and added an absolute layer after the first convolutional layer for subsequent modeling. The last three convolutional layers used convolution with kernel size to limit modeling intensity. Xu-Net achieved performance comparable to the traditional steganalysis method SRM [9]. Ye et al. [18] proposed Ye-Net, an alternative CNN-based method, referring to the three steps of residual calculation, feature extraction, and binary classification in traditional steganalysis methods. Ye-Net widened the convolution in the preprocessing layer from using 1 filter in SRM to 30 filters and introduced a new activation function, the truncated linear unit (TLU), to adapt to the characteristics of embedded signals in the range. Additionally, they designed a method for integrating Selection Channel into the CNN network. Results showed that Ye-Net outperformed traditional methods such as SRM [9] and maxSRM [28] in performance. Boroumand et al. [20] argued that the optimal filter should be learned rather than handcrafted, leading to the proposal of SRNet. SRNet initializes all convolutional kernel weight parameters randomly to avoid manual intervention, disables pooling in shallow convolutions to prevent signal suppression, and adds residual connections to improve detection accuracy. SRNet achieved significant improvements in both spatial and JPEG domains. Zhang et al. [21] introduced a new CNN architecture, Zhu-Net, employing various techniques. Zhu-Net standardized the sizes of 30 convolutional kernels in the preprocessing layer and initialized them using the basic filters from SRM [9]. During training, the preprocessing kernels were also optimized. Additionally, Zhu-Net utilized two depthwise separable convolution blocks in the shallow layers and applied spatial pyramid pooling [29] after the convolutional layers to compress feature maps.Deng et al. [19] proposed a Covariance Pooling Steganalysis Network (CPSN). CPSN designed a fast and effective network architecture through extensive experiments and further improved network performance using covariance pooling [30], achieving better detection results than SRNet. You et al. [31] introduced SiaStegNet, capable of detecting images of arbitrary sizes. SiaStegNet divided the input image into two subregions, each inputted into one of the two subnetworks of Siamese [32]. The outputs of each subnetwork were concatenated to distinguish between cover and stego results. Moreover, the outputs of each subnetwork were compared using contrastive loss [33] to extract the similarity between the two subregions, adapting better to images of arbitrary sizes. Liu et al. [22] explored methods to reduce the suppression of weak signals embedded during convolution and proposed FPNet. FPNet introduced the efficient significant expansion front part (FEPM) to enhance the transmission of weak signals and designed the attention downsampling module (ADM) to mitigate the suppression of weak signals by pooling. Weng et al. [34] presented the lightweight LWENet, which utilized 30 fixed filter kernels from SRM, followed by bottleneck residual blocks (BRBs) to further improve the residual signal-to-noise ratio. LWENet also employed multi-view global pooling (MGP) and depthwise separable convolution (DWSConv) techniques to reduce parameter count while enhancing detection performance.

3. Materials and Methods

To enhance the accuracy of image steganalysis by effectively identifying weak embedded signals, this paper proposes a convolutional neural network named “WSERNet.”

3.1. Overview of Method

As shown in Figure 1, the proposed WSERNet consists of three parts: (1) the preprocessing part, represented by the LFCHP (learnable filters constrained by high-pass prior) module in Figure 1, composed of a high-pass prior constraint branch (HPCB) and a learnable filter bank (LFB), responsible for preprocessing the input image; (2) the Feature Extraction part, comprising one Block 1 module, four FEMs (Feature Enhancement Modules), one SoftPool, and two AvgPool; (3) the classifier, composed of GSP (Global SoftPool), BN (Batch Normalization); and a fully connected layer (FC). The network structure of WSERNet differs between the training and inference stages. In the training stage, LFCHP includes HPCB, and each FEM contains a second-order signal auxiliary branch (SSAB). In the inference stage, neither HPCB nor SSAB affects the results of the main network; thus, these branch structures are removed.

3.2. The LFCHP Module: Learnable Filters with High-Pass Prior Constraint

In the preprocessing stage, existing methods usually use manually designed filter kernels or randomly initialize the convolution kernel. Manually designed filter kernels can strictly focus on complex areas where information is embedded, but manual experience cannot represent all situations, and the feature extraction is not completely enough. The method of randomly initializing the convolution kernel automatically searches for the optimal parameters in theory, but the CNN pays more attention to mid-frequency and low-frequency information, and does not pay enough attention to high-frequency information [35], that is, the complex area where information is embedded. In order to combine the advantages of the two methods, we propose learnable filters constrained by high-pass prior, as shown in the LFCHP part in Figure 1, which consists of two parts: LFB and HPCB.
The LFB part consists of three CBT (c) modules, where c denotes the number of output channels. The parameters of the convolutional layers in each CBT module are randomly initialized, allowing each convolutional kernel to automatically search for the optimal parameters. The output channel numbers of the three CBT modules are 64, 32, and 32. The first convolutional module with 64 channels is used to extract more diversified features directly from the input image. For spatial image steganography, embedding secret information is achieved by modifying the grayscale values of the carrier image pixels. In this case, using the ReLU activation function may lead to the loss of a considerable amount of information. Therefore, we choose the hyperbolic tangent function (TanH) as the activation function to better adapt to the characteristics of the steganographic embedding signal.
In the HPCB part, adaptive steganography embeds secret information into the complex high-frequency areas of images. Therefore, we need to encourage the steganalysis network to focus on high-frequency information. The loss function enables the network to learn to target, while the high-pass filter can strictly focus on the high-frequency area, but the high-pass filter is difficult to use as a direct learning target. Therefore, we use the output image of the high-pass filter as the learning target; that is, F r learns from F g . The high-pass filter uses the classic setting [19]. Specifically, we use 30 filters from SRM as an additional branch, with each filter normalized and the filter weights fixed. Then, a linear truncation activation function (TLU) with a threshold of 3 is used to filter out large-value elements with low signal-to-noise ratios. After TLU, an output F g is formed with a channel number of 30, serving as one input to Equation (1). Additionally, the number of output channels of LFB is 32, which is inconsistent with the target number of 30 channels. Therefore, the number of channels needs to be adjusted; that is, a convolution layer branch with a convolution kernel size of 1 and an output channel number of 30 is introduced after the LFB to form the output F r , which is another input to Equation (1). This convolutional layer does not have batch normalization (BN) and activation functions. In addition to adjusting the number of channels, adding this convolutional layer branch also allows each channel of the third CBT module to learn from all high-pass filters, so that the convolution kernel in the LFB has richer prior knowledge. Finally, F r and F g are entered into the mean square error loss function L M S E , and F g is constrained to F r by L M S E ; that is, F r learns from F g , where the eigengraph size of F r and F g is M · N , and the loss function L M S E is defined as
L M S E = 1 M · N i = 1 M i = 1 N ( F r ( i , j ) F g ( i , j ) ) 2

3.3. Feature Enhancement Module

Since the embedding signal is distributed in the high-frequency area of the image, this will cause the weak embedding signal to disappear during the convolution stacking process. Therefore, we need to maintain attention to the high-frequency area in the network. The greater the difference in adjacent element values of the feature map, the more dispersed the distribution of element values will naturally be, and the calculated element value variance will usually be higher. Then, using variance and covariance in the network will make the network pay more attention to high-frequency areas. GCP can calculate the variance of the feature map itself and the covariance between different feature maps, that is, the second-order information of the feature map. Therefore, using GCP can enable the network to maintain focus on high-frequency areas, thus avoiding the problem of weak embedded signals disappearing during the convolution stacking process. We drew on the usage of GCP in the image classification method Detachable Second-Order Pooling (DSOP) [36], proposed a method called SSAB, and built a weak signal feature enhancement module (FEM).
Each FEM comprises a Block 1/2 and an SSAB branch. Block 1/2 indicates the option to choose between Block 1 and Block 2 (Block 1 is used when the input and output channel numbers are equal, and Block 2 is used otherwise).
In image steganalysis tasks, downsampling is typically disabled in shallow convolutions to avoid signal suppression. Moreover, due to computational constraints, the channel numbers of shallow feature maps cannot be excessive. These factors often result in shallow feature maps characterized by high resolution and low channel numbers. Based on this feature, we improved the structure of DSOP and named it SSAB, as illustrated by the SSAB of FEM in Figure 1.
SSAB initially reduces the feature map size of the input with a size of H × W × c using N CBR (Conv+BN+ReLU) and AvgPool layers until the output feature map size becomes 32 × 32 , i.e., ( H / 2 N ) × ( W / 2 N ) = 32 × 32 . Therefore, the sizes of N in the 4 FEMs are 3, 2, 1, and 0 in order. Subsequently, the integrated average-pooled information is processed by a CBR. The channel numbers remain unchanged throughout these processes. After dimension reduction of the feature map size, global covariance pooling (GCP) is employed to extract global second-order information, preserving only the channel dimension. Finally, classification and training are performed through a fully connected layer. Specifically, each SSAB branch is equipped with a classifier for convenient training of SSAB.
Block 1 consists of a CBR, Convolution, BN, residual connection, and ReLU. In contrast to Block 1, the residual connection part of Block 2 contains Convolution and BN for adjusting the channel number.
We apply FEM hierarchically in the network, reinforcing weak signal features in multiple stages to effectively prevent signal loss. Moreover, all SSAB branches are removed during the inference phase, avoiding an increase in computational costs during actual network applications.
The loss function used for each FEM during training is cross-entropy loss ( L C E , n , n = 1 , 2 , 3 , 4 ). The total loss function L F E M across the four FEMs is defined by the following Equation (2):
L F E M = i = 1 4 L C E , n
Each L C E , n corresponds to a different loss function in the different FEMs.
The loss function used for the classifiers is cross-entropy loss L C E , c l f . Since both the FEMs and classifiers use cross-entropy loss, we do not introduce additional weighting hyperparameters for L F E M . Therefore, the total loss function is given by the following Equation (3):
L = L C E , c l f + λ · L M S E + L F E M

3.4. SoftPool and Global SoftPool (GSP)

For most CNN-based steganalysis networks, average pooling is typically employed between convolutional layers. Pooling reduces the model size, improves computational speed, and enhances robustness. However, conventional pooling methods may not adapt well to the characteristics of steganalysis, leading to weakening of the embedded signals.
To address the limitations of conventional pooling layers, we utilize SoftPool [27] for feature map downsampling in the network. The calculation method is illustrated by the following Equation (4):
a ˜ = i R 0 e a i j R e a j a i
In this context, a i represents an element in the feature map, R 0 denotes the kernel neighborhood for downsampling, and a ˜ signifies the sampling result. SoftPool offers several advantages: compared with average pooling, it can effectively highlight important features; compared with max pooling, it avoids discarding excessive information, making it better suited for steganalysis by preserving critical features more effectively during the pooling process.
Additionally, we employ SoftPool as the Global SoftPool (GSP) after the final convolution module, enhancing its diversity by introducing a coefficient. Its expression is depicted in the following Equation (5):
a ˜ = i R e n a i j R e n a j a i
At this point, R represents the entire feature map, and n denotes the coefficient being added. We set n to four different values: n = 1 , 0 , 1 , + . When n = 1 , it emphasizes weaker elements; when n = 0 , it is equivalent to global average pooling (GAP), emphasizing all elements equally; when n = 1 , it highlights representative elements while also integrating others; and when n = + , it is equivalent to global max pooling (GMP), only emphasizing representative elements. This configuration of GSP enables a comprehensive assessment of the feature map’s characteristics, facilitating a better collection of information from the feature map.

4. Experimental Results and Analysis

4.1. Dataset and Environment Settings

In this section, we conducted extensive experiments to demonstrate the effectiveness of WERSNet. The datasets used in our experiments are BOSSbase 1.01 [37] and BOWS2 [38], each containing 10,000 grayscale images of size 512 × 512 . Due to limitations in GPU computational capacity, we resized the image resolution from 512 × 512 to 256 × 256 . We use steganography to generate stego images after the image resizing is completed. Therefore, the operation of resizing the image will not have a potential impact on the experimental results.
We employed three different spatial domain image steganography methods in our experiments: WOW [11], S-UNIWARD [39], and HILL [12]. Each steganography method includes four different payload sizes: 0.1 bpp, 0.2 bpp, 0.3 bpp, and 0.4 bpp. For each steganography method, each embedding rate generated 10,000 cover-stego pairs. The dataset used in Section 4.2, Section 4.3 and Section 4.4 is BOSSbase, with a training–validation–testing split ratio of 4:1:5. Each sample pair is randomly partitioned. Additionally, in Section 4.2, Section 4.3 and Section 4.4, only BOSSbase 1.01 is used. In Section 4.5, the training set also added 10,000 pairs of images from BOSW2.
All experiments employed two data augmentations: random horizontal flipping and random rotation at four angles ( 0 , 90 , 180 , 270 ). The optimizer used during training is SGD, with weight decay and momentum values set to 0.0005 and 0.9, respectively. The batch size is 32 (16 pairs of cover-stego). L2 regularization is disabled for all convolutional and fully connected layers’ bias terms. The initial learning rate is 0.01, divided by 10 at epochs 80, 140, and 180, for a total of 300 epochs. To ensure fair comparison, all steganalysis models were implemented using PyTorch on identical hardware facilities. All experiments were repeated three times, and the average value was taken as the final result. The GPU used is NVIDIA RTX3090, with Python version 3.7 and PyTorch version 1.7.1.

4.2. Ablation Study of LFCHP

LFCHP is a newly designed preprocessing structure for steganalysis that we have developed. In this section, we will examine the effectiveness of LFCHP. Initially, we will consider the impact of hyperparameters λ in Equation (3) on the entire network. The results in Table 1 show that when the hyperparameter λ = 0.1 , the detection accuracy is 85.25%, achieving the best detection performance. We will set λ = 0.1 in all subsequent experiments. For other cases, detection performance is weaker but still yields good results. We attempted to set λ = 0 , meaning not introducing high-pass prior knowledge, as a comparative experiment with other values. However, in this scenario, WERSNet fails to converge.
To further investigate the enhancement of detection performance by LFCHP, we compared several different preprocessing approaches, as shown in Figure 2.
HPF in Figure 2 is composed of a convolution layer and TLU (truncation threshold T = 3 ) activation function, where the parameters of the convolution layer are initialized and fixed by 30 different high-pass filters in SRM, which can effectively filter out low-frequency signals and enhance the signal-to-noise ratio of embedded signals, and is a classic preprocessing layer structure in steganography.
LFCHP in Figure 2 is our proposed preprocessing structure. LFB is randomly initialized and trained jointly by dataset and HPCB, which not only solves the possible defects of manual design, but also integrates high-pass prior knowledge and enables the preprocessing layer to focus on the complex region where the embedded signal is located.
The preprocessing layer scheme in Figure 2b is used as an ablation scheme. As can be seen from Table 2, HPF+2conv only improves by 0.09% and 0.18% compared to HPF, indicating that blindly adding convolutional layers cannot significantly improve the performance of the network, while our proposed LFCHP improves by 1.97% and 2.00% compared with HPF, which is a significant improvement, indicating that the improvement achieved by LFCHP is not caused by the additional two layers of convolutional layers, reflecting the rationality and superiority of the LFCHP design.
To more intuitively demonstrate the advantages of our proposed LFCHP, we visualized the output results of two preprocessing layers, HPF (a high-pass filter with 30 fixed parameters) and LFCHP (32 channels in the output feature map) in Figure 2, as shown in Figure 3. Compared with HPF, LFCHP offers the following advantages:
(1) After HPF processing, many points are truncated by the TLU activation function (depicted as white points in Figure 3c). These truncated pixels have relatively low signal-to-noise ratios. While truncating these points helps the network focus on high signal-to-noise ratio points and aids in network training, it also results in loss of embedded information, potentially impacting network performance improvement. LFCHP, after optimization through training, adapts better to the characteristics of the embedding regions. Consequently, fewer points are forcibly truncated after LFCHP processing, indicating that LFCHP preserves more embedded information compared with HPF.
(2) The grayscale values of the filtering results after HPF processing are limited to a few fixed amplitudes, potentially causing the network to overlook many details. In contrast, the grayscale values of the feature maps after LFCHP are not limited to a few values; they are more evenly distributed and comprehensive, enabling more detailed extraction of embedded signal features.
(3) The variations between different filtering results after HPF processing are not very distinct, resulting in considerable information redundancy. In contrast, LFCHP exhibits significant variations between different channels, leading to richer feature extraction and significantly reduced information redundancy.

4.3. Ablation Study for the SSAB and the Pooling Layer

For the issue of weak signal features possibly disappearing as they propagate from shallow to deep networks, existing methods typically employ residual connections to aid feature propagation. In addition to residual connections, we also considered methods to enhance the strength of weak signal features. To validate this idea, we removed SSAB from WERSNet while keeping the remaining structure unchanged, and then compared it with the cover WERSNet. The results in Table 3 indicate that after removing SSAB, WERSNet’s performance decreased by 1.21% and 1.12% at payload rates of 0.2 bpp and 0.4 bpp, respectively.
The faintness of steganographic signals is an inherent property for security reasons, yet it also leads to the suppression of weak signals during convolutional stacking in steganalysis. Therefore, enhancing the features of weak signals can effectively improve classification performance. Through SSAB, we extract second-order information such as variance and covariance from the network, which can better capture subtle variations in feature maps, thus fully extracting the features of weak signals. By constructing FEM with SSAB and hierarchically applying FEM to WERSNet, we can enhance the features of weak signals at multiple levels, mitigating the inhibitory effect on weak signals during convolutional stacking.
Due to the fact that common pooling methods tend to suppress the features of weak signals, we attempted different combinations of pooling and global pooling to alleviate this issue, as shown in Table 4. Experimental results show that the accuracy of MaxPool+GAP is 82.78%, which is the lowest among all methods. This is because MaxPool will lose a lot of information, resulting in a significant decrease in performance. The accuracy of AvgPool+GAP is 84.31%, which is significantly improved compared with MaxPool+GAP. This is because AvgPool can utilize all elements in the kernel field, which alleviates the problem of excessive information loss to a certain extent, but it is still falls short. SoftPool, building upon the utilization of kernel field information, emphasizes key elements, further enhancing performance; specifically, the combination of SoftPool+GAP outperforms AvgPool+GAP by 0.39%. Though the performance gain is not substantial, it does not entail additional parameter overhead.
In the Feature Extraction section, replacing the other two AvgPools with SoftPool did not yield further performance improvements; hence, we refrain from replacing these AvgPools. Furthermore, the extension of SoftPool to GSP achieved a notable performance boost. In summary, the combination of SoftPool+GSP improves performance by 0.94% compared with AvgPool+GAP and by 2.47% compared with MaxPool+GAP. This suggests that the new pooling methods effectively alleviate the inhibitory effect on weak signal features during network propagation compared with existing pooling methods, thereby enhancing network performance. SoftPool not only utilizes all information within the kernel neighborhood but also emphasizes representative features, reducing information loss during downsampling. GSP, building upon SoftPool, captures feature map information from multiple perspectives, enriching the features of global pooling.

4.4. Comparison of Generalization Performance Across Different Steganography Techniques

To assess the generalization performance of WERSNet, we conducted experiments with mismatched steganographic sources. In this section of experiments, the steganalysis methods used were trained on a single steganographic method’s training set and tested on test sets with different steganographic methods, all constructed with a payload of 0.4 bpp.
Table 5 presents the results of WERSNet compared with SRNet, Zhu-Net, CPSN, FPNet, and LWENet in detecting mismatched steganographic sources under three steganographic techniques. Overall, all detection results of WERSNet outperform those of other steganalysis networks. When the test set employs the easiest-to-detect method, WOW, all steganalysis networks demonstrate good generalization performance. For a training set using HILL and a test set using WOW, all steganalysis networks show higher detection accuracy compared with when the test set uses HILL; however, our model’s detection accuracy is significantly superior to other networks. With the test set using S-UNIWARD, all steganalysis networks experience varying degrees of performance loss, with ours exhibiting the lowest loss. When the test set utilizes the most challenging to detect, HILL, our model’s performance loss is lower than FPNet and LWENet, higher than SRNet and Zhu-Net, comparable to CPSN, yet still superior to other methods in terms of detection accuracy. Overall, our model outperforms the other five methods in terms of generalization performance across different steganography techniques, making it more suitable for practical applications.

4.5. Performance Comparison

To validate the effectiveness of WERSNet, we conducted a fair comparison with five state-of-the-art steganalysis networks, including SRNet, Zhu-Net, CPSN, FPNet, and LWENet. In addition to the BOSSbase dataset, we added the BOWS2 dataset to ensure better representation of different methods’ performance. The steganographic algorithms employed were WOW, S-UNIWARD, and HILL, with each steganalysis algorithm using the same dataset split (as detailed in Section 4.1). Each steganographic algorithm utilized payloads ranging from 0.1 bpp to 0.4 bpp.
The experimental results presented in Table 6 and Figure 4 demonstrate that for any steganographic algorithm and payload used, WERSNet achieved higher detection accuracy compared with the other five steganalysis networks. Specifically, WERSNet exhibited a 2.45% to 4.51% improvement over SRNet, a 6.22% to 9.55% improvement over Zhu-Net, a 1.26% to 4.14% improvement over CPSN, a 2.38% to 4.92% improvement over FPNet, and a 1.08% to 2.96% improvement over LWENet.
Additionally, we depicted the validation set accuracy curves for these six methods with a payload of 0.4 bpp, as shown in Figure 5. The validation set accuracy curves accurately reflect the convergence status of various methods. WERSNet exhibited convergence speeds similar to CPSN and LWENet, both of which utilize fixed high-pass filtering kernels, but with superior accuracy performance. These three methods converged faster than FPNet, Zhu-Net, and SRNet, which employ trainable high-pass filtering kernels and randomly initialized kernels. This observation highlights that LFCHP not only demonstrates excellent performance in accuracy but also exhibits competitive convergence speeds.

5. Conclusions

In this paper, we introduce a convolutional neural network called WERSNet for steganalysis of spatial domain images. Starting from enhancing weak steganographic signal features, we design a novel preprocessing structure, LFCHP, which significantly improves preprocessing performance by introducing high-pass prior constraints while being randomly initialized. Furthermore, utilizing SSAB, we construct an FEM capable of enhancing weak signal features and hierarchically apply FEM to WERSNet, thereby enhancing weak signal features at multiple levels and alleviating the suppression effect on weak signals during convolution stacking. Additionally, we replace some AvgPool operations with SoftPool and extend SoftPool to GSP. SoftPool emphasizes representative features while utilizing information from the entire kernel neighborhood, reducing information loss during downsampling. GSP, based on SoftPool, captures feature map information from multiple angles, enriching global pooling features. Compared with existing steganalysis networks, WERSNet achieves significant improvements and demonstrates outstanding generalization performance across different steganography techniques.
In future work, we will explore the application of LFCHP in various steganalysis domains such as text and audio steganalysis. Additionally, we will investigate methods to enhance the generalization performance of steganalysis networks, which is crucial for practical application scenarios.

Author Contributions

Conceptualization, W.L. and Q.L.; methodology, W.L.; software, W.L.; validation, W.L.; formal analysis, W.L. and Q.L.; investigation, W.L.; resources, Q.L.; data curation, Q.L.; writing—original draft preparation, W.L.; writing—review and editing, W.L. and Q.L.; visualization, W.L.; supervision, Q.L.; project administration, Q.L.; funding acquisition, Q.L. All authors have read and agreed to the published version of the manuscript.

Funding

This is a part research is supported by National Natural Science Foundation of China(61871138) and Guangxi Science and Technology Base and Talent Special Project (No. AD25069071).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Code available: https://github.com/LWL139/WSERNet green(accessed on: 5 February 2026). Additional data are available upon request to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LFCHPlearnable filter constrained by high-pass prior
SSABsecond-order signal auxiliary branch
CNNconvolutional neural network
GNCNNGaussian-Neuron CNN
TLUtruncated linear unit
CPSNCovariance Pooling Steganalysis Network
HPCBhigh-pass prior constraint branch
LFBlearnable filter bank
FEMfeature enhancement module
BNbatch normalization
FCfully connected
TanHhyperbolic tangent function
GSPglobal Soft Pooling
GCPglobal covariance pooling
GAPglobal average pooling
GMPglobal max pooling

References

  1. Xiang, S.; Luo, X. Reversible Data Hiding in Encrypted Image Based on Homomorphic Public Key Cryptosystem. J. Softw. 2016, 27, 1592. [Google Scholar] [CrossRef]
  2. Wani, M.A.; Sultan, B. Deep learning based image steganography: A review. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2023, 13, e1481. [Google Scholar] [CrossRef]
  3. Kombrink, M.H.; Geradts, Z.J.M.H.; Worring, M. Image Steganography Approaches and Their Detection Strategies: A Survey. ACM Comput. Surv. 2024, 57, 1–40. [Google Scholar] [CrossRef]
  4. Chen, F.; Fu, Z.; Zhang, W.; Cheng, X.; Sun, X. Review of image steganalysis based on deep learning. Ruan Jian Xue Bao 2021, 32, 551–578. [Google Scholar] [CrossRef]
  5. Gao, S.; Zhang, Z.; Li, Q.; Ding, S.; Iu, H.H.C.; Cao, Y.; Xu, X.; Wang, C.; Mou, J. Encrypt a Story: A Video Segment Encryption Method Based on the Discrete Sinusoidal Memristive Rulkov Neuron. IEEE Trans. Dependable Secur. Comput. 2025, 22, 8011–8024. [Google Scholar] [CrossRef]
  6. Pevný, T.; Filler, T.; Bas, P. Using High-Dimensional Image Models to Perform Highly Undetectable Steganography. In Proceedings of the Information Hiding; Böhme, R., Fong, P.W.L., Safavi-Naini, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 161–177. [Google Scholar]
  7. Cheng, X.; Wang, H.; Luo, X.; Guan, Q.; Ma, B.; Wang, J. Re-cropping Framework: A Grid Recovery Method for Quantization Step Estimation in Non-aligned Recompressed Images. IEEE Trans. Circuits Syst. Video Technol. 2025, early access. [Google Scholar] [CrossRef]
  8. Pevný, T.; Bas, P.; Fridrich, J. Steganalysis by Subtractive Pixel Adjacency Matrix. In MM&Sec ’09: Proceedings of the 11th ACM Workshop on Multimedia and Security; ACM Digital Library: New York, NY, USA, 2009; pp. 75–84. [Google Scholar] [CrossRef]
  9. Fridrich, J.; Kodovsky, J. Rich Models for Steganalysis of Digital Images. IEEE Trans. Inf. Forensic Secur. 2012, 7, 868–882. [Google Scholar] [CrossRef]
  10. Holub, V.; Fridrich, J. Random Projections of Residuals for Digital Image Steganalysis. IEEE Trans. Inf. Forensic Secur. 2013, 8, 1996–2006. [Google Scholar] [CrossRef]
  11. Holub, V.; Fridrich, J. Designing steganographic distortion using directional filters. In Proceedings of the 2012 IEEE International Workshop on Information Forensics and Security (WIFS); IEEE: New York, NY, USA, 2012; pp. 234–239. [Google Scholar] [CrossRef]
  12. Li, B.; Wang, M.; Huang, J.; Li, X. A new cost function for spatial image steganography. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP); IEEE: New York, NY, USA, 2014; pp. 4206–4210. [Google Scholar] [CrossRef]
  13. Zhang, L.; Lu, Y.; Li, T.; Lu, G. Joint adjustment image steganography networks. Signal Process. Image Commun. 2023, 118, 117022. [Google Scholar] [CrossRef]
  14. Wang, T.; Cheng, H.; Liu, X.; Xu, Y.; Chen, F.; Wang, M.; Chen, J. Lossless image steganography: Regard steganography as super-resolution. Inf. Process. Manag. 2024, 61, 103719. [Google Scholar] [CrossRef]
  15. Tan, S.; Li, B. Stacked convolutional auto-encoders for steganalysis of digital images. In Proceedings of the Signal and Information Processing Association Annual Summit and Conference (APSIPA), 2014 Asia-Pacific; IEEE: New York, NY, USA, 2014; pp. 1–4. [Google Scholar] [CrossRef]
  16. Qian, Y.; Dong, J.; Wang, W.; Tan, T. Deep learning for steganalysis via convolutional neural networks. In Proceedings of the Media Watermarking, Security, and Forensics 2015; Alattar, A.M., Memon, N.D., Heitzenrater, C.D., Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2015; Volume 9409, p. 94090J. [Google Scholar] [CrossRef]
  17. Xu, G.S.; Wu, H.Z.; Shi, Y.Q. Structural Design of Convolutional Neural Networks for Steganalysis. IEEE Signal Process. Lett. 2016, 23, 708–712. [Google Scholar] [CrossRef]
  18. Ye, J.; Ni, J.; Yi, Y. Deep Learning Hierarchical Representations for Image Steganalysis. IEEE Trans. Inf. Forensic Secur. 2017, 12, 2545–2557. [Google Scholar] [CrossRef]
  19. Deng, X.; Chen, B.; Luo, W.; Luo, D. Fast and Effective Global Covariance Pooling Network for Image Steganalysis. In IH&MMSec’19: Proceedings of the ACM Workshop on Information Hiding and Multimedia Security; ACM Digital Library: New York, NY, USA, 2019; pp. 230–234. [Google Scholar] [CrossRef]
  20. Boroumand, M.; Chen, M.; Fridrich, J. Deep Residual Network for Steganalysis of Digital Images. IEEE Trans. Inf. Forensic Secur. 2019, 14, 1181–1193. [Google Scholar] [CrossRef]
  21. Zhang, R.; Zhu, F.; Liu, J.; Liu, G. Depth-Wise Separable Convolutions and Multi-Level Pooling for an Efficient Spatial CNN-Based Steganalysis. IEEE Trans. Inf. Forensic Secur. 2020, 15, 1138–1150. [Google Scholar] [CrossRef]
  22. Liu, J.; Jiao, G.; Sun, X. Feature Passing Learning for Image Steganalysis. IEEE Signal Process. Lett. 2022, 29, 2233–2237. [Google Scholar] [CrossRef]
  23. Fu, T.; Chen, L.; Fu, Z.; Yu, K.; Wang, Y. CCNet: CNN model with channel attention and convolutional pooling mechanism for spatial image steganalysis. J. Vis. Commun. Image Represent. 2022, 88, 103633. [Google Scholar] [CrossRef]
  24. Boureau, Y.L.; Ponce, J.; LeCun, Y. A theoretical analysis of feature pooling in visual recognition. In Proceedings of the 27th International Conference on Machine Learning (ICML-10); ACM Digital Library: New York, NY, USA, 2010; pp. 111–118. [Google Scholar]
  25. Zhang, Q.; Cui, Z.; Niu, X.; Geng, S.; Qiao, Y. Image segmentation with pyramid dilated convolution based on ResNet and U-Net. In Proceedings of the Neural Information Processing: 24th International Conference, ICONIP 2017, Guangzhou, China, 14-18 November 2017; Proceedings, Part II 24; Springer: Cham, Switzerland, 2017; pp. 364–372. [Google Scholar]
  26. Wu, T.; Ren, W.; Li, D.; Wang, L.; Jia, J. JPEG steganalysis based on denoising network and attention module. Int. J. Intell. Syst. 2022, 37, 5011–5030. [Google Scholar] [CrossRef]
  27. Stergiou, A.; Poppe, R.; Kalliatakis, G. Refining activation downsampling with SoftPool. In Proceedings of the IEEE/CVF International Conference on Computer Vision; IEEE: New York, NY, USA, 2021; pp. 10357–10366. [Google Scholar]
  28. Xu, X.; Song, X.; Zhao, R. Selection-Channel-Aware Scale Co-occurrence Matrix Feature to Detect Content-Adaptive JPEG Steganography. In Proceedings of the 2018 2nd International Conference on Computer Science and Artificial Intelligence; ACM Digital Library: New York, NY, USA, 2018; pp. 375–379. [Google Scholar]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  30. Li, P.; Xie, J.; Wang, Q.; Gao, Z. Towards faster training of global covariance pooling networks by iterative matrix square root normalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: New York, NY, USA, 2018; pp. 947–955. [Google Scholar]
  31. You, W.; Zhang, H.; Zhao, X. A Siamese CNN for Image Steganalysis. IEEE Trans. Inf. Forensic Secur. 2021, 16, 291–306. [Google Scholar] [CrossRef]
  32. Chopra, S.; Hadsell, R.; LeCun, Y. Learning a similarity metric discriminatively, with application to face verification. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05); IEEE: New York, NY, USA, 2005; Volume 1, pp. 539–546. [Google Scholar]
  33. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06); IEEE: New York, NY, USA, 2006; Volume 2, pp. 1735–1742. [Google Scholar]
  34. Weng, S.; Chen, M.; Yu, L.; Sun, S. Lightweight and Effective Deep Image Steganalysis Network. IEEE Signal Process. Lett. 2022, 29, 1888–1892. [Google Scholar] [CrossRef]
  35. Lin, Z.; Gao, Y.; Sang, J. Investigating and Explaining the Frequency Bias in Image Classification. arXiv 2022, arXiv:2205.03154. [Google Scholar] [CrossRef]
  36. Li, L.; Xie, J.; Li, P.; Zhang, L. Detachable second-order pooling: Toward high-performance first-order networks. IEEE Trans. Neural Networks Learn. Syst. 2021, 33, 3400–3414. [Google Scholar] [CrossRef]
  37. Bas, P.; Filler, T.; Pevnỳ, T. “Break our steganographic system”: The ins and outs of organizing BOSS. In Proceedings of the International Workshop on Information Hiding; Springer: Berlin/Heidelberg, Germany, 2011; pp. 59–70. [Google Scholar]
  38. Furon, P. BOWS-2. 2007. Available online: http://bows2.ec-lille.fr (accessed on 1 December 2025).
  39. Holub, V.; Fridrich, J.; Denemark, T. Universal distortion function for steganography in an arbitrary domain. EURASIP J. Inf. Secur. 2014, 2014, 1. [Google Scholar] [CrossRef]
Figure 1. The overall structure of WSERNet. WSERNet consists of three parts: LFCHP, Feature Extraction, and Classifier. LFCHP includes LFB and HPCB. LFB eliminates the defects of fixed convolution kernels through random initialization. In HPCB, 30HPF+TLU is a commonly used preprocessing method, its output is F g , and the number of channels is 30. “Conv1, 30” changes the number of output channels of LFB from 32 to 30, aligning the number of channels of F r and F g . Through L M S E , HPCB is responsible for imposing high-pass prior constraints on LFB to keep LFB focused on complex areas. SSAB, a second-order signal auxiliary branch in FEM, enhances weak signal features using second-order statistical information from feature maps. The numbers in Feature Extraction and CBT (c) represent the number of channels. “Block 1/2” indicates the option to choose between Block 1 and Block 2, with Block 1 used when the input and output channel numbers are equal, and Block 2 used otherwise. “(Conv1/3, c)” denotes convolutional layers with a kernel size of 1 or 3 and output channel numbers of c. During the inference phase, all HPCB and SSAB branches are removed.
Figure 1. The overall structure of WSERNet. WSERNet consists of three parts: LFCHP, Feature Extraction, and Classifier. LFCHP includes LFB and HPCB. LFB eliminates the defects of fixed convolution kernels through random initialization. In HPCB, 30HPF+TLU is a commonly used preprocessing method, its output is F g , and the number of channels is 30. “Conv1, 30” changes the number of output channels of LFB from 32 to 30, aligning the number of channels of F r and F g . Through L M S E , HPCB is responsible for imposing high-pass prior constraints on LFB to keep LFB focused on complex areas. SSAB, a second-order signal auxiliary branch in FEM, enhances weak signal features using second-order statistical information from feature maps. The numbers in Feature Extraction and CBT (c) represent the number of channels. “Block 1/2” indicates the option to choose between Block 1 and Block 2, with Block 1 used when the input and output channel numbers are equal, and Block 2 used otherwise. “(Conv1/3, c)” denotes convolutional layers with a kernel size of 1 or 3 and output channel numbers of c. During the inference phase, all HPCB and SSAB branches are removed.
Sensors 26 01329 g001
Figure 2. Different preprocessing layers. HPF is the classical preprocessing method. HPF+2Conv adds two convolution layers after HPF for comparison with LFCHP. LFCHP is our proposed preprocessing method.
Figure 2. Different preprocessing layers. HPF is the classical preprocessing method. HPF+2Conv adds two convolution layers after HPF for comparison with LFCHP. LFCHP is our proposed preprocessing method.
Sensors 26 01329 g002
Figure 3. Visualization of different preprocessing results. (a) is the input image. (b) is the result image obtained after the input image has been preprocessed by 30HPF+TLU. The image difference is not significant and has high redundancy. (c) shows the result image obtained after LFCHP preprocessing of the input image. The image difference is more obvious, reducing information redundancy, and the focus on complex areas is also maintained.
Figure 3. Visualization of different preprocessing results. (a) is the input image. (b) is the result image obtained after the input image has been preprocessed by 30HPF+TLU. The image difference is not significant and has high redundancy. (c) shows the result image obtained after LFCHP preprocessing of the input image. The image difference is more obvious, reducing information redundancy, and the focus on complex areas is also maintained.
Sensors 26 01329 g003
Figure 4. WERSNet comparison of performances with SRNet, Zhu-Net, CPSN, FPNet, and LWENet under the three steganographic algorithms.
Figure 4. WERSNet comparison of performances with SRNet, Zhu-Net, CPSN, FPNet, and LWENet under the three steganographic algorithms.
Sensors 26 01329 g004
Figure 5. The accuracy curves of six networks on the validation set under HILL at 0.4 bpp.
Figure 5. The accuracy curves of six networks on the validation set under HILL at 0.4 bpp.
Sensors 26 01329 g005
Table 1. Detection accuracy (%) of WERSNet under different λ values.
Table 1. Detection accuracy (%) of WERSNet under different λ values.
λ0.070.080.090.10.20.30.4
Acc85.0784.8785.2085.2585.1085.1485.02
Table 2. Detection accuracy of different preprocessing layer schemes.
Table 2. Detection accuracy of different preprocessing layer schemes.
PreprocessingHILL (0.2 bpp)HILL (0.4 bpp)
HPF73.4083.25
HPF+2conv73.4983.43
LFCHP75.3785.25
Table 3. Ablation experiment of SSAB.
Table 3. Ablation experiment of SSAB.
NetworkHILL (0.2 bpp)HILL (0.4 bpp)
WSERNet without SSAB74.1684.13
WSERNet with SSAB75.3785.25
Table 4. Impact of different pooling methods on performance with the HILL embedding method at a payload of 0.4 bpp.
Table 4. Impact of different pooling methods on performance with the HILL embedding method at a payload of 0.4 bpp.
Pooling+Global PoolingAccuracyNumber of Classifier Parameters
MaxPool+GAP82.78514
AvgPool+GAP84.31514
SoftPool+GAP84.70514
AvgPool+GSP84.892050
SoftPool+GSP85.252050
Table 5. The experiments of mismatched steganographic sources on the BOSSbase dataset.
Table 5. The experiments of mismatched steganographic sources on the BOSSbase dataset.
NetworkTrain MethodTest Method
WOWS-UNIWARDHILL
SRNetWOW87.4178.3965.88
S-UNIWARD84.1285.1973.41
HILL81.8373.4381.59
Zhu-NetWOW85.7476.2763.82
S-UNIWARD81.7682.5367.42
HILL80.1172.8479.26
CPSNWOW89.5182.4465.42
S-UNIWARD88.8488.3973.18
HILL84.1076.7283.34
FPNetWOW88.1275.3062.51
S-UNIWARD83.4886.1767.79
HILL81.1873.6080.09
LWENetWOW89.8381.1064.62
S-UNIWARD88.7388.5671.98
HILL83.9777.1183.58
oursWOW91.4384.7267.47
S-UNIWARD90.4189.9073.67
HILL86.0280.4185.25
Table 6. Comparison of accuracy (Acc) of six methods under three steganographic techniques at four embedding rates.
Table 6. Comparison of accuracy (Acc) of six methods under three steganographic techniques at four embedding rates.
NetworkWOWS-UNIWARDHILL
0.10.20.30.40.10.20.30.40.10.20.30.4
SRNet [20]73.6782.9187.4190.0868.2078.8884.8688.4366.3474.4980.3183.78
Zhu-Net [21]71.3481.1086.2588.6962.4371.5878.4484.6362.4670.5976.2081.10
CPSN [19]74.0483.3788.0791.2568.5279.2685.2589.8567.2075.8881.2285.25
FPNet [22]74.6383.0987.9590.7667.1476.8884.5087.3666.7574.0779.4782.96
LWENet [34]76.1684.1488.2392.0669.0379.7285.6689.7867.8775.2781.3485.59
Ours78.1887.1090.4793.1472.0681.3387.3591.6369.7977.1483.3587.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, W.; Li, Q. Steganalysis Network for Weak Steganographic Signal Extraction and Enhancement. Sensors 2026, 26, 1329. https://doi.org/10.3390/s26041329

AMA Style

Liang W, Li Q. Steganalysis Network for Weak Steganographic Signal Extraction and Enhancement. Sensors. 2026; 26(4):1329. https://doi.org/10.3390/s26041329

Chicago/Turabian Style

Liang, Weilin, and Qingguang Li. 2026. "Steganalysis Network for Weak Steganographic Signal Extraction and Enhancement" Sensors 26, no. 4: 1329. https://doi.org/10.3390/s26041329

APA Style

Liang, W., & Li, Q. (2026). Steganalysis Network for Weak Steganographic Signal Extraction and Enhancement. Sensors, 26(4), 1329. https://doi.org/10.3390/s26041329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop