Abstract
In image splicing tamper detection, forgery operations simultaneously introduce macroscopic semantic inconsistencies and microscopic tampering artifacts. Conventional methods often treat semantic understanding and low-level artifact perception as separate tasks, which impedes their effective synergy. Meanwhile, frequency-domain information, a crucial clue for identifying traces of tampering, is frequently overlooked. However, a simplistic fusion of frequency-domain and spatial features can lead to feature conflicts and information redundancy. To resolve these challenges, this paper proposes an Adaptive Multi-Scale Edge-Aware Network (AMSEANet). This network employs a synergistic enhancement cascade architecture, recasting semantic understanding and artifact perception as a single, frequency-aware process guided by deep semantics. It leverages data-driven adaptive filters to precisely isolate and focus on edge artifacts that signify tampering. Concurrently, the dense fusion and enhancement of cross-scale features effectively preserve minute tampering clues and boundary details. Extensive experiments demonstrate that our proposed method achieves superior performance on several public datasets and exhibits excellent robustness against common attacks, such as noise and JPEG compression.
1. Introduction
With the rapid development of digital image processing technologies, current techniques for digital image generation, editing, and dissemination have permeated every aspect of modern life. As the primary medium for information transmission, images play an indispensable role in critical domains such as photojournalism, judicial forensics, and national security, where their authenticity and integrity directly impact public perception and social stability. However, the evolution of image editing software, typified by Adobe Photoshop, has significantly lowered the barrier to creating sophisticated forgeries. This enables even non-expert users to easily produce highly realistic image manipulations. Among these techniques, image splicing—which involves constructing false semantic content by copying and pasting regions from multiple source images—has become a primary method of manipulation. It is important to note that while image editing serves many legitimate creative purposes, the scope of this paper is focused on the forensic challenge of detecting “malicious forgeries” intended to deceive, as illustrated in Figure 1. Such tampering not only disrupts social order but can also be exploited for criminal activities, including political manipulation and financial fraud, thereby highlighting the urgent need for robust image splicing detection technologies.
Figure 1.
Example of image splicing forgery. (a) The authentic image. (b) The forged image. (c) The ground truth.
Compared to other tampering methods, such as copy-move or content removal, image splicing achieves semantic deception by fusing content from different image domains, which significantly increases its detection difficulty. On the one hand, forgers can employ sophisticated post-processing techniques (e.g., noise addition, color adjustment) to eliminate the traces of physical inconsistency relied upon by traditional detection methods. On the other hand, the complex semantic relationships and cross-domain feature discrepancies among source images often cause existing algorithms to generate false positives and false negatives when localizing splicing boundaries and tracing the provenance of forged regions. Furthermore, deep learning-driven generative techniques, such as GAN-based image synthesis, further blur the visual boundaries between authentic and forged content, placing higher demands on the robustness and generalization capabilities of detection models.
During their formation, digital images acquire unique digital imprints from the optical characteristics of the imaging device [1], the physical environment, and digital processing procedures [2]. These medium-specific features, unrelated to the visual content, are termed “image digital fingerprints” due to their uniqueness and can serve as identifiers for image source tracing. In the field of digital image forensics, the core principle of splicing forgery detection lies in identifying attribute inconsistencies between the tampered and authentic regions. This inconsistency arises from the inherent feature differences between various source images. Existing research has developed multiple feature extraction techniques to capture such discrepancies, which are mainly categorized into two approaches: traditional handcrafted feature methods [3,4], and deep learning-based CNN feature extraction methods [5,6,7,8]. Traditional methods involve extracting features through manual design, followed by comparative or statistical analysis of these features to determine if an image has been tampered with and to locate the spliced region. However, the performance of these traditional image forensic methods heavily relies on predefined, singular image fingerprints (such as sensor noise or traces of interpolation). These features are vulnerable to common post-processing operations like blurring and JPEG compression, which can lead to feature invalidation and result in missed detections.
Inspired by the success of deep learning in other computer vision domains, researchers began to actively explore its application in image tamper detection [9]. The essence of image tamper detection lies in the synergy of localizing tampering artifacts (e.g., inconsistent edge seams, abnormal noise patterns) and verifying the consistency of pixel origins. This task requirement spurred the development of patch-level analysis frameworks [10,11]. These methods segment the image into independent small patches for processing, relying on local feature extraction mechanisms to avoid excessive focus on semantic content unrelated to traces of tampering, thereby enhancing sensitivity to subtle artifacts like resampling patterns and compression noise [12]. However, their core deficiency lies in a fragmented processing logic. While the local field of view of a single patch can isolate irrelevant semantic content, it consequently fails to capture tamper-related spatial information and global semantic associations (such as cross-regional physical plausibility and lighting consistency), thus ignoring the relationships between patches. For instance, artifacts are prone to appear at patch boundaries, necessitating post-processing algorithms for repair; simultaneously, the patch size must be manually preset, making it difficult to adapt to multi-scale tampering scenarios (e.g., small targets being missed or large targets being fragmented). More critically, these methods rely on a disjointed “divide → classify → merge” workflow, which decouples the feature extraction and localization tasks, thereby weakening the model’s generalization ability for complex tampering patterns (such as low-contrast splicing).
To achieve better global modeling, researchers proposed methods based on end-to-end networks [13,14,15,16,17]. Instead of partitioning the image, they treat splicing tamper detection as a pixel-wise binary classification problem: classifying each pixel as either belonging to a “tampered region” or an “authentic region”. Such methods can accept images of any size as input, thereby eliminating the feature discontinuities caused by manual patching and overcoming the drawbacks of patch-based approaches to some extent. However, constrained by the inherent architecture of CNNs, deep networks still prioritize local features and have insufficient global semantic verification capabilities. Concurrently, local and global information are difficult to effectively complement, making it challenging to capture the contextual associations of traces of tampering at different scales.
Against this backdrop, dual-branch network architectures emerged [18,19,20,21], attempting to break the limitations of single-stream modeling through heterogeneous feature fusion strategies. These methods typically construct parallel processing paths to separately extract spatial-domain features (e.g., RGB image features) and frequency-domain features (e.g., noise, compression artifacts), compensating for the global perception shortcomings of CNNs through cross-modal feature interaction. However, their core bottleneck is the information isolation introduced by the parallel branches. In this paradigm, the frequency-domain analysis branch operates without knowledge of the image’s semantic content, while the spatial-domain branch cannot leverage low-level artifact clues to aid its judgment. This ’working in isolation’ during the feature extraction stage poses a significant challenge for subsequent fusion modules (whether simple concatenation or complex attention mechanisms), making it difficult to fundamentally resolve the feature conflicts and semantic disconnects caused by information isolation. Furthermore, existing attention mechanisms struggle to establish a synergistic enhancement relationship between local details (such as edge artifacts) and global semantics (such as the distribution of the tampered region), leading to frequent issues of missed small-scale tampered targets and blurred boundary localization.
To address the aforementioned shortcomings, this paper proposes an Adaptive Multi-Scale Edge-Aware Network (AMSEANet). This network discards the inherent drawbacks of the traditional parallel dual-branch paradigm in feature interaction and innovatively designs a single-backbone synergistic enhancement framework aimed at organically unifying the two previously separate tasks of deep semantic understanding and low-level tampering artifact perception. A fundamental prerequisite for realizing this synergistic framework is to address the problem of information dilution, where key features are inherently diluted as they pass through deep layers of a single-backbone network. To tackle this challenge, we first introduce the Cross-Scale Dense Residual Fusion Block (CSDRFBlock). This module forms the foundation for feature preservation, ensuring the efficient cross-level transmission of critical clues, such as minute tampering artifacts, from the source. Building upon this high-fidelity feature stream, the network must then precisely discern traces of tampering from a complex background. Considering the limitations of traditional fixed filters for this task, we subsequently designed the Edge-Aware Spatial-Frequency Fusion Module (ESFFM), driven by an Adaptive Smooth Filter. This module takes the enhanced features from the preceding module as input and dynamically focuses on authentic tampering boundaries in a data-driven manner. Finally, to effectively integrate and reorganize these multi-stage enhanced and refined features in the decoder, we propose a novel cross-scale attention mechanism (MGFA). This mechanism facilitates the seamless fusion of features from different scales and semantic levels, thereby ensuring the global consistency and boundary precision of the final segmentation mask.
In summary, the main contributions of this paper are as follows:
- We propose a novel model architecture for image splicing detection, built upon a “spatial-guided frequency perception” philosophy and implemented within a single-backbone synergistic framework. This architecture fundamentally differs from traditional dual-branch networks, which suffer from isolated feature extraction and late-stage fusion. Instead, our framework uses deep spatial semantics to actively guide the perception and enhancement of high-frequency artifacts within a unified feature stream, which is realized through a combination of supervision with an edge loss, adaptive frequency perception, and multi-scale feature fusion.
- We design a CSDRFBlock that employs a progressive cascade mechanism to drive the dynamic inheritance and reinforcement of multi-scale features. This avoids the feature conflicts typical of dual-branch fusion and significantly improves the cross-level transmission efficiency of critical clues, such as small-scale tampering artifacts.
- We design an MGFA mechanism. Through strategies including bi-directional channel-spatial feature calibration, local-global semantic focusing, and dynamic weight aggregation, it establishes strong relationships between tampered regions and the background within multi-scale feature maps, effectively mitigating issues of blurred boundaries and semantic disconnection.
- We design an ESFFM with an ASF at its core. This module learns to optimally separate high- and low-frequency tampering artifacts in a data-driven manner, overcoming the limitations of traditional fixed filters and significantly improving the model’s localization precision for tampering boundaries.
2. Related Work
The core strategy of traditional methods focuses on the extraction and analysis of specific image fingerprints, which are mainly divided into three technical approaches. The first is based on sensor pattern noise (SPN) [1,2], which performs source detection by analyzing the inherent noise characteristics of imaging devices; however, a convergence of noise features among similar devices can lead to detection failure. The second is based on the quantization effects of Discrete Cosine Transform (DCT) coefficients. Zhao et al. (2014) [4] proposed an image splicing forgery detection method based on a two-dimensional Markov model, which is applied in the block DCT domain and the Discrete Multiwavelet Transform (DMWT) domain, and utilizes cross-domain characteristics as classification features. Although this method offers enhanced robustness against median filtering and JPEG compression, it is inherently complex. The third category of methods focuses on detecting traces introduced by geometric transformations. For example, Popescu and Farid [22] proposed a method to detect image resampling (such as scaling and rotation). This technique identifies tampered regions by analyzing the periodic correlations among pixels. Since these traditional methods primarily target a single type of tampering feature and rely on the design of handcrafted statistical features, they can be easily circumvented by targeted manipulations, resulting in suboptimal detection performance.
In recent years, deep learning methods have been implemented to solve numerous scientific problems and have gained immense popularity. This is because their performance in classification problems, regression problems, and segmentation challenges has proven to be outstanding. Consequently, many deep learning-based methods for image tamper detection have been proposed. Initially, researchers focused on using convolutional neural networks for splicing tamper localization at the patch level. Bappy et al. [10] proposed a hybrid architecture combining a CNN encoder with an LSTM to achieve tampered region localization by simultaneously extracting spatial and frequency-domain features. Experiments showed that it could achieve pixel-level tamper localization without relying on handcrafted feature design, surpassing traditional handcrafted feature methods in detection accuracy. Xiao et al. [11] proposed a coarse-to-fine convolutional neural network (C2R-Net) based on non-overlapping image patches, which segments the input image into fixed-size, non-overlapping patches, uses a CNN to extract multi-scale features, and combines this with an adaptive clustering algorithm to localize tampered regions. However, such methods require segmenting the image into independent small patches for processing, leading to significant redundant computation, and the local field of view of a single patch makes it difficult to capture global semantic associations.
The proposal of Fully Convolutional Networks (FCNs) directly addressed the aforementioned bottlenecks. Through a fully convolutional encoder–decoder architecture, they achieve end-to-end mapping from an entire input image to a pixel-level output, completely abandoning the patching operation. Liu and Pun [15] proposed a splicing tamper localization model based on a cascade of a multi-scale Fully Convolutional Network and a Conditional Random Field (FCNs–CRF). It extracts deep features using three FCNs with different scale sensitivities and combines them with a CRF to adaptively fuse the multi-scale prediction results. However, FCNs–CRF lacks the ability to learn the tampered region’s edges, leading to suboptimal feature fusion performance. Considering that image splicing detection is essentially a pixel-level binary classification task, Bi [13] proposed a method combining a Ringed Residual U-Net (RRU-Net). RRU-Net constructs cross-level feature interaction channels through forward propagation in residual paths and backward connections in residual feedback paths, achieving multi-level reuse of features within the network. However, experimental results showed that on the COLUMBIA dataset, the detected tampered regions exhibited discontinuities and blurred boundaries, indicating a limited ability to capture fine-grained structures. Although the ManTra-Net model proposed by Wu et al. [14] extracts feature patterns of 385 known manipulation types through self-supervised learning, it lacks detailed information for localizing the tampered region.
In recent years, multi-scale strategies have been widely applied across various computer vision tasks. For instance, models such as DDMSNet [23] and GridFormer [24] adopt strategies that process input image pyramids, while others like ESTINet [25] construct feature pyramids internally through encoder–decoder structures. These approaches all aim to promote robust fusion between high-resolution details and low-resolution global context. However, strategies like DDMSNet and GridFormer that process multi-resolution inputs may introduce interpolation artifacts during the image scaling process, thereby interfering with or corrupting original traces of tampering. Another class of methods, such as MB-TaylorFormer V2 [26], innovates at a finer granular level by generating multi-scale representations at the token level within Transformer modules, thereby efficiently obtaining flexible receptive fields. Although these multi-scale architectures perform excellently in their respective tasks, their design philosophy is fundamentally intended to handle global or semi-global natural degradations (such as rain, snow, haze), and their inherent feature fusion logic presents a fundamental conflict with the requirements of image splicing forensics tasks. In image restoration tasks, the dominant role of multi-scale fusion is played by low-frequency global information: networks utilize stable scene structures (low-frequency) as priors to guide and constrain the reconstruction of high-frequency details, with the goal of enforcing consistency between high and low-frequency features to generate visually coherent images. However, in splicing forensics, the key evidence is precisely those high-frequency artifacts (such as splicing seams and noise anomalies) that are inconsistent with the surrounding low-frequency semantic background. Therefore, networks designed for restoration tasks, with their inherent bias toward pursuing “consistency”, tend to treat these high-frequency tampering traces as “noise” and smooth or suppress them, which runs counter to the forensic objective of preserving and amplifying “inconsistencies”. Meanwhile, strategies like DDMSNet and GridFormer that rely on input image scaling may themselves corrupt the extremely fine forensic fingerprints originating from device-level sources through their interpolation processes.
With the growing popularity of attention mechanisms and dual-branch networks, some methods add attention mechanisms to the backbone network, while others prefer to use a dual-branch backbone to extract different types of image features. Zhou et al. [19] employ a dual-stream feature synergy framework that achieves cross-modal detection and localization of tampered regions by extracting low-order noise residuals (sensor pattern noise) and high-order semantic features (deep CNN representations) in parallel and fusing them via feature concatenation. D-Net [21] adopts a dual-encoder architecture (with a fixed encoder extracting pixel structure information and a learnable encoder learning image fingerprints) and synergistically enhances the comprehensive detection capability for tampered regions through multi-level feature fusion and a spatial pyramid global feature enhancement (SPGFE) module. The ViT-VAENet model, proposed in [27], combines the self-attention mechanism of the Vision Transformer (ViT) with a Variational Autoencoder (VAE). It leverages ViT’s global context awareness and VAE’s latent feature generation capabilities to achieve localization of tampered regions in images without requiring a fixed input size. Subsequently, an increasing number of researchers have focused on methods that combine image features with edge features. MVSS-Net [5] achieves effective fusion of noise and edge features by using a Sobel filter and an additional edge-supervision branch. EMF-Net [28] better localizes tampered regions by combining RGB image features and edge features. Building on the previous two, UGEE-Net [29] adds uncertainty guidance to improve the model’s sensitivity to subtle tampering.
3. Methodology
To address the limitations of prior work, particularly the information isolation inherent in dual-branch architectures, we propose a novel methodological framework in this section. Unlike paradigms that process different features in separate streams, our approach is built upon a single-backbone synergistic enhancement design. The core philosophy is to create a unified process where high-level semantic understanding and low-level artifact perception are not treated as separate tasks but are deeply integrated. This is achieved through a cascade of specialized modules designed for multi-scale feature preservation, adaptive edge-aware analysis, and effective feature fusion, which will be detailed in the following subsections.
3.1. Overall Architecture
As shown in Figure 2, the proposed AMSEANet adopts a standard encoder–decoder architecture to efficiently learn feature distributions at various scales. The entire AMSEANet is primarily divided into three stages: semantic feature enhancement, image feature extraction, and edge feature perception. In the semantic feature enhancement stage, we use a CSDRFBlock before the first down-sampling layer to prioritize the processing of high-resolution features. This pre-emptively reinforces and preserves multi-scale contextual information before it enters the deep network, laying a solid foundation for subsequent tampering artifact detection. In the image feature extraction stage, we employ two standard residual blocks from ResNet-18 to perform deep feature extraction on the feature map output by the preceding module and to further extract image features, passing a feature map rich in semantic and image information to the next stage of edge feature perception. In the edge feature perception module, we utilize ESFFM to dynamically perceive and enhance tampering-related edge artifact information within the feature stream. Finally, skip connections are added between the down-sampling and up-sampling processes, and MGFA is used to strengthen the relationships among feature information from different scales and from different receptive fields within the same scale. This bridges the gap between high-level semantics and low-level details, ensuring the precision of the final segmentation mask.
Figure 2.
The overall architecture of the proposed AMSEANet. The network follows a standard encoder–decoder structure, incorporating three main innovative modules: the Cross-Scale Dense Residual Fusion Block (CSDRFBlock) for multi-scale feature enhancement, the Edge-Aware Spatial-Frequency Fusion Module (ESFFM) for guided trace perception, and the MGFA mechanism in the skip connections.
3.2. Multi-Scale Global–Local Fused Attention
To address the issues of semantic information loss during feature extraction in the encoder architecture, weakened relationships between multi-scale features, and the lack of diversity in receptive fields at the same scale, this paper proposes a multi-scale feature enhancement attention mechanism. Its core consists of three stages that operate between the encoder and decoder: Multi-Scale Interaction, Local Receptive Field Enhancement, and Global Context Modeling. The structure is shown in Figure 3. The Multi-Scale Interaction module overcomes the limitation of traditional attention mechanisms that rely only on features from a single encoder layer by implementing cross-level feature calibration through a bi-directional channel-spatial attention mechanism to identify tampered regions. The Local Receptive Field Enhancement module is used to capture tampering trace features within the local context, while the Global Context Modeling module is responsible for extracting global features at the current level.
Figure 3.
The architecture of the proposed MGFA module.
3.2.1. Multi-Scale Feature Interaction
Given the high-level feature input from the n-th encoder layer and the low-level feature input from the -th encoder layer, a channel attention (CA) mechanism is first applied to . A channel attention weight vector is generated via global average pooling and a convolution to facilitate cross-channel information selection. Subsequently, a convolution compresses the channel dimension to C, and the result is upsampled to resolution, yielding the refined low-level features . Finally, and are concatenated along the channel dimension to obtain the fused features , thereby achieving cross-scale feature recombination. A spatial attention (SA) mechanism is then employed to compute a pixel-wise attention map, strengthening the response in key regions. Finally, a convolution reduces the channel dimension to C, outputting the multi-scale fused features .
3.2.2. Local-Global Feature Focusing
To capture tampering trace features and local artifacts within the local context, a dilated convolution with a dilation rate of 2 is applied to the original features . A gating mechanism is then used to perform feature selection and suppress irrelevant noise, outputting the locally enhanced features with a channel dimension of .
Concurrently, a convolution and a are employed to extract cross-channel global dependencies from , generating globally enhanced semantic features with a channel dimension of .
3.2.3. Dynamic Feature Aggregation
The multi-scale features , local features , and global features are concatenated along the channel dimension. A convolution followed by a activation function is then applied to integrate cross-dimensional information, outputting the integrated features . Subsequently, an adaptive attention map is generated using a Sigmoid function. The calculation process is formulated as follows:
where denotes the Sigmoid function and represents the channel-wise concatenation operation. Finally, the features are enhanced via a residual connection:
where ⊗ denotes element-wise multiplication and ⊕ denotes element-wise addition. This design, through progressive weight optimization, preserves the integrity of the original features while strengthening cross-scale feature relationships and enabling comprehensive modeling of features from multiple receptive fields at the same scale.
3.3. Cross-Scale Dense Residual Fusion Block
Image splicing operations typically introduce visual artifacts, specifically manifested as unnatural edges and texture discontinuities in the tampered region. These artifacts provide crucial clues for the precise localization of the forgery. However, traditional convolutional neural networks (CNNs) perform poorly in the task of identifying and localizing tampered regions due to their limited receptive fields and the insufficiency of feature map information, making it difficult to effectively capture the global contextual information of the forged area. To this end, inspired by EfficientNet [30], this paper proposes an improved CSDRFBlock, which aims to enhance the network’s detection and localization accuracy for tampered regions by fusing feature information from different scales.
As shown in Figure 4, the multi-scale feature extraction module, CSDRFBlock, is primarily composed of three components: a multi-scale feature generation layer, a cross-scale interaction branch, and a bottom-up residual enhancement module. The multi-scale feature generation layer decomposes the input feature map into a high-resolution feature map , a medium-scale feature map , and a low-scale feature map through two progressive downsampling operations (indicated by red arrows). To ensure that each scale can maximally leverage the feature information from higher-level scales, the CSDRFBlock incorporates an inter-level feature inheritance mechanism, namely the cross-scale interaction branch. During processing in the scale space (where ), each scale level , prior to performing feature extraction at its own scale, receives a compressed feature map from the higher level via a downsampling operation.
Figure 4.
The architecture of the proposed CSDRFBlock. The blocks on the left represent the input multi-scale feature maps. The module processes them through its three main components: a multi-scale feature generation layer, a top-down cross-scale interaction branch, and a bottom-up residual enhancement path to produce the fused output. The red arrows indicate downsampling operations, while the dark red arrows within the residual path represent upsampling operations.
The process can be formulated as:
Here, denotes a spatial downsampling operation that includes channel compression, with the spatial dimensions defined as and . This feature is then concatenated with the original input feature at the current scale along the channel dimension. This operation is represented as follows, where denotes channel-wise concatenation:
To mitigate the gradient vanishing problem in deep networks, the module employs a bottom-up residual aggregation path. The output of each layer is upsampled (denoted by ) and then added element-wise to the output of the preceding higher-level layer, ultimately achieving multi-dimensional enhancement of the required features. For the i-th branch (where ), its output features are upsampled to be fused with the features from the preceding higher-level scale. The upsampling process for each branch is defined as:
Finally, the features from each scale are aggregated through element-wise addition to produce the final output, x:
3.3.1. SimpleGate
SimpleGate is a lightweight yet efficient non-linear feature transformation module. Its core idea is to achieve adaptive feature modulation through channel splitting and element-wise multiplication (Hadamard product). As illustrated in Figure 5, given an input feature tensor , where B represents the batch size, C is the number of channels, and H and W represent the spatial height and width, respectively, x is first split into two equal sub-tensors along the channel dimension: .
Figure 5.
A diagram of the structure of SimpleGate.
The gating modulation operation achieves feature interaction through the Hadamard product (⊗), and the final output is given by:
By splitting the input feature map evenly into two parts, and , along the channel dimension, SimpleGate forces the network to decouple two independent sets of semantic representations during the feature propagation process. This splitting essentially divides the original feature space into two complementary subspaces. The subsequent element-wise multiplication (Hadamard product) is not merely an act of information compression but rather constructs a non-linear cross-subspace interaction mechanism. The multiplication of and is equivalent to establishing an adaptive feature-selection gate along the channel dimension, enabling the network to automatically identify key response areas within both feature sets that are sensitive to the current task. Unlike traditional activation functions (such as ), which perform fixed-threshold truncation, SimpleGate achieves a dynamic non-linear transformation through the multiplication of feature subspaces.
3.3.2. Simplified Channel Attention Block
As illustrated in Figure 6, the SCABlock introduces a lightweight channel attention mechanism designed to generate channel weights through global average pooling (GAP) and a series of convolutional operations, thereby effectively extracting salient features. Specifically, to stabilize the feature distribution and accelerate training, the module first normalizes the input feature map across the batch dimension using BatchNorm and then uses a convolution to extract high-level features. Subsequently, a depth-wise convolution (DConv) and a standard convolution are employed to further extract spatial features, which are then activated through a SimpleGate (SG) module. Next, global information is acquired via global average pooling, and channel weights are subsequently adjusted using a convolution followed by Layer Normalization, which performs normalization across the channel dimension. These weights are then multiplied by the original feature map to achieve dynamic channel-wise weighting. This design not only effectively captures the interdependencies between different channels but also enhances the feature representation capability while maintaining a lightweight model architecture.
Figure 6.
The architecture of the proposed Simplified Channel Attention Block (SCABlock) and the Multi-Receptive Field Block (MRFBlock).
The SCABlock ensures that key features are enhanced and redundant information is suppressed through multiple stages of feature calibration. Specifically, this includes an initial feature extraction stage, where features are preliminarily refined using BatchNorm and a convolution; a deep feature enhancement stage, which utilizes a depth-wise convolution and a standard convolution to further enhance feature representation; a global information fusion stage, where GAP and a subsequent convolution integrate global information to generate a channel attention weight vector; and a final feature calibration stage, where Layer Norm and a convolution perform the final calibration to obtain a precise feature vector, ensuring that the output features have a stronger representative capacity. Given an input feature tensor , the feature outputs from the four stages (where ) can be, respectively, represented as:
3.3.3. Multi-Receptive Field Block
As illustrated in Figure 6, the Multi-Receptive Field Block (MRFBlock) effectively captures both local and global image information by designing multiple parallel branches to process features at different scales. Specifically, the MRFBlock contains three main branches, which, respectively, employ two consecutive layers of and depth-wise convolutions (DConv), and a standard convolution, combined with convolutions and the SimpleGate (SG) activation function, to extract features from different receptive fields. After their respective convolutional operations, the branches exchange information via a cross-connection mechanism, which enhances the diversity and complementarity of the features. To ensure the thorough fusion of features from different scales, the MRFBlock is designed with an efficient internal information interaction mechanism. This includes cross-connections between the branches and the fusion of branch features using a convolution and Layer Normalization to generate the final multi-scale feature map. Additionally, the final output of each layer within the MRFBlock is aggregated through bottom-up element-wise addition, ultimately achieving effective feature enhancement. Given the output features from the SCABlock, they are passed through Layer Normalization and a point-wise convolution before being fed into three parallel branches to obtain contextually rich features . The entire process can be represented as follows:
This bottom-up feature aggregation strategy not only enhances the model’s performance on the image splicing detection task but also ensures the efficiency and robustness of the feature extraction process
3.4. Edge Feature Perception Module
This module is constructed from two series-connected Edge-Aware Feature Enhancement Blocks (EFEBlocks), detailed in Figure 7, where each block comprises a core ESFFM module and a ConvBlock. The design of the core ESFFM is elaborated below.
Figure 7.
The architecture of the Edge-Aware Feature Enhancement Block (EFEBlock). The block is composed of two ESFFM modules and a ConvBlock in series, with a weight-sharing strategy applied between corresponding modules.
3.4.1. Edge-Aware Spatial-Frequency Fusion Module
In the image tamper detection task, the boundary between the tampered and authentic regions often contains critical traces of forgery. These traces can manifest as unnatural edge transitions, anomalous noise patterns, or subtle textural inconsistencies, and they are typically rich in frequency-domain information. For instance, EMF-Net [28] employs the Spatial Rich Model (SRM) [31], a classic forensic method that uses a set of handcrafted high-pass filters to extract noise-like features, to capture high-frequency information. However, the weights of SRM are predefined and cannot be learned or updated. To overcome this issue, MSNP-Net [18] employs a learnable constrained convolution [32] to extract high-frequency features; however, during the learning process, it tends to over-rely on high-frequency features while neglecting low-frequency ones. The research by UGEE-Net [29] demonstrated that low-frequency information, which typically includes pixels with slight intensity variations, is equally important for image splicing localization. The subtle pixel intensity changes in smooth regions, contained within low-frequency information, are crucial for a complete understanding of the image content, identifying semantic inconsistencies, and localizing tampered regions that have undergone smoothing operations.
To effectively capture and enhance these indicative edge features, and to fully leverage information from different frequency domains, we have designed the ESFFM. As illustrated in Figure 8, the core of this module lies in the introduction of a learnable ASF, which dynamically decomposes the input features into their low-frequency (smooth) and high-frequency (detail) components based on the data. Subsequently, these separated components are combined with the original feature information and refined through subsequent convolutional and attention mechanisms, thereby achieving a comprehensive enhancement of the indicative edge tampering features.
Figure 8.
Detailed architecture of the Edge-Aware Module (ESFFM) and its core component, the Adaptive Smooth Filter (ASF). The diagram illustrates how ESFFM uses frequency decomposition to generate an attention map for feature enhancement. The dashed box details the working principle of its core component, the ASF, which dynamically generates a smoothing kernel from learnable parameters.
3.4.2. Adaptive Smooth Filter
The ASF aims to learn an optimal low-pass filtering operation in a data-driven manner. This unit contains a set of learnable raw weight parameters , where is the spatial size of the convolutional kernel (e.g., ). These parameters are transformed via a Softmax function to generate an effective smoothing (low-pass) convolutional kernel . For the raw weight kernel of each channel i, its corresponding effective kernel weight is calculated as follows:
where is the flattened form of , and is a mapping from the 2D index to a 1D index. This Softmax transformation ensures that possesses the key mathematical properties of a local weighted averaging operator: non-negativity () and normalization (i.e., the weights sum to one, ). This makes the subsequent convolution operation, (where * denotes the convolution operation, is the feature map and is a learnable bias), essentially an adaptive local weighted average applied to each input channel , thereby yielding its low-frequency component . The properties of this low-pass filtering operation, , are determined entirely by the data and the learning objective, rather than being predefined. Correspondingly, the high-frequency component can be obtained through a difference operation:
where is the identity operator, and is the complementary, implicitly learned adaptive high-pass filtering operator to . To ensure the smoothing kernel learned by the Adaptive Smooth Filter is robust and possesses favorable mathematical properties, we introduce an optional regularization term. This regularization term guides the learning by minimizing the mean squared error between the learned effective smoothing kernel and a predefined target Gaussian kernel (with a standard deviation of ):
This regularization can be viewed as introducing a Gaussian prior centered at into the parameter space of . This encourages the model to favor the formation of filters with classic smoothing characteristics while optimizing the main task loss.
After obtaining the adaptive low-frequency component and high-frequency component , the ESFFM module integrates this information through a specific fusion strategy. In the current design, we first precisely reconstruct through the operation . This reconstructed feature is then concatenated with the original along the channel dimension, forming . This concatenation allows a subsequent convolutional layer to perform a weighted combination and dimensionality reduction, enabling it to learn how to optimally utilize these frequency-analyzed features derived from .
3.5. Multi-Scale Edge Dice Loss
The traditional Dice loss function [33], as illustrated in Figure 9, is primarily used to measure the overall degree of overlap between the predicted mask and the ground truth mask. A significant limitation of this method is that even suboptimal segmentation results with obvious inaccuracies can still produce a considerable overlap area. Consequently, the Dice loss value can sometimes yield misleadingly high scores, which diminishes its effectiveness as a guiding metric for model optimization, especially when dealing with cases of partially correct predictions.
Figure 9.
Diagram illustrating the principle of the traditional Dice loss function. The method measures the quality of a segmentation result by comparing the Predicted Forgery Mask and the Ground Truth Forgery Mask based on their intersection and non-overlapping regions.
Therefore, this paper proposes an improved Dice loss function, which we term the Multi-Scale Edge Dice Loss (ME-Dice). This function enhances the model’s sensitivity to key traces of forgery by applying multi-scale gradient supervision to the tampering boundaries.
In our preliminary explorations, we found that directly using the standard 2D Sobel operator to compute the edge loss yielded unsatisfactory results. We analyze that this is because the standard 2D Sobel operator structurally couples differentiation with a smoothing operation. This inherent smoothing property, while effective for suppressing noise in natural images, can conversely blur or weaken the sharp yet subtle traces of forgery that serve as key evidence in the tamper detection task. It must be emphasized that rejecting the smoothing property at the loss function level is to ensure the absolute precision of the supervisory signal. This does not contradict our strategy of using the ASF module (Section 3.4.2) to adaptively separate and utilize high- and low-frequency information during the feature extraction stage. The goal of the latter is to provide the network with rich feature representations for analysis; the roles they play and the stages at which they operate within the network are entirely different.
To address the aforementioned problem, we propose a strategy that decouples smoothing from differentiation, directly employing pure one-dimensional (1D) central difference convolutional kernels to compute the gradient, thereby maximally preserving the original sharpness of the edges. Furthermore, considering that the diversity of tampering techniques can lead to boundaries exhibiting different characteristics, we have designed two sets of 1D convolutional kernels at different scales to achieve a multi-scale perception of the tampering boundaries.
The two sets of convolutional kernels we use are defined as follows:
- Small-scale () convolutional kernel:Used to capture the highest-frequency and sharpest pixel-level edge transitions.
- Large-scale () convolutional kernel: Used to perceive edges that have become relatively smooth due to post-processing, such as slight blurring or compression.
Let I be the input image, P be the prediction, and T be the target (ground truth). The edge detection process involves convolving the image with horizontal () and vertical () convolutional kernels.
Next, the gradient magnitude is obtained by:
and are the predicted and target edge maps obtained using the convolutional kernel, respectively. The same logic applies to the convolutional kernel.
Finally, we introduce a weighting factor to balance the two edge losses from different scales. In our experiments, we empirically set to 0.5, aiming to assign equal importance to both the fine, high-frequency edge features (captured by the kernel) and the smoother, wider-span edge features (captured by the kernel). The total edge loss function, , is calculated as follows:
4. Experiments and Results Analysis
To evaluate the performance of the proposed AMSEANet, we compared it with several other image splicing forgery detection methods. We conducted a series of comparative experiments to benchmark its overall performance and ablation studies to demonstrate the effectiveness of its main components. Furthermore, robustness experiments were carried out to verify the stability of our network against various common attacks.
4.1. Datasets
The proposed image splicing forgery detection method is analyzed and evaluated on four public datasets: CASIA v1.0, CASIA v2.0 [34], NIST’16 [35], and COLUMBIA [36]. To prevent result bias caused by source image overlap, we adhere to a strict source-based partitioning principle for all datasets. Specifically, we ensure that all tampered samples derived from the same original authentic image are assigned exclusively to a single data split (i.e., training, validation, or test set). This strategy is designed to prevent the model from learning source-specific, non-tampering features, thereby allowing for a fairer evaluation of its generalization capability in detecting common tampering artifacts. In the COLUMBIA dataset, the spliced regions are large and simple, resulting in low detection difficulty. Compared to CASIA v1.0, CASIA v2.0 contains many more subtle forgeries. The forged images in the NIST’16 dataset have been post-processed to conceal any visible tampering traces. To ensure a comprehensive and targeted evaluation, the experiments in this paper employ two different setups. The first setup is used for the main performance evaluation, covering the “Comparison with State-of-the-Art” in Section 4.4, the “Ablation Study” in Section 4.5, and the “Effectiveness Validation of the Multi-scale Edge Dice Loss” in Section 4.6. These experiments are conducted on the CASIA v2.0, NIST’16, and COLUMBIA datasets, and data augmentation techniques such as random flipping, scaling, and rotation are applied during the training process. The second setup is used for the performance validation of a specific module, namely the “Performance Comparison of Different Frequency-Domain Information Fusion Strategies” in Section 4.7. To ensure a pure evaluation of the module’s performance, this experiment uses the CASIA v1.0 and COLUMBIA datasets with a 65% training and 35% testing split, and no data augmentation strategies are employed in this phase. Table 1 summarizes the detailed dataset partitioning for the first main experimental setup and also lists the specific attack parameters used in the subsequent “Robustness Tests” in Section 4.8.
Table 1.
Dataset partitioning and parameter settings for robustness tests.
4.2. Evaluation Metrics
To quantitatively evaluate the performance of our proposed AMSEANet model and other comparative methods on the task of image splicing forgery detection, we treat this task as a pixel-level binary classification problem (i.e., classifying each pixel as either “tampered” or “authentic”). We employ a series of widely recognized evaluation metrics, including Precision, Recall, and the F-measure. These metrics are calculated based on the following three fundamental quantities: the number of tampered pixels correctly detected as tampered (True Positives, TPs); the number of authentic pixels incorrectly detected as tampered (False Positives, FPs); and the number of tampered pixels incorrectly identified as authentic (false negatives, FNs).
4.3. Implementation Details and Loss Function
Our model is implemented in PyTorch 2.4.1 and trained on a GeForce GTX 5070Ti GPU. To validate the model’s learning capability, we trained it from scratch without using any pre-trained weights. We use the Adam optimization algorithm to optimize the model, with an initial learning rate of 0.0002, beta parameters set to (0.5, 0.999), and a batch size of 8. The model is trained for a total of 150 epochs. A cosine learning rate scheduler is employed, and all input images are resized to pixels. Notably, the choice of the initial learning rate is critical for the model’s final performance. Our proposed ME-Dice edge loss significantly accelerates model convergence by providing fine-grained boundary supervision. It was observed experimentally that an excessively low learning rate could cause the model to converge prematurely to a local optimum without the ability to escape. Therefore, a moderate learning rate of 0.0002 was chosen to ensure that the model, while converging rapidly, retains sufficient exploratory momentum to avoid suboptimal solutions, thereby finding a more optimal parameter space during training.
To effectively train the proposed AMSEANet model, we designed a composite loss function, , which is a weighted combination of the binary cross-entropy loss (), our proposed Multi-Scale Edge Dice Loss (), and the regularization loss () from the Adaptive Smooth Filter (ASF). The total loss function is defined as follows:
where , , and . Since itself is relatively small, the corresponding weight is set to a larger value; however, its impact on the overall loss function is negligible.
4.4. Comparison with State-of-the-Art Methods
To evaluate the performance of the proposed AMSEANet, we selected two traditional methods and five advanced deep learning methods as comparison algorithms: NOI [37], CFA [38], ManTra-Net [14], RRU-Net [13], C2R-Net [11], CAT-Net [39], and PSCC-Net [40].
As shown in Table 2, traditional detection methods such as NOI achieve high Recall but suffer from low Precision and F-measure. This is typically attributed to the misclassification of numerous irrelevant regions as tampered, as an excessively high Recall score loses its significance without a correspondingly high Precision. Among the deep learning-based methods, CAT-Net shows a balanced performance, notably achieving the highest Precision (0.924) on the NIST’16 dataset. This suggests the model’s strength in suppressing false positives, though its slightly lower Recall indicates it may miss some traces of tampering.
Table 2.
Performance comparison of our proposed method (Ours) against several state-of-the-art forgery detection methods. The best results are highlighted in bold.
AMSEANet demonstrates competitive and comprehensive performance in these comparative experiments. On all three datasets, our model achieves the highest F-measure while maintaining a good balance between Precision and Recall. Our method achieves optimal performance on the COLUMBIA dataset, which can be attributed to AMSEANet’s simultaneous capabilities for deep semantic understanding and edge detection, enabling it to easily identify large tampered regions. On the CASIA v2.0 dataset, AMSEANet obtains the highest Precision (0.887) and F-measure (0.877). It is noteworthy that our model’s Recall (0.865) is marginally lower than that of CAT-Net (0.869); however, our Precision is substantially higher. We believe this highlights a performance trade-off on this specific dataset: CAT-Net’s higher Recall may come at the cost of lower precision, reflected in its relatively low Precision of 0.833 on this dataset, which can lead to more false positives. In contrast, AMSEANet, while maintaining a high Recall, demonstrates a stronger ability to distinguish between subtle artifacts and complex authentic textures. This gives it an advantage in providing high-confidence predictions and reducing misclassifications, ultimately leading to the best overall performance in the F-measure.
The robustness of AMSEANet is thoroughly validated on the NIST’16 dataset. Images in this dataset typically undergo complex post-processing designed to conceal overt traces of tampering, posing a significant challenge to models that rely on local edge or low-level artifacts. Consequently, we observe a notable performance degradation for some models, such as RRU-Net, whose F-measure drops from 0.843 on CASIA v2.0 to 0.765 on NIST’16. This suggests that the detection capabilities of these models are substantially compromised when low-level tampering features are suppressed. In contrast, AMSEANet demonstrates exceptional stability in this adversarial context, achieving an F-measure of 0.919 and a Recall of 0.921, both of which are the highest among all compared methods. This performance gap is attributed to our model’s advanced architecture. The effectiveness of our key modules, validated in the ablation study in Section 4.5, directly translates to this robust performance. When low-level edge cues become ambiguous or unreliable, AMSEANet can more effectively integrate high-level semantic context. For instance, the CSDRFBlock provides a high-fidelity feature foundation for the model; our ablation study has already demonstrated it to be the most critical component, improving the F-measure from 0.787 to 0.834 on its own (see Table 3). Building upon this, the ESFFM and MGFA modules further refine these features, culminating in the model’s superior ability to analyze the logical and structural plausibility of image content from a global perspective. Therefore, even when traces of tampering are intentionally concealed, our method can still effectively identify and localize the majority of forged regions, demonstrating its application potential in complex scenarios. To provide a more intuitive comparison of the localization performance of the various methods, we visualize the results on several representative samples in Figure 10.
Table 3.
Results of the ablation study on the core components of AMSEANet, conducted on the CASIA v2.0 dataset. The checkmark (🗸) indicates that the module is included in the configuration, while the cross (×) indicates its exclusion.
Figure 10.
Tamper localization results of AMSEANet and other methods on public datasets. The rows, from top to bottom, show two representative samples each from the CASIA v2.0, NIST’16, and COLUMBIA datasets, respectively.
4.5. Ablation Study
To systematically validate the incremental contributions of each core component in our proposed AMSEANet, we designed a series of ablation studies and comparative experiments. These experiments were conducted to verify the effectiveness of the CSDRFBlock, the ESFFM, the MGFA mechanism, and the ME-Dice. All experiments were performed on the challenging CASIA v2.0 dataset using the same training configuration to ensure a fair comparison. The detailed experimental results are presented in Table 3.
We first establish a baseline model (Baseline), a standard encoder–decoder network without any of our proposed modules, which achieves an F-measure of 0.709. By solely incorporating our proposed ME-Dice loss for edge supervision (the Edgebase model), the F-measure significantly improves to 0.787. This provides strong evidence that applying explicit supervision to tampering boundaries effectively guides the model to focus on critical artifacts, thereby substantially reducing false positives. Building upon the Edgebase, we further evaluated the individual contributions of the core modules. The introduction of CSDRFBlock yielded the largest performance gain, increasing the F-measure to 0.834. This indicates that CSDRFBlock, through its efficient cross-scale feature fusion and residual connection mechanisms, builds more robust multi-level feature maps, significantly enhancing the network’s ability to perceive and capture tampered regions of various sizes. When the ESFFM module was introduced (F-measure of 0.811), the performance improvement was primarily attributed to the spatial-frequency decomposition performed by its integrated ASF, which enables the network to accurately isolate and focus on high-frequency edge artifacts indicative of tampering. This study also reveals the synergistic effects between modules, particularly the interaction between CSDRFBlock and MGFA. By comparing the Edgebase + MGFA + ESFFM model (F-measure of 0.839) with our full model (F-measure of 0.877), it is evident that adding CSDRFBlock to the former configuration provides a substantial performance boost. This clearly demonstrates that CSDRFBlock serves as a foundational module providing high-fidelity feature streams for the subsequent MGFA. Building on this foundation, MGFA can then effectively strengthen the correlations between features from different scales and receptive fields to ensure optimal fusion of high-resolution boundary details, with both modules working in concert to unleash their full potential. These results ultimately validate our design rationale: each of our proposed modules is individually effective, and the synergy between them is key to the model’s final leading performance.
4.6. Effectiveness Validation of the Multi-Scale Edge Dice Loss
To systematically validate the effectiveness of our proposed ME-Dice and demonstrate its superiority over traditional loss functions, we designed the following ablation study. All experiments were conducted on the CASIA v2.0 dataset, with the baseline model uniformly using the Binary Cross-Entropy (BCE) loss as the base loss.
The experimental results in Table 4 clearly reveal the contribution of each component of our proposed loss function. First, compared to the baseline model that only uses , the addition of any edge loss improves the model’s performance. Notably, both of our proposed single-scale 1D gradient losses (with F-scores of 0.859 and 0.865, respectively) significantly outperform the edge loss using the standard 2D Sobel operator (F-score of 0.831). This strongly validates our core hypothesis: that decoupling the smoothing and differentiation operations can more effectively supervise the network in learning tampering artifacts.
Table 4.
Ablation study for the edge-guided loss function. denotes the edge loss using the standard 2D Sobel operator. and represent the edge losses using only the 1 × 3 and 1 × 5 1D gradient kernels, respectively. is the combination of and .
It is worth noting that the two single-scale gradient kernels exhibit distinct complementary characteristics. The scale loss, , achieves the highest Precision (0.924), indicating that its wider receptive field and smoother gradient calculation enable it to robustly identify tampering boundaries and avoid misclassifying noise as artifacts. Meanwhile, the scale loss, , obtains the highest Recall (0.876), demonstrating that its sensitivity to high-frequency, sharp changes allows it to maximally detect all potential tampered regions.
Ultimately, our proposed Multi-Scale Edge Dice Loss, , achieves the highest F-measure (0.877), outperforming any single component. This clearly demonstrates the synergistic effect of the multi-scale strategy. By combining the high Precision of the kernel with the high Recall of the kernel, strikes the optimal balance between precision (“getting it right”) and completeness (“finding it all”), thereby achieving the best overall tamper localization performance. Figure 11 visually demonstrates this progressive performance improvement, where the result from our final scheme most closely approximates the ground truth mask in terms of boundary details.
Figure 11.
Visual comparison of the loss function ablation study. (a) Input image and (g) ground truth. (b–f) show the prediction results under different loss function configurations, where (f) is the result from our final Multi-Scale Edge Dice loss (ME-Dice).
To more comprehensively evaluate the performance of the ME-Dice loss, we further compared it against three advanced loss functions widely used in related fields: Boundary Loss, a classic boundary-refinement loss; Focal Loss, a popular solution for addressing class imbalance; and Active Contour Loss, which is inspired by traditional contour models. As shown in Table 5, we compared these three loss functions with our final ME-Dice.
Table 5.
Performance comparison of the ME-Dice loss against other advanced loss functions. The best results are highlighted in bold.
The experimental results indicate that all three advanced loss functions achieved competitive performance, with F-measures of 0.849, 0.837, and 0.859, respectively. All of them significantly outperform the traditional Sobel-based method, demonstrating their effectiveness. Among them, Active Contour Loss exhibits the highest precision (0.894), albeit with a relatively lower recall. Despite the strong performance of these methods, our ME-Dice, specifically tailored for the tampering detection task, still achieved the best overall performance with the highest F-measure of 0.877, striking a better balance between precision and recall. This comprehensively demonstrates that our strategy of using multi-scale gradient supervision to precisely capture traces of tampering is more suitable for this task compared to general-purpose boundary refinement or class-balance losses.
4.7. Performance Comparison of Different Frequency-Domain Information Fusion Strategies
To validate the synergistic effect of multi-frequency information and the role of the Adaptive Smooth Filter, as illustrated in Figure 12, we designed a series of comparative experiments. We compare our method with several alternative frequency-domain information fusion schemes on two datasets of varying difficulty (CASIA v1.0 and COLUMBIA). The CASIA v1.0 dataset contains a total of 461 images; we partitioned it, along with the COLUMBIA dataset, into a 65% training set and a 35% testing set. It should be noted that the experimental setup here is used only for the module comparison in this section and differs from the main experiments in previous sections. All comparison schemes are implemented on the same baseline network architecture, differing only in the feature fusion module. The specific descriptions of each comparison method are as follows:
Figure 12.
Comparative diagrams of different frequency-domain information fusion strategies. The figure illustrates the six different feature fusion methods based on fixed filters (SRM, Gaussian filter) and the adaptive filter (ASFilter) used in the comparative experiments.
- Normal: As the baseline model, this method only uses basic semantic features extracted via a convolution, without introducing any additional frequency-domain information.
- SRM-AddHF: This method uses a fixed Spatial Rich Model (SRM) filter to extract high-frequency (HF) features from the image, which are then fused with the basic semantic features via element-wise addition.
- SRM-CatDiff: This method first extracts high-frequency features using an SRM filter, then calculates the difference map between them and the basic semantic features. This difference map is then concatenated with the original basic semantic features along the channel dimension.
- GaussFix-AddHF: This method utilizes a fixed Gaussian blur kernel to extract the low-frequency (LF) components of the input. The high-frequency features are obtained by subtracting the low-frequency components from the basic semantic features. Finally, the extracted high-frequency features are added element-wise to the basic semantic features.
- LS-AddHF: This method replaces the fixed Gaussian kernel with our proposed Adaptive Smooth Filter (ASF) to learn and extract low-frequency features. Subsequently, the high-frequency features are obtained via a difference operation and added to the basic semantic features.
- LS-AddLF: This method directly fuses the low-frequency features, which are adaptively learned by the ASF, with the basic semantic features via element-wise addition.
- Ours: This is the final scheme proposed in this paper. First, the ASF is utilized to adaptively extract the low-frequency component, and the high-frequency component is obtained by taking the difference from the basic semantic features. Next, the low- and high-frequency components are added together to reconstruct the features. Finally, this reconstructed feature is concatenated with the original basic semantic features along the channel dimension, aiming to provide the network with a dual view that contains both original semantics and explicit frequency decomposition information.
The experimental results in Table 6 show that the baseline Normal method performs reasonably well on both datasets, demonstrating the fundamental effectiveness of deep semantic features for the tamper detection task. However, its performance also indicates that relying solely on semantic information is insufficient to capture all tampering traces accurately. Compared to the baseline, the performance of methods using fixed filters, such as SRM-AddHF and GaussFix-AddHF, degrades. This highlights a core deficiency of fixed filters: their weights are predefined and non-learnable, rendering them unable to adapt to varying image content and diverse tampering techniques. Simply superimposing fixed high-frequency details onto semantic features can introduce task-irrelevant interference, which may confuse the model and lower its precision. Notably, the SRM-CatDiff method outperforms other fixed-filter schemes. By computing the difference between basic semantic features and SRM-derived high-frequency features, it effectively generates an “inconsistency map” that explicitly highlights potential tampering traces. Although this difference-based strategy is more informative than direct addition, its ultimate effectiveness remains constrained by the non-adaptive nature of the SRM filter.
Table 6.
Performance comparison of different frequency-domain information fusion strategies. The best results are highlighted in bold.
In sharp contrast, both the LS-AddHF and LS-AddLF methods, which employ our proposed Adaptive Smooth Filter (ASF), achieve significant performance improvements on both datasets. This strongly demonstrates that using a learnable, data-driven filter to dynamically separate high- and low-frequency information is a correct and efficient strategy. The model can adaptively adjust its filtering based on the input content to extract the features most relevant to the tampering task. Our proposed final scheme, designated as ‘ours’, stands out among all compared methods by achieving the best overall performance on both datasets. On the more challenging CASIA dataset, its Precision reaches 0.8671, significantly surpassing all other methods and showcasing a strong capability for suppressing false positives. On the less difficult COLUMBIA dataset, our scheme maintains its leading edge, even as the performance gap between the methods narrows.
4.8. Robustness Tests
To further validate the effectiveness and robustness of AMSEANet, we compared its performance against existing detection methods under various types of attacks: noise addition, JPEG compression, and resizing. It is noteworthy that none of the test sets used for the robustness experiments were included during training.
Noise is often introduced during image transmission or post-processing, which can interfere with a detection model’s ability to judge underlying artifacts. Therefore, a model’s stability against noise is an important performance indicator. In this experiment, we added Gaussian white noise with varying variances to the test images. As seen in the experimental results in Figure 13, the Recall of traditional methods increases dramatically, as they tend to classify nearly the entire image as tampered under severe noise attacks. Among the CNN-based methods, AMSEANet achieves the highest Precision and F-measure on both the CASIA v2.0 and COLUMBIA datasets, and its performance is comparable to that of CAT-Net on NIST’16, demonstrating its robustness against noise attacks across the three datasets.
Figure 13.
Comparison results under noise addition attack. The three columns represent Precision, Recall, and F-measure, respectively. (a–c) show the experiment results on the CASIA v2.0, COLUMBIA, and NIST16 datasets, respectively.
JPEG compression is a common attack used to evaluate the robustness of image tamper detection models and is ubiquitous in real-world applications. Therefore, we tested various models under different levels of JPEG compression. As shown in Figure 14, we observe that JPEG compression severely impacts the performance of all models on the CASIA v2.0 dataset, with some models exhibiting a precipitous drop in their detection rates for tampered pixels. The impact of JPEG compression is less pronounced on the COLUMBIA and NIST’16 datasets. This is because the tampered regions in COLUMBIA are large and easily discernible, while the NIST’16 dataset already consists of images that have undergone multiple post-processing steps, presenting a high level of intrinsic difficulty. Notably, the performance of CAT-Net, which is otherwise robust, degrades significantly under JPEG compression. This is likely because the JPEG attack directly disrupts the analytical basis for CAT-Net’s DCT stream by overwriting the original, fine-grained “frequency fingerprints”. In contrast, PSCC-Net constructs a dual-verification system using cross-scale relationships and Self-Correlation Calculation (SCC). By examining the consistency of image statistical features from both inter-scale and intra-scale dimensions, it can accurately decouple local, non-stationary artifacts originating from tampering from the global, stationary artifacts caused by JPEG compression, thus exhibiting more stable performance. Generally, as the Quality Factor (QF) decreases, the Precision and F-measure of most models decline. Among all the compared models, AMSEANet is less affected than its counterparts. While its Recall is slightly lower than that of the traditional methods, this is due to the high rate of false positives in the latter. Overall, AMSEANet maintains commendable robustness against JPEG compression attacks.
Figure 14.
Comparison results under JPEG compression attack. The three columns represent Precision, Recall, and F-measure, respectively. (a–c) show the experiment results on the CASIA v2.0, COLUMBIA, and NIST16 datasets, respectively.
Image scaling, a common post-processing technique, poses a unique challenge as its inherent interpolation and resampling processes can alter crucial forensic traces. As shown in Figure 15, all deep learning-based methods exhibit some performance degradation under scaling attacks. Nevertheless, our AMSEANet demonstrates considerable stability across most scenarios, maintaining a high level of performance on the CASIA v2.0 and COLUMBIA datasets. However, on the NIST’16 dataset, our model’s performance is slightly inferior to that of CAT-Net. We attribute this phenomenon to a direct conflict between the attack’s nature and our model’s core detection mechanisms. Specifically, the interpolation process tends to blur the sharp, high-frequency artifacts found at tampering boundaries. This directly compromises the effectiveness of our Edge-Aware Spatial-Frequency Fusion Module (ESFFM), which is designed to adaptively separate these very high-frequency signals. When the original tampering artifacts are weakened, or masked by new, globally consistent interpolation artifacts, the discriminative ability of the ESFFM is inherently limited. This effect is particularly pronounced on the heavily post-processed NIST’16 dataset, which explains the model’s slightly inferior performance in this specific scenario. Despite this vulnerability in a specific context, AMSEANet’s overall strong performance across different datasets and scaling ratios validates its robust design against this challenging attack.
Figure 15.
Comparison results under resize operation attack. The three columns represent Precision, Recall, and F-measure, respectively. (a–c) show the experiment results on the CASIA v2.0, COLUMBIA, and NIST16 datasets, respectively.
4.9. Computational Complexity Analysis
In addition to detection performance, the computational complexity of a model is a crucial indicator of its practical value. To this end, we evaluated the parameters, FLOPs, and inference speed of AMSEANet and compared them with several mainstream methods, with the results presented in Table 7. For an input size of 256 × 256, our model has 52.08 M parameters and a computational load of 124.86 GFLOPs. On a single NVIDIA GeForce RTX 5070 Ti GPU, its inference speed reaches 35.1 FPS.
Table 7.
Computational complexity comparison of different methods.
The comparative analysis in Table 7 reveals the design trade-offs between performance and efficiency among different models. Compared to lightweight models such as ManTra-Net and PSCC-Net, AMSEANet has a relatively larger parameter count and computational load. We believe this moderate increase in complexity is an important factor for the model to achieve better detection performance. As the results in Table 2 show, the detection accuracy of AMSEANet, especially on challenging datasets like NIST’16, outperforms these lightweight models.
On the other hand, compared to the heavyweight model MVSS-Net (146.88 M), our AMSEANet is more compact in model size and more parameter-efficient, while its practical inference speed (35.1 FPS vs. 37.3 FPS) remains at a comparable level.
In conclusion, we believe that AMSEANet makes a reasonable trade-off between performance and complexity. While ensuring a necessary level of architectural complexity, it achieves competitive detection performance and avoids the parameter redundancy of oversized models, striking a favorable balance among accuracy, efficiency, and model size.
4.10. Limitations Analysis
In our research, we noted that tampering involving large-area natural background replacements (e.g., sky, water) poses a significant challenge to current detection algorithms. Our proposed AMSEANet model makes some useful attempts to address this problem. In many test samples, the model can perform relatively accurate segmentation along the complex natural contours of foreground objects (such as mountains or buildings). As shown in Figure 16, the detection results in these cases are highly consistent with the ground truth, which preliminarily demonstrates the potential of our designed method.
Figure 16.
Successful detection results of AMSEANet on natural background replacement samples. Subfigure (a) shows the original images, while the model’s prediction (b) is highly consistent with the ground truth (c), demonstrating its effectiveness.
Of course, we must also acknowledge that the model’s performance still has room for improvement when the complexity and intricacy of these natural boundaries reach extremely high levels. We preliminarily attribute the detection difficulties in such scenarios to the following two aspects:
- Challenges from visual realism and low-contrast boundaries: This type of splicing often achieves a high degree of realism by replacing the background, which occupies a large portion of the image. Forgers use advanced post-processing techniques that result in extremely low contrast at the boundary between the forged background and the authentic foreground, making it visually seamless. This renders detection methods based on local artifacts highly prone to failure.
- Segmentation challenges from the highly complex contours of authentic foregrounds: The core difficulty of the task lies in the fact that the model does not need to segment a newly introduced forgery edge, but rather the natural, extremely complex, and irregular contour of the authentic foreground object itself. As shown in Figure 17, the fine structure of a castle turret or the minute contour of a distant person are authentic yet extremely complex details preserved from the original image. The algorithm must perfectly delineate these details, much like tracing an outline, to accurately define the replaced background area. This places extremely high demands on the model’s edge perception and precise localization capabilities.
Figure 17. Limitation analysis of AMSEANet on highly challenging samples. Subfigure (a) shows the original images, while red boxes highlight discrepancies between the prediction (b) and the ground truth (c), showing that the model may produce incomplete detections on contours with extremely fine-grained details.
As can be seen from the areas marked by red boxes in Figure 17, when faced with fine structures like the castle turret or the contour of a small, distant person, our model sometimes produces partial missed detections (false negatives) and incomplete results. This indicates that while AMSEANet can handle complex boundaries overall, there are still shortcomings in its precision for capturing fine-grained details.
In summary, although AMSEANet has made some exploratory efforts and achieved certain results in handling natural background replacement tasks, there is still room for improvement in its robustness and precision when faced with complex, authentic contours featuring extremely fine-grained details and very low contrast. This also points the way for our future research: how to enhance the model’s comprehensive perception of weak signals and its segmentation precision for ultra-fine-grained structures.
5. Conclusions
To address the challenges of image splicing tamper detection, this paper proposes an Adaptive Multi-Scale Edge-Aware Network (AMSEANet). This network employs a novel single-backbone synergistic enhancement framework, which discards the traditional parallel dual-branch structure and achieves a deep integration of several key technologies within a unified feature stream. Specifically, the framework first ensures the high-fidelity transmission of micro-level tampering traces within the deep network via the Cross-Scale Dense Residual Fusion Block (CSDRFBlock). Subsequently, the Edge-Aware Spatial-Frequency Fusion Module (ESFFM) utilizes its integrated data-driven Adaptive Smooth Filter (ASF) to precisely separate and focus on high- and low-frequency tampering artifacts from the enhanced features. Concurrently, in the skip connections of the decoding path, the MGFA effectively establishes multi-level feature relationships between the tampered regions and the background. Finally, the entire network is supervised end-to-end by a custom Multi-Scale Edge Dice Loss (ME-Dice) to enhance the localization precision of tampering boundaries. Comprehensive experimental evaluations on several public benchmark datasets demonstrate that the performance of AMSEANet surpasses that of various existing state-of-the-art methods, and it exhibits superior robustness against common attacks such as noise and JPEG compression. Despite its excellent performance, the model still has limitations: its performance degrades when facing image scaling attacks; concurrently, on some datasets, the model achieves high Precision at the cost of a slightly lower Recall, revealing a potential trade-off between suppressing false positives and ensuring detection completeness.
Author Contributions
Conceptualization, Y.Y.; Data curation, W.L.; Formal analysis, Y.H.; Funding acquisition, Y.H.; Investigation, X.M.; Methodology, Y.Y.; Supervision, J.C. and H.W.; Visualization, Y.Y.; Writing—original draft, Y.Y.; Writing—review and editing, Y.H. and H.W. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Jiangxi Province 03 Special Project and 5G Program (grant number 20232ABC03A09), the Graduate Innovation Special Fund Project of East China University of Technology (grant number DHYC-2025070), and the Jiangxi Engineering Laboratory on Radioactive Geoscience and Big Data Technology Open Fund (grant number JELRGBDT202207). The APC was funded by the Jiangxi Province 03 Special Project and 5G Program.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data are available upon request to the correspondence e-mail.
Acknowledgments
The authors would like to express their gratitude to all colleagues who provided valuable suggestions and support during the research process.
Conflicts of Interest
The authors declare no conflicts of interest.
Abbreviations
| Abbreviation | Full Name |
| AMSEANet | Adaptive Multi-Scale Edge-Aware Network |
| ASF | Adaptive Smooth Filter |
| CSDRFBlock | Cross-Scale Dense Residual Fusion Block |
| ESFFM | Edge-Aware Spatial-Frequency Fusion Module |
| MGFA | Multi-Scale Global–local Fused Attention |
| ME-Dice | Multi-Scale Edge Dice Loss |
| CNN | Convolutional neural network |
| FCN | Fully Convolutional Network |
| SPN | Sensor pattern noise |
| DCT | Discrete Cosine Transform |
| DMWT | Discrete Multiwavelet Transform |
| SRM | Spatial Rich Model (Filter) |
| CA | Channel attention |
| SA | Spatial attention |
| GAP | Global average pooling |
| GMP | Global Max Pooling |
| TP | True Positives |
| FP | False Positives |
| FN | False Negatives |
| BCE | Binary Cross-Entropy (Loss) |
| QF | Quality Factor |
References
- Lukáš, J.; Fridrich, J.; Goljan, M. Digital camera identification from sensor pattern noise. IEEE Trans. Inf. Forensics Secur. 2006, 1, 205–214. [Google Scholar] [CrossRef]
- Tomioka, Y.; Ito, Y.; Kitazawa, H. Robust Digital Camera Identification Based on Pairwise Magnitude Relations of Clustered Sensor Pattern Noise. IEEE Trans. Inf. Forensics Secur. 2013, 8, 1986–1995. [Google Scholar] [CrossRef]
- Fang, Z.; Wang, S.; Zhang, X. Image Splicing Detection Using Color Edge Inconsistency. In Proceedings of the 2010 International Conference on Multimedia Information Networking and Security, Nanjing, China, 4–6 November 2010; pp. 923–926. [Google Scholar]
- Zhao, X.; Wang, S.; Li, S.; Li, J. Passive Image-Splicing Detection by a 2-D Noncausal Markov Model. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 185–199. [Google Scholar] [CrossRef]
- Dong, C.; Chen, X.; Hu, R.; Cao, J.; Li, X. MVSS-Net: Multi-View Multi-Scale Supervised Networks for Image Manipulation Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 3539–3553. [Google Scholar] [CrossRef]
- Anwar, M.A.; Tahir, S.F.; Fahad, L.G.; Kifayat, K. Image forgery detection by transforming local descriptors into deep-derived features. Appl. Soft Comput. 2023, 147, 110730. [Google Scholar] [CrossRef]
- Bi, X.; Zhang, Z.; Xiao, B. Reality Transform Adversarial Generators for Image Splicing Forgery Detection and Localization. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 14274–14283. [Google Scholar]
- Yu, Z.; Ni, J.; Lin, Y.; Deng, H.; Li, B. DiffForensics: Leveraging Diffusion Prior to Image Forgery Detection and Localization. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 12765–12774. [Google Scholar]
- Zhou, P.; Han, X.; Morariu, V.I.; Davis, L.S. Learning Rich Features for Image Manipulation Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1053–1061. [Google Scholar]
- Bappy, J.H.; Simons, C.; Nataraj, L.; Manjunath, B.S.; Roy-Chowdhury, A.K. Hybrid LSTM and Encoder-Decoder Architecture for Detection of Image Forgeries. IEEE Trans. Image Process. 2019, 28, 3286–3300. [Google Scholar] [CrossRef] [PubMed]
- Xiao, B.; Wei, Y.; Bi, X.; Li, W.; Ma, J. Image splicing forgery detection combining coarse to refined convolutional neural network and adaptive clustering. Inf. Sci. 2020, 511, 172–191. [Google Scholar] [CrossRef]
- Chen, H.; Chang, C.; Shi, Z.; Lyu, Y. Hybrid features and semantic reinforcement network for image forgery detection. Multimedia Syst. 2022, 28, 363–374. [Google Scholar] [CrossRef]
- Bi, X.; Wei, Y.; Xiao, B.; Li, W. RRU-Net: The Ringed Residual U-Net for Image Splicing Forgery Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 30–39. [Google Scholar]
- Wu, Y.; AbdAlmageed, W.; Natarajan, P. ManTra-Net: Manipulation Tracing Network for Detection and Localization of Image Forgeries With Anomalous Features. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 9535–9544. [Google Scholar]
- Liu, B.; Pun, C.-M. Locating splicing forgery by fully convolutional networks and conditional random field. Signal Process. Image Commun. 2018, 66, 103–112. [Google Scholar] [CrossRef]
- Salloum, R.; Ren, Y.; Kuo, C.-C.J. Image splicing localization using a multi-task fully convolutional network (MFCN). J. Vis. Commun. Image Represent. 2018, 51, 201–209. [Google Scholar] [CrossRef]
- Zhuang, P.; Li, H.; Tan, S.; Li, B.; Huang, J. Image tampering localization using a dense fully convolutional network. IEEE Trans. Inf. Forensics Secur. 2021, 16, 2986–2999. [Google Scholar] [CrossRef]
- Zhang, D.; Jiang, N.; Li, F.; Chen, J.; Liao, X.; Yang, G.; Ding, X. Multi-scale noise-guided progressive network for image splicing detection and localization. Expert Syst. Appl. 2024, 257, 124975. [Google Scholar] [CrossRef]
- Zhou, P.; Han, X.; Morariu, V.I.; Davis, L.S. Two-Stream Neural Network for Tampered Face Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 40–49. [Google Scholar]
- Wang, H.; Deng, J.; Lin, X.; Tang, W.; Wang, S. CDS-Net: Cooperative dual-stream network for image manipulation detection. Pattern Recognit. Lett. 2023, 176, 167–173. [Google Scholar] [CrossRef]
- Yang, Z.; Liu, B.; Bi, X.; Xiao, B.; Li, W.; Wang, G.; Gao, X. D-Net: A dual-encoder network for image splicing forgery detection and localization. Pattern Recognit. 2024, 155, 110727. [Google Scholar] [CrossRef]
- Popescu, A.C.; Farid, H. Exposing digital forgeries by detecting traces of resampling. IEEE Trans. Signal Process. 2005, 53, 758–767. [Google Scholar] [CrossRef]
- Zhang, K.; Li, R.; Yu, Y.; Luo, W.; Li, C. Deep Dense Multi-Scale Network for Snow Removal Using Semantic and Depth Priors. IEEE Trans. Image Process. 2021, 30, 7419–7431. [Google Scholar] [CrossRef] [PubMed]
- Wang, T.; Zhang, K.; Shao, Z.; Luo, W.; Stenger, B.; Lu, T.; Kim, T.-K.; Liu, W.; Li, H. GridFormer: Residual Dense Transformer with Grid Structure for Image Restoration in Adverse Weather Conditions. Int. J. Comput. Vision 2024, 132, 4541–4563. [Google Scholar] [CrossRef]
- Zhang, K.; Li, D.; Luo, W.; Ren, W.; Liu, W. Enhanced Spatio-Temporal Interaction Learning for Video Deraining: Faster and Better. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 1287–1293. [Google Scholar] [CrossRef] [PubMed]
- Jin, Z.; Qiu, Y.; Zhang, K.; Li, H.; Luo, W. MB-TaylorFormer V2: Improved Multi-Branch Linear Transformer Expanded by Taylor Formula for Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 5990–6005. [Google Scholar] [CrossRef]
- Atak, I.G.; Yasar, A. Image forgery detection by combining Visual Transformer with Variational Autoencoder Network. Appl. Soft Comput. 2024, 165, 112068. [Google Scholar] [CrossRef]
- Ren, R.; Hao, Q.; Gu, F.; Niu, S.; Zhang, J.; Wang, M. EMF-Net: An edge-guided multi-feature fusion network for text manipulation detection. Expert Syst. Appl. 2024, 249, 123548. [Google Scholar] [CrossRef]
- Hao, Q.; Ren, R.; Niu, S.; Wang, K.; Wang, M.; Zhang, J. UGEE-Net: Uncertainty-guided and edge-enhanced network for image splicing localization. Neural Netw. 2024, 178, 106430. [Google Scholar] [CrossRef] [PubMed]
- Gao, H.; Ma, B.; Zhang, Y.; Yang, J.; Yang, J.; Dang, D. Emphasizing crucial features for efficient image restoration. arXiv 2024, arXiv:2405.11468. [Google Scholar] [CrossRef]
- Fridrich, J.; Kodovsky, J. Rich Models for Steganalysis of Digital Images. IEEE Trans. Inf. Forensics Secur. 2012, 7, 868–882. [Google Scholar] [CrossRef]
- Bayar, B.; Stamm, M.C. Constrained Convolutional Neural Networks: A New Approach Towards General Purpose Image Manipulation Detection. IEEE Trans. Inf. Forensics Secur. 2018, 13, 2691–2706. [Google Scholar] [CrossRef]
- Milletari, F.; Navab, N.; Ahmadi, S.-A. V-Net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
- Dong, J.; Wang, W.; Tan, T. Casia image tampering detection evaluation database. In Proceedings of the 2013 IEEE China Summit and International Conference on Signal and Information Processing, Beijing, China, 6–10 July 2013; pp. 422–426. [Google Scholar]
- Guan, H.; Kozak, M.; Robertson, E.; Lee, Y.; Yates, A.N.; Delgado, A.; Zhou, D.; Kheyrkhah, T.; Smith, J.; Fiscus, J. MFC datasets: Large-scale benchmark datasets for media forensic challenge evaluation. In Proceedings of the 2019 IEEE Winter Applications of Computer Vision Workshops (WACVW), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 63–72. [Google Scholar]
- Ng, T.-T.; Hsu, J.; Chang, S.-F. Columbia Image Splicing Detection Evaluation Dataset; DVMM Lab, Columbia University: New York, NY, USA, 2009. [Google Scholar]
- Mahdian, B.; Saic, S. Using noise inconsistencies for blind image forensics. Image Vis. Comput. 2009, 27, 1497–1503. [Google Scholar] [CrossRef]
- Ferrara, P.; Bianchi, T.; De Rosa, A.; Piva, A. Image forgery localization via fine-grained analysis of CFA artifacts. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1566–1577. [Google Scholar] [CrossRef]
- Kwon, M.-J.; Yu, I.-J.; Nam, S.-H.; Lee, H.-K. CAT-Net: Compression artifact tracing network for detection and localization of image splicing. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 375–384. [Google Scholar]
- Liu, X.; Liu, Y.; Chen, J.; Liu, X. PSCC-Net: Progressive spatio-channel correlation network for image manipulation detection and localization. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 7505–7517. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).