Next Article in Journal
Deep Learning-Based Sound Source Localization: A Review
Previous Article in Journal
Application of Fermented Wheat, Acorns, and Sorghum in Processing of Couscous: Effect on Culinary Quality, Pasting Properties, and Microstructure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detecting AI-Generated Images Using a Hybrid ResNet-SE Attention Model

by
Abhilash Reddy Gunukula
1,†,
Himel Das Gupta
2,† and
Victor S. Sheng
1,*
1
Department of Computer Science, Texas Tech University, Lubbock, TX 79409, USA
2
Department of Mathematics and Computer Science, Louisiana State University of Alexandria, Alexandria, LA 71302, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(13), 7421; https://doi.org/10.3390/app15137421
Submission received: 16 May 2025 / Revised: 14 June 2025 / Accepted: 30 June 2025 / Published: 2 July 2025
(This article belongs to the Special Issue Advanced Signal and Image Processing for Applied Engineering)

Abstract

The rapid advancements in generative artificial intelligence (AI), particularly through models like Generative Adversarial Networks (GANs) and diffusion-based architectures, have made it increasingly difficult to distinguish between real and synthetically generated images. While these technologies offer benefits in creative domains, they also pose serious risks in terms of misinformation, digital forgery, and identity manipulation. This paper presents a novel hybrid deep learning model for detecting AI-generated images by integrating the ResNet-50 architecture with Squeeze-and-Excitation (SE) attention blocks. The proposed SE-ResNet50 model enhances channel-wise feature recalibration and interpretability by integrating Squeeze-and-Excitation (SE) blocks into the ResNet-50 backbone, enabling dynamic emphasis on subtle generative artifacts such as unnatural textures and semantic inconsistencies, thereby improving classification fidelity. Experimental evaluation on the CIFAKE dataset demonstrates the model’s effectiveness, achieving a test accuracy of 96.12%, precision of 97.04%, recall of 88.94%, F1-score of 92.82%, and an AUC score of 0.9862. The model shows strong generalization, minimal overfitting, and superior performance compared with transformer-based models and standard architectures like ResNet-50, VGGNet, and DenseNet. These results confirm the hybrid model’s suitability for real-time and resource-constrained applications in media forensics, content authentication, and ethical AI governance.

1. Introduction

In recent years, significant advancements in artificial intelligence (AI) have enabled the creation of highly realistic digital media, including images, videos, and audio. Generative models such as Generative Adversarial Networks (GANs) and diffusion models have become increasingly capable of producing AI-generated images indistinguishable from real images [1,2,3]. While these developments provide opportunities in art, entertainment, and virtual environments [1], they also pose severe threats related to misinformation, identity manipulation, and social trust erosion [3,4]. The potential misuse of synthetic media has already been observed in numerous cases, including the dissemination of fake news, fraudulent identity creation, and even manipulating political discourse and public opinion [3,4].
The widespread availability and ease of use of generative AI technologies have raised concerns among researchers and policymakers alike [5]. For instance, recent controversies arose when an AI-generated image won an art prize. Such an unexpected outcome sparked debates regarding authenticity and creative ownership. To summarize, it illustrates broader societal anxieties about the implications of AI-generated content [1]. Therefore, it has become critically important to develop robust methods capable of accurately identifying synthetic images to preserve trust in digital content.
Existing approaches for the detection of AI-generated images predominantly rely on traditional deep learning architectures such as ResNet, VGGNet, and DenseNet [6,7,8]. These models have demonstrated considerable success across multiple image classification tasks; however, they still encounter difficulties when dealing with subtle, realistic AI-generated images and can be vulnerable to adversarial perturbations [9]. Moreover, standard deep learning methods often lack interpretability and require extensive computational resources, making them challenging to deploy effectively in real-time or resource-constrained environments [10,11,12].
Considering these limitations, this research aims to develop an innovative hybrid deep learning framework combining the ResNet architecture with an attention mechanism. The attention mechanism is designed to dynamically identify and focus on regions of the image most indicative of AI-generated artifacts, enhancing the model’s accuracy and interpretability [10]. This proposed approach seeks to mitigate the shortcomings associated with traditional deep learning models, enabling more precise and computationally efficient identification of synthetic images.
The main objectives of this research are as follows:
  • To evaluate the effectiveness of widely used deep learning models (ResNet, VGGNet, DenseNet) in accurately detecting AI-generated images.
  • To propose and validate a novel hybrid model integrating ResNet with an attention-based mechanism to achieve superior accuracy and robustness compared with baseline models.
  • To rigorously assess the performance of the proposed model using the CIFAKE dataset [6], evaluating its potential for real-world applications.
The core contribution of this research lies in augmenting the ResNet-50 architecture with Squeeze-and-Excitation (SE) attention blocks to create a hybrid model—SE-ResNet50—designed specifically for AI-generated image detection. These SE blocks dynamically recalibrate feature channels during learning, enabling the network to selectively emphasize subtle generative cues such as unnatural textures, semantic inconsistencies, or repetitive patterns commonly associated with diffusion-based and GAN-generated content. This channel-wise attention mechanism significantly improves the model’s ability to distinguish synthetic imagery, resulting in enhanced precision and reduced classification ambiguity, especially in borderline cases.
This paper contributes significantly to the domain by proposing an efficient, interpretable, and robust method for synthetic image identification, addressing critical issues raised by the rapidly evolving landscape of generative AI. Furthermore, the presented research lays the groundwork for future developments in AI-generated media detection and helps build a resilient framework against malicious use of AI-generated content.
The remainder of this paper is structured as follows: Section 2 presents a comprehensive literature review, Section 3 describes the methodology and experimental setup, Section 4 details the experimental results and discusses the findings, and Section 5 concludes the paper and suggests avenues for future research.

2. Literature Review

This section provides a comprehensive review of existing research related to AI-generated image detection. The purpose is to systematically analyze prominent methodologies, evaluate key datasets, and identify critical gaps that motivate this study.

2.1. Existing Techniques and Methodologies

Various deep learning methodologies have been employed for the identification of AI-generated images. Among these, Convolutional Neural Networks (CNNs) and their advanced architectures have emerged as dominant approaches due to their remarkable capability in pattern recognition and classification tasks.
ResNet architectures, introduced by He et al. [13], leverage residual connections to alleviate the vanishing gradient problem inherent in deep neural networks. These connections facilitate the training of deeper architectures and enable more sophisticated feature extraction, significantly enhancing classification performance. However, while ResNet effectively captures deep hierarchical features, it may struggle to detect subtle differences between real and synthetic images [2,14].
VGGNet, introduced by Simonyan and Zisserman [15], is widely recognized for its structural simplicity, employing homogeneous convolutional layers of small receptive fields (3 × 3). VGGNet demonstrates strong generalization capabilities across diverse image datasets, achieving notable performance in general classification tasks. Nevertheless, it requires substantial computational resources due to the dense nature of its architecture [15].
Another important architecture, DenseNet, introduced by Huang et al. [16], utilizes densely connected convolutional layers, significantly enhancing feature reuse and gradient flow throughout the network. This characteristic enables DenseNet to identify subtle visual features that differentiate AI-generated images from authentic ones. Despite achieving high accuracy, DenseNet’s computational cost can become prohibitive, limiting its deployment in resource-constrained environments [16].
The integration of attention mechanisms in deep learning has emerged as a promising approach to further enhance classification accuracy and interpretability. Vaswani et al. [17] introduced the concept of attention to dynamically weigh features, significantly improving model performance, especially in scenarios requiring the identification of subtle discriminative patterns. Attention mechanisms have successfully been applied in deepfake detection [18,19] and other related fields, indicating their suitability for AI-generated image recognition tasks.

2.2. Datasets and Benchmarks

Standard datasets play a critical role in assessing and benchmarking AI-generated image detection techniques. Among these datasets, CIFAKE [20] has been widely used for evaluating synthetic image identification models. It contains a curated mix of authentic and AI-generated images, specifically tailored for evaluating detection capabilities in distinguishing subtle differences.
Other relevant benchmarks include Celeb-DF and DeepFake Detection (DFD) datasets [4]. These datasets primarily focus on facial manipulation and synthetic faces, enabling robust evaluation of algorithms’ effectiveness against high-quality Generative Adversarial Networks (GANs) and diffusion-based methods [2]. Utilization of these datasets has substantially contributed to algorithmic improvements, providing well-established performance baselines in the literature.

2.3. Gaps and Limitations in Existing Approaches

Despite significant progress, current methods exhibit several critical limitations. Many deep learning models, including ResNet and VGGNet, exhibit vulnerability to adversarial perturbations, which may compromise their reliability [21,22,23,24]. Moreover, these models often encounter difficulties generalizing learned features to novel, previously unseen data or adversarially modified inputs [25].
Additionally, computational complexity remains a significant barrier to real-time deployment and resource-limited applications [3,26]. DenseNet, although demonstrating high accuracy, demands extensive computational resources, complicating its practical deployment. Recent journal studies have applied efficient Squeeze-and-Excitation networks to domains like medical imaging and remote sensing, demonstrating their potential for lightweight, high-performance detection [27,28]. Moreover, existing models lack interpretability, posing difficulties in understanding their internal decision-making processes, crucial for sensitive applications such as forensic investigations [21,26].
These limitations necessitate the development of innovative models capable of accurately detecting synthetic images while addressing computational constraints and interpretability concerns effectively.

2.4. Summary and Motivation for Proposed Research

The review of existing literature demonstrates the need for models capable of overcoming current limitations in accuracy, interpretability, robustness, and efficiency. Existing models, although effective, fall short in consistently identifying realistic synthetic images and remain computationally intensive, restricting their practical deployment.
Motivated by these gaps, this study proposes a hybrid approach, combining ResNet’s deep feature extraction capability with an attention mechanism to dynamically emphasize image regions critical to accurate synthetic image identification. This approach aims to address the gaps, significantly improving the detection accuracy, robustness against adversarial perturbations, computational efficiency, and interpretability.

3. Methodology

This research adopts a systematic experimental approach to identify AI-produced images. The study first assesses the baseline deep learning models that include ResNet, VGGNet, and DenseNet, and then introduces a new hybrid model that incorporates the ResNet architecture with an attention mechanism. The CIFAKE dataset is used for experimentation, as it contains a balanced mix of real and fake images to enable detailed analysis [20].

3.1. Dataset Description

The dataset employed in this research is the CIFAKE dataset [20], consisting of 50,000 labeled images, evenly distributed between real and AI-generated categories. CIFAKE is specifically curated to assess the performance of AI-generated image detection models, incorporating subtle synthetic image features generated using advanced generative models.
For experimental analysis, the dataset is partitioned into two subsets: 80% (40,000 images) for training and validation, and 20% (10,000 images) reserved strictly for testing. Each image is resized to 224 × 224 pixels to comply with the input constraints of standard CNN architectures such as ResNet and VGGNet. Representative samples of real and AI-generated images from the CIFAKE dataset are illustrated in Figure 1.

3.2. Model Architectures and Techniques

Baseline Models

The research first evaluates three prominent CNN architectures as baselines:
  • ResNet employs residual learning with skip connections to enhance gradient flow in deep architectures, mitigating the vanishing gradient problem and effectively capturing hierarchical features [15].
  • VGGNet consists of uniform convolutional layers with small receptive fields ( 3 × 3 ), enabling efficient feature extraction with strong generalization capabilities, despite a relatively higher computational load [16].
  • DenseNet utilizes densely interconnected layers, promoting effective feature reuse and robust gradient propagation, thus achieving high accuracy in distinguishing minute differences between synthetic and real images [29].
These baseline models provide a comparative foundation for evaluating the performance of the proposed hybrid model.

3.3. Experimental Setup

Experiments were conducted using a cloud-based infrastructure on Google Colab, leveraging Tesla GPUs for accelerated training. Tensorflow 2.18.0 was the primary framework used for model development, training, and evaluation, owing to its flexibility and strong community support. TensorFlow was additionally used for validation and visualization tasks.
The training pipeline involved standardized hyperparameters to ensure fair comparison across models mentioned in Table 1. Key configurations include the following:
  • Optimizer: Adam optimizer, selected for its adaptive learning rate and fast convergence properties [6].
  • Learning Rate: 1 × 10 4 for all models, including the hybrid.
  • Batch Size: 32 images per batch, balanced for efficient GPU utilization and memory stability.
  • Loss Function: Cross-Entropy Loss, appropriate for binary classification tasks.
The Adam optimizer [30] was consistently applied for all models, and training was performed under the same computational setup for reproducibility. Notably, the SE-ResNet50 hybrid model—despite being trained for only 10 epochs—demonstrated faster convergence and stronger generalization compared with its baseline counterparts, due in part to the effectiveness of transfer learning and the attention-enhanced SE blocks.
These choices were based on iterative tuning and ensured optimal convergence while maintaining training stability and consistency across architectures.

3.4. Proposed Hybrid Model (ResNet + Attention Mechanism)

To address limitations observed in baseline models—particularly their inability to effectively differentiate subtle visual patterns between real and AI-generated images—this research introduces a novel hybrid deep learning model. This hybrid architecture leverages the established feature-extraction strengths of the ResNet-50 architecture [20] in conjunction with a powerful attention mechanism known as the Squeeze-and-Excitation (SE) block [15].

3.5. Model Architecture

The core of the proposed architecture is based on ResNet-50, a variant of Residual Network (ResNet) consisting of multiple residual blocks interconnected by shortcut (skip) connections [20]. Each residual block contains convolutional layers followed by batch normalization and ReLU activation functions. Skip connections enable efficient gradient flow, allowing deeper networks to learn complex hierarchical image representations while addressing the vanishing gradient problem prevalent in traditional CNN architectures [16].
To further enhance the discriminative capability of ResNet-50 in identifying subtle features inherent to synthetic images, an SE attention mechanism is integrated into each residual block. This integration produces a residual block with channel-wise attention, known as the SE-ResNet block.

3.6. Squeeze-and-Excitation (SE) Attention Mechanism

The SE attention mechanism explicitly recalibrates the importance of each feature channel dynamically, thus enabling the network to enhance focus on the most informative features while suppressing less informative ones [15]. The detailed implementation of the SE mechanism is presented in Algorithm 1. The SE block achieves channel-wise attention through two critical operations:
Algorithm 1 Forward Pass of SELayer.
1:
Input: Feature map X R C × H × W
2:
Output: Attention-weighted feature map X ˜
3:
z c 1 H · W i = 1 H j = 1 W X c ( i , j )                                              ▷ Global Average Pooling (GAP)
4:
z [ z 1 , z 2 , . . . , z C ]                                                                      ▷ Channel descriptor vector
5:
s σ W 2 · δ ( W 1 · z )                          ▷ Two FC layers with ReLU and sigmoid activation
6:
X ˜ c s c · X c for all c = 1 , 2 , , C                                           ▷ Element-wise channel scaling
7:
return  X ˜
  • Legend:
  • X: Input feature map of shape C × H × W .
  • C: Number of channels.
  • H , W : Height and width of the feature map.
  • z c : Average-pooled scalar for channel c.
  • z: Channel descriptor vector ( R C × 1 ).
  • W 1 , W 2 : Fully connected (FC) layer weights.
  • δ : ReLU activation function.
  • σ : Sigmoid activation function.
  • s c : Channel-wise attention weight.
  • X ˜ c : Attention-weighted feature map for channel c.
  • X ˜ : Output feature map after SE attention.

3.6.1. Squeeze Operation

This step aggregates spatial information across each feature channel by performing Global Average Pooling (GAP). GAP generates channel-wise statistics, compressing each feature map from spatial dimensions H × W to a single scalar value per channel. Mathematically, the squeeze operation is described as follows:
z c = F s q ( u c ) = 1 H × W i = 1 H j = 1 W u c ( i , j )
where z c represents the squeezed scalar of the c t h channel, and u c ( i , j ) represents the feature values at spatial coordinates ( i , j ) within channel c.

3.6.2. Excitation Operation

Following the squeeze operation, the excitation step recalibrates channel importance using two fully connected (FC) layers, separated by a ReLU activation function, followed by a sigmoid activation. This enables the model to learn nonlinear interactions between channels:
s = F e x ( z , W ) = σ ( W 2 · δ ( W 1 · z ) )
where W 1 R C r × C and W 2 R C × C r are weight matrices of two FC layers, r is a reduction ratio (commonly set to 16), δ denotes the ReLU activation function, and σ represents the sigmoid activation function. The output s R C indicates channel-wise attention scores.
The channel-wise attention score vector s subsequently recalibrates the original input feature maps u c :
u ˜ c = s c · u c
where u ˜ c is the recalibrated channel feature map.

3.7. Hybrid Architecture: SE-ResNet50

Integrating SE blocks into each residual block of ResNet-50 results in the proposed SE-ResNet50 model. Specifically, each standard residual block in ResNet-50 is modified by adding an SE attention module at the end of each block. This modification enriches residual blocks with adaptive, channel-wise attention capability while preserving the robust feature extraction characteristics of the original residual connections. The complete architecture initialization and forward pass procedure is detailed in Algorithm 2 and Figure 2.
Algorithm 2 AttentionResNet: Initialization and Forward Pass.
1:
  In:  I R 3 × 224 × 224
2:
  Out:  y R C
3:
  procedure InitModel
4:
       R Pretrain ( R 50 )
5:
      for  B i { L 1 , L 2 , L 3 , L 4 }  do
6:
          for  L j B i  do
7:
              L j L j + SE
8:
          end for
9:
      end for
10:
     W f c R d × C , b f c R C
11:
end procedure
12:
procedure FwdPass(I)
13:
     F CB ( I )                                                 ▷  CB = Conv + MP
14:
    for  B i { L 1 , L 2 , L 3 , L 4 }  do
15:
         F B i ( F )
16:
    end for
17:
     F GAP ( F )
18:
     z W f c · F + b f c
19:
     y SM ( z )
20:
    return y
21:
end procedure
Legend:
  • I: Input image tensor.
  • y: Output class probabilities.
  • R 50 : ResNet-50.
  • L i : Residual block layers ( i = 1 to 4).
  • L j : Convolutional layer within block.
  • SE : Squeeze-and-Excitation Layer
  • CB : Convolution + Max Pooling.
  • GAP : Global Average Pooling.
  • SM : Softmax function.
  • W f c , b f c : Weights and bias of final fully connected layer.
  • d: Feature vector dimension after GAP .
  • C: Number of output classes.
The overall hybrid architecture comprises the following:
  • Initial convolutional layer:  7 × 7 convolution, stride = 2; followed by batch normalization and ReLU activation; max pooling with 3 × 3 kernel, stride = 2.
  • Stacked SE-ResNet blocks:
    3 SE-residual blocks (64 channels)
    4 SE-residual blocks (128 channels)
    6 SE-residual blocks (256 channels)
    3 SE-residual blocks (512 channels)
  • Classification head: Global Average Pooling (GAP) followed by a fully connected (FC) layer for binary classification (real vs. AI-generated).
The final fully connected layer outputs probability scores using a softmax activation function:
y pred = softmax ( W f c · x G A P + b f c )
where W f c and b f c represent the weights and bias of the final classification layer, respectively, and x G A P is the output from the global average pooling layer.
The methodology section presented a structured approach to effectively identify AI-generated images. The CIFAKE dataset was employed, providing a robust benchmark for evaluation. Baseline models (ResNet, VGGNet, and DenseNet) were initially evaluated to establish performance baselines. Subsequently, a novel hybrid model, SE-ResNet50, integrating the ResNet-50 architecture with the Squeeze-and-Excitation (SE) channel-wise attention mechanism, was proposed to enhance the detection of subtle synthetic image features.
This explanation included the architectural integration of SE-blocks within residual blocks, clearly articulated equations and algorithmic pseudocode, and highlighted the advantages of the hybrid approach concerning computational complexity, interpretability, and improved accuracy. Experimental configurations, including software tools, computational environment, and hyperparameters, were explicitly detailed. Comprehensive evaluation metrics were also defined to rigorously quantify and compare the performance of the proposed model with baseline models.

4. Results and Discussion

This section presents detailed experimental results obtained from evaluating both baseline models (ResNet-50, VGGNet, and DenseNet) and the proposed hybrid model (SE-ResNet50) on the CIFAKE dataset. Results are analyzed and compared across critical performance metrics, including accuracy, precision, recall, F1-score, ROC-AUC scores, confusion matrices, and training efficiency.

4.1. Performance Evaluation of Baseline Models

To establish a strong comparative foundation for evaluating the proposed hybrid architecture, three well-established deep learning models—ResNet-50, VGGNet, and DenseNet—were re-implemented and evaluated using the CIFAKE dataset. This dataset, comprising a balanced set of real and AI-generated images, served as a reliable benchmark to assess each model’s capability in accurately distinguishing between synthetic and authentic visual content.
Each model was trained under identical experimental conditions using 80% of the dataset for training and 20% for testing. This uniform configuration ensured a fair and consistent evaluation across all architectures.
The SE-ResNet50 model emerged as the best-performing architecture among all evaluated models, achieving a test accuracy of 96.12%, precision of 97.04%, recall of 88.94%, F1-score of 92.82%, and a ROC-AUC score of 0.9862. This superior performance underscores the impact of integrating Squeeze-and-Excitation (SE) blocks into the ResNet-50 backbone. By recalibrating feature channels dynamically, the SE mechanism allows the network to focus more effectively on subtle and discriminative cues typically introduced by generative models, such as minor texture inconsistencies or unnatural patterns, thereby enhancing the model’s ability to distinguish between real and synthetic images with greater precision.
VGGNet, although architecturally simpler with its fixed-size ( 3 × 3 ) convolutional layers, achieved a test accuracy of 94.06%. Its deep and sequential design contributes to strong generalization but comes at the cost of higher computational demand and slower training. DenseNet, with its densely connected layers that promote efficient gradient flow and feature reuse, slightly outperformed VGGNet with a test accuracy of 95.98%. However, it lagged behind SE-ResNet50 in precision, recall, and F1-score, highlighting its relative weakness in maintaining classification balance between real and fake image categories.
Overall, while both VGGNet and DenseNet performed competitively, SE-ResNet50 clearly outperformed them across most key metrics, demonstrating not only high classification accuracy but also strong generalization, low false-positive rates, and improved focus on semantically relevant features. These characteristics make it especially suitable for high-stakes applications such as digital content authentication, where both performance and interpretability are critical.
Table 2 summarizes the performance metrics of baseline models evaluated on the CIFAKE dataset.
Results indicate that the ResNet-50 model, enhanced with Squeeze-and-Excitation (SE) attention blocks, achieved the best performance among all the evaluated models. With an accuracy of 96.12%, precision of 97.04%, recall of 88.94%, and an F1-score of 92.82%, SE-ResNet50 outperformed both DenseNet and VGGNet across all major classification metrics. Its high ROC-AUC score of 0.9862 further confirms its strong discriminative capability between real and AI-generated images.
This superior performance can be attributed to the SE blocks, which dynamically recalibrate the importance of feature channels. By emphasizing informative features and suppressing less useful ones, the model becomes more sensitive to subtle visual discrepancies, such as texture artifacts or unnatural transitions, commonly found in synthetic images.
VGGNet, while architecturally simpler and effective in generating deep representations through stacked 3 × 3 convolutional layers, achieved a lower accuracy of 94.06%. Despite its good precision, the model’s large parameter count and lack of efficient connectivity resulted in slower training and higher memory usage. DenseNet showed marginally better results than VGGNet with an accuracy of 95.98%, benefiting from its densely connected layers that encourage feature reuse. However, it still lagged behind SE-ResNet50 in precision, F1-score, and computational efficiency.
To further analyze classification behavior, a confusion matrix for the ResNet-50 model is presented in Table 3. It provides insights into the distribution of true positives, false positives, true negatives, and false negatives.
The model correctly classified 9729 real images and 8894 synthetic images, with relatively low rates of misclassification—271 false positives and 1106 false negatives. These results demonstrate the model’s balanced classification capability and reliability in distinguishing between visually similar image classes. Compared with VGGNet and DenseNet, SE-ResNet-50 shows clear improvements not only in numerical performance but also in reducing both types of classification errors.
Therefore, the proposed SE-ResNet-50 hybrid model sets a new benchmark for detecting AI-generated images in this context, combining strong predictive accuracy with efficient computation and enhanced feature sensitivity.

4.2. Comparison with Prior Works

To highlight the contribution of the proposed SE-ResNet-50 model, a comparative evaluation with existing state-of-the-art models on the CIFAKE dataset is presented in Table 4. This comparison includes standard architectures such as ResNet-50, VGGNet-16, and DenseNet-121, evaluated under identical experimental settings. Additionally, we include results from a transformer-based Vision Transformer (ViT) model as reported by Kumar [2] for a comprehensive comparison with newer architectures.
As observed, the proposed SE-ResNet50 model significantly outperforms both the baseline ResNet-50 and the transformer-based ViT architecture (by 4.12% in accuracy), while also showing marginal improvements over other conventional CNN architectures. Although AUC scores were not available for the ViT model, our SE-ResNet50 achieved the highest AUC (0.9862) among all compared models with available metrics. The improvement over ResNet-50 is attributed to the integration of Squeeze-and-Excitation (SE) attention mechanisms, which enable the model to focus more effectively on subtle generative artifacts. The model demonstrates strong precision–recall balance and robust separability across classes, establishing it as a superior architecture for AI-generated image detection.

4.3. Statistical Significance Testing

To assess whether the performance improvement of SE-ResNet50 over DenseNet is statistically significant, McNemar’s test was conducted on their respective predictions across the CIFAKE test set. A 2×2 contingency table (Table 5) was constructed based on classification disagreements:
McNemar’s chi-square statistic was calculated as
χ 2 = ( | b c | 1 ) 2 b + c = ( | 918 189 | 1 ) 2 918 + 189 478.6
This yields a p-value < 0.0001, indicating that the improvement of SE-ResNet50 over DenseNet is statistically significant. In future work, we plan to complement this with confidence intervals and multiple runs to further assess the robustness of model performance.

4.4. Performance of Proposed Hybrid Model (SE-ResNet-50)

To overcome the limitations identified in the baseline models, particularly in handling nuanced visual cues and computational trade-offs, a hybrid model—SE-ResNet50—has been proposed and rigorously evaluated on the same CIFAKE dataset. This architecture augments the traditional ResNet-50 framework with Squeeze-and-Excitation (SE) blocks, a channel-wise attention mechanism designed to dynamically recalibrate the importance of feature maps. The goal was to enable the network to focus more selectively on image regions that contain subtle generative artifacts often missed by traditional convolutional networks.
The SE-ResNet-50 model demonstrated the best classification performance among all tested models, achieving an accuracy of 96.12%, a precision of 97.04%, a recall of 88.94%, and an F1-score of 92.82%. Additionally, the model attained a high ROC-AUC score of 0.9862, indicating excellent discrimination capability between real and synthetic images across classification thresholds. These results confirm SE-ResNet50 as the most balanced and effective model for the detection task.
Table 6 presents a detailed comparison of core evaluation metrics.
The superior performance of the SE-ResNet50 model—achieving 96.12% accuracy and a ROC-AUC score of 0.9862—is primarily attributed to the incorporation of SE blocks into the ResNet-50 architecture. These blocks perform dynamic, channel-wise recalibration of feature maps by leveraging global context through the squeeze operation and reweighting informative features via excitation. As a result, the model is able to highlight fine-grained patterns and subtle visual anomalies—such as edge inconsistencies or low-level artifacts—that are often overlooked by conventional convolutional layers. This architectural enhancement improves class separability and reduces misclassification, particularly in challenging or visually ambiguous cases, contributing directly to the model’s high classification fidelity.
The precision score of 97.04% reflects the model’s ability to reduce false positives, while the recall score of 88.94% demonstrates its effectiveness in detecting AI-generated images. The F1-score of 92.82% highlights its overall robustness in balancing sensitivity and specificity. With the highest test performance across all evaluated models, SE-ResNet-50 emerges as the optimal choice for real-world applications requiring both accuracy and efficiency.

4.4.1. Impact of SE Attention Mechanism

The significant performance boost can be directly attributed to the integration of SE blocks, which apply channel-wise attention to modulate feature responses based on their relative importance. Unlike traditional convolutional layers that treat all feature channels equally, SE blocks allow the network to attend to channels that encode critical visual cues—such as edge inconsistencies, textural anomalies, and spatial distortions—often introduced by generative models like GANs or diffusion networks [2,3].
Through global average pooling, each SE block captures a compact summary of the global context of each feature map, followed by excitation through a pair of fully connected layers. This mechanism enables the model to suppress uninformative or noisy channels while amplifying those that carry meaningful signals indicative of synthetic content.

4.4.2. Comparative Analysis and Discussion

To visualize comparative performance, ROC curves and confusion matrices for the baseline models and the proposed SE-ResNet50 are analyzed.
Figure 3 illustrates the comparative ROC curves of the hybrid and baseline models. The SE-ResNet50 model achieves an area under the curve (AUC) of 0.98, demonstrating superior classification performance compared with VGGNet and DenseNet. Its ROC curve remains consistently above those of the baselines, particularly in the low false-positive rate region, reaffirming its heightened sensitivity and specificity in detecting AI-generated images.
To provide a comparative diagnostic analysis of classification behavior, Table 7 presents a side-by-side confusion matrix comparison between ResNet-50 and the proposed SE-ResNet50 model. The SE-ResNet50 model exhibits a notable reduction in both false positives and false negatives, further evidencing its enhanced discriminatory capability and robustness in practical deployment scenarios.
Furthermore, Figure 4 presents the graphical representation of confusion matrices for both ResNet-50 and SE-ResNet50. The hybrid model demonstrates significantly fewer misclassifications, especially in detecting synthetic (AI-generated) images. This improvement is particularly vital for high-stakes applications in digital forensics, media integrity verification, and identity protection.
The remarkable improvement observed in SE-ResNet50 can be attributed to the integration of Squeeze-and-Excitation (SE) blocks. These attention mechanisms dynamically recalibrate feature maps across channels, enabling the network to better capture subtle, class-discriminative features commonly embedded in AI-generated imagery.

4.4.3. Training Stability and Computational Efficiency

The model also demonstrates training stability, achieving convergence within 20 epochs using a batch size of 32 and an initial learning rate of 1 × 10 4 . The integration of SE blocks introduces only a modest increase in computational overhead compared with plain ResNet-50, while being significantly faster than DenseNet. As detailed in Table 8, SE-ResNet-50 provides an efficient trade-off between accuracy and training time.
SE-ResNet-50 maintains efficient computational performance (318 s/epoch), significantly faster than DenseNet (646 s/epoch), despite similar accuracy levels. While slightly slower than basic ResNet-50, the substantial accuracy improvements justify the moderate increase in computational demand. This efficiency makes the proposed model more suitable for deployment in real-world applications, including mobile edge devices and cloud-based platforms, where computational resources may be constrained.
To quantify its efficiency, the model’s training time and GPU utilization were recorded and are shown in Table 9. The average training time per epoch was 318 s, and GPU utilization was moderate at 48 units, reflecting a good balance between performance and resource consumption. Additionally, the model maintained accuracy stability with a standard deviation of ±0.25%, indicating robust and consistent learning behavior across training runs.
Figure 5 illustrates the training dynamics of the SE-ResNet-50 model. The training curves show consistent improvement across epochs with minimal overfitting, though a slight divergence is observed in recall and validation loss after epoch 6, possibly due to increased sensitivity to subtle generative features. The stability in training loss and precision confirms robust convergence. This data supports the feasibility of deploying the model in both cloud-based and resource-constrained edge environments. Compared with DenseNet—which, while slightly more accurate, incurs nearly double the training cost—the SE-ResNet-50 model achieves an optimal balance of speed, accuracy, and hardware efficiency.

4.5. Qualitative Analysis of Classification Behavior

False Positives (FP): real images with compression noise or poor lighting were occasionally misclassified as fake, likely due to superficial texture anomalies.
False Negatives (FN): some AI-generated images with high photorealism and clean edge consistency evaded detection, leading to misclassification.
True Positives (TP): synthetic images with repetitive patterns, artifacts, or blurred textures were consistently flagged as fake.
True Negatives (TN): natural scenes with complex lighting and realistic depth cues were correctly classified as authentic.
This qualitative analysis supports the numerical metrics and reveals edge cases where generative quality or real-world noise affects prediction reliability.

4.5.1. Interpretability and Transparency

Beyond numerical performance, the model also enhances interpretability, a critical factor in deep learning-based security applications. The learned attention weights from SE blocks provide insight into the model’s decision-making, allowing practitioners to visualize and understand which parts of the image contributed most to its classification. This capability not only supports debugging and refinement but also enhances trustworthiness, especially in forensic and legal scenarios.

4.5.2. Discussion of Implications and Limitations

The experimental results obtained from the proposed SE-ResNet-50 model not only highlight its superior performance but also underline several broader implications and practical considerations for the real-world deployment of AI-generated image detection systems. Future investigations could address these limitations through enhanced optimization strategies, more robust data augmentation methods, or by exploring other lightweight attention mechanisms.
Real-World Applicability and Societal Impact
The increasing prevalence of AI-generated images in digital art, journalism, advertising, and social media introduces significant ethical and operational challenges, including misinformation, identity fraud, and digital forgery [1,3]. The high accuracy (96.12%) and precision (97.04%) of the proposed hybrid model suggest it can serve as a reliable automated tool for detecting such synthetic content across various industries, particularly in content moderation, digital forensics, law enforcement, and media authentication.
Importantly, the model demonstrates not just statistical excellence but operational robustness. Its ability to operate with moderate computational demands while retaining state-of-the-art accuracy positions it well for real-time or embedded applications, such as browser plugins, mobile applications, or edge AI devices. Moreover, its enhanced interpretability via attention weights allows stakeholders to better trust, audit, and explain decisions made by the system—an essential requirement in regulated environments like legal and governmental contexts.
Ethical Considerations and Explainability
As the arms race between generative AI and detection systems accelerates, explainability becomes paramount. The integration of SE blocks introduces a level of transparency uncommon in traditional CNNs. By providing insight into which feature channels the model emphasizes, developers and analysts gain the ability to interpret, debug, and validate model predictions—bridging the gap between black-box AI and human interpretability [15].
This attention-guided insight can also support efforts toward ethical AI governance, where transparency and accountability in decision-making processes are increasingly required by both institutions and emerging legislation worldwide.
Limitations and Areas for Caution
Despite its high performance, the SE-ResNet-50 model presents certain limitations that warrant further research and engineering optimization:
  • Computational overhead: Although more efficient than DenseNet, the hybrid model still introduces additional parameters due to the SE blocks. This could be a constraint for ultra-low-power devices or latency-sensitive real-time systems.
  • Data sensitivity: The model’s performance is closely tied to the quality and diversity of training data. Biases or limitations within the CIFAKE dataset could influence its generalization ability across unseen generative models or content types.
  • Overfitting to visual artifacts: There is a risk that the model may be overfit to the generative patterns present in the current dataset (e.g., GAN fingerprints or compression artifacts), reducing its adaptability to emerging or more advanced synthetic generation techniques that minimize such artifacts [29,31].
  • Adversarial vulnerability: Like most deep learning models, SE-ResNet-50 may remain vulnerable to adversarial attacks, where carefully crafted perturbations can manipulate outputs. Adversarial robustness was not the focus of this study and remains a critical direction for future investigation [17].
  • Limited architecture comparison: While our study compares with traditional CNN architectures, a comprehensive comparison with newer transformer-based models was not fully explored. Although our current results show competitive performance against selected transformer architectures, a more extensive evaluation across the full spectrum of vision transformers and diffusion-aware models would provide additional insights into relative strengths and weaknesses.

5. Conclusions and Future Work

5.1. Conclusions

The rapid advancement of generative artificial intelligence has introduced new challenges in discerning synthetic visual content from authentic media. In this study, a hybrid deep learning model—SE-ResNet-50—was proposed to address these challenges by combining the hierarchical feature extraction capabilities of ResNet-50 with the dynamic feature recalibration of Squeeze-and-Excitation (SE) attention mechanisms. Extensive experiments conducted on the CIFAKE dataset demonstrated that the proposed hybrid model outperforms standard CNN-based architectures such as VGGNet and DenseNet and even improves upon the baseline ResNet-50 model. Comparative analysis with transformer-based architectures shows our approach achieves competitive performance while maintaining computational efficiency. The SE-ResNet-50 model achieved a classification accuracy of 96.12%, a precision of 97.04%, a recall of 88.94%, an F1-score of 92.82%, and a high ROC-AUC score of 0.9862—confirming its robust ability to distinguish real from AI-generated images. Additionally, the inclusion of the SE attention mechanism enhanced both the discriminative power and interpretability of the model. This feature enables transparency in decision-making, which is crucial for sensitive applications such as digital forensics, identity verification, content authentication, and regulatory compliance. While current limitations include evaluation on a single dataset and the absence of adversarial testing, future work will address these constraints through cross-dataset validation and robust augmentation strategies. By presenting a high-performing, interpretable, and computationally efficient solution, this research contributes to the growing field of AI-generated content detection. The proposed SE-ResNet-50 model shows strong promise for practical deployment in real-world scenarios, including edge devices, mobile applications, and secure cloud-based platforms.

5.2. Future Work

While the proposed approach yielded promising results, several avenues remain open for exploration and enhancement:
  • Generalization Across Datasets and Domains:Future research should validate the model on larger and more diverse datasets such as Celeb-DF, FaceForensics++, and DeepFake Detection Challenge (DFDC) datasets to assess cross-domain generalization and robustness across different types of generative models and modalities [32].
  • Adversarial Robustness: As adversarial attacks pose a growing threat to deep learning-based classifiers, future work could explore adversarial training, robust optimization, or defensive distillation to enhance the model’s resilience against targeted manipulations and image perturbations.
  • Integration of Multi-modal Signals: Extending the current framework to incorporate multi-modal information—such as audio, text metadata, or temporal cues in video—could enhance performance for multimedia deepfake detection and support broader misinformation mitigation strategies.
  • Lightweight and Real-time Deployment: Optimization for real-time inference on edge devices and mobile platforms using techniques like quantization, pruning, or knowledge distillation will improve the model’s applicability in real-world, resource-constrained environments.
  • Explainable AI and Visual Attribution: To further improve trust and transparency, future work could incorporate visual attribution techniques such as Grad-CAM or saliency mapping to visualize which specific regions of the image influenced the model’s decision. This will benefit human-in-the-loop verification systems.
  • Future studies will evaluate the proposed model across diverse benchmarks such as Celeb-DF, FaceForensics++, and DFDC to assess cross-domain generalization and robustness against emerging generative models.
  • The current model was trained without data augmentation or adversarial perturbation strategies, which may affect its resilience in noisy or manipulated inputs. Incorporating adversarial training and augmentation (e.g., noise, rotations, occlusions) is an avenue for future improvement.

Author Contributions

All authors contributed equally to the conception, design, execution, data analysis, and writing of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study is publicly available at Kaggle: https://www.kaggle.com/datasets/birdy654/cifake-real-and-ai-generated-synthetic-images (accessed on 15 January 2025). The code used for analysis is not publicly available but can be obtained from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Roose, K. An AI-generated picture won an art prize. Artists aren’t happy. The New York Times, 2 September 2022. [Google Scholar]
  2. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 10684–10695. [Google Scholar]
  3. Pennycook, G.; Rand, D.G. The psychology of fake news. Trends Cogn. Sci. 2021, 25, 388–402. [Google Scholar] [CrossRef] [PubMed]
  4. Singh, B.; Sharma, D.K. Predicting image credibility in fake news over social media using multi-modal approach. Neural Comput. Appl. 2022, 34, 21503–21517. [Google Scholar] [CrossRef] [PubMed]
  5. Bonettini, N.; Bestagini, P.; Milani, S.; Tubaro, S. On the use of Benford’s law to detect GAN-generated images. In Proceedings of the 25th International Conference on Pattern Recognition (ICPR), Virtual Event, 10–15 January 2021; pp. 5495–5502. [Google Scholar]
  6. Deb, D.; Zhang, J.; Jain, A.K. AdvFaces: Adversarial face synthesis. In Proceedings of the IEEE International Joint Conference on Biometrics (IJCB), Houston, TX, USA, 28 September–1 October 2020; pp. 1–10. [Google Scholar]
  7. Khosravy, M.; Nakamura, K.; Hirose, Y.; Nitta, N.; Babaguchi, N. Model inversion attack: Analysis under gray-box scenario on deep learning based face recognition system. KSII Trans. Internet Inf. Syst. 2021, 15, 1100–1118. [Google Scholar]
  8. Bird, J.J.; Naser, A.; Lotfi, A. Writer-independent signature verification; evaluation of robotic and generative adversarial attacks. Inf. Sci. 2023, 633, 170–181. [Google Scholar] [CrossRef]
  9. Ramesh, A.; Pavlov, M.; Goh, G.; Gray, S.; Voss, C.; Radford, A.; Chen, M.; Beltagy, I. Zero-Shot Text-to-Image Generation. arXiv 2021, arXiv:2102.12092. [Google Scholar]
  10. Saharia, C.; Chan, W.; Saxena, S.; Li, L.; Whang, J.; Denton, E.L.; Ghasemipour, K.; Gontijo Lopes, R.; Karagol Ayan, B.; Salimans, T.; et al. Photorealistic text-to-image diffusion models with deep language understanding. arXiv 2022, arXiv:2205.11487. [Google Scholar]
  11. Chambon, P.; Bluethgen, C.; Langlotz, C.P.; Chaudhari, A. Adapting pretrained vision-language foundational models to medical imaging domains. arXiv 2022, arXiv:2210.04133. [Google Scholar]
  12. Schneider, F.; Kamal, O.; Jin, Z.; Schölkopf, B. Moûsai: Text-to-music generation with long-context latent diffusion. arXiv 2023, arXiv:2301.11757. [Google Scholar]
  13. Wang, J.; Wu, Z.; Ouyang, W.; Han, X.; Chen, J.; Jiang, Y.-G. M2TR: Multi-modal multi-scale transformers for Deepfake detection. In Proceedings of the International Conference on Multimedia Retrieval(ICMR), Newark, NJ, USA, 27–30 June 2022; pp. 615–623. [Google Scholar]
  14. Amerini, I.; Galteri, L.; Caldelli, R.; Del Bimbo, A. Deepfake video detection through optical flow based CNN. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1205–1207. [Google Scholar]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  16. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  17. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  18. Guo, C.; Dou, Y.; Bai, T.; Dai, X.; Wang, C.; Wen, Y. ArtVerse: A paradigm for parallel human–machine collaborative painting creation in Metaverses. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 2200–2208. [Google Scholar] [CrossRef]
  19. Güera, D.; Delp, E.J. Deepfake video detection using recurrent neural networks. In Proceedings of the 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018; pp. 1–6. [Google Scholar]
  20. Bird, J.J.; Lotfi, A. CIFAKE: Image Classification and Explainable Identification of AI-Generated Synthetic Images. IEEE Trans. Neural Netw. Learn. Syst. 2023, 12, 15642–15650. [Google Scholar] [CrossRef]
  21. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  22. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the Conference on North American Chapter of the Association for Computational Linguistics (NAACL-HLT), Minneapolis, MN, USA, 3–5 June 2019; pp. 4171–4186. [Google Scholar]
  23. Noack, B.; Schleich, C.; Wiatowski, T. An Empirical Study on the Relation Between Network Interpretability and Adversarial Robustness. SN Appl. Sci. 2020, 2, 32. [Google Scholar] [CrossRef]
  24. Boopathy, A.P.J.; Saha, A.; Chilimbi, T.M.; Varshney, K.R.; Sattigeri, P. Proper Network Interpretability Helps Adversarial Robustness in Classification. arXiv 2020, arXiv:2002.12254. [Google Scholar]
  25. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  26. Paszke, A. PyTorch: An imperative style, high-performance deep learning library. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 8–14 December 2019; pp. 8026–8037. [Google Scholar]
  27. Gong, J.; Liu, J.; Zhu, X.; Shi, Y.; Lv, H. Automated Pulmonary Nodule Detection in CT Images Using 3D Deep Squeeze-and-Excitation Networks. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1247–1254. [Google Scholar] [CrossRef] [PubMed]
  28. Ait El Asri, O.; Ennaji, M.; Oukili, M.; Regragui, B. Advanced Squeeze-and-Excitation Residual Network Based Methodology for Building Extraction. Int. J. Electr. Comput. Eng. 2024, 14, 1113–1123. [Google Scholar] [CrossRef]
  29. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 4700–4708. [Google Scholar]
  30. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  31. Yi, D.; Guo, C.; Bai, T. Exploring painting synthesis with diffusion models. In Proceedings of the IEEE 1st International Conference on Digital Twins and Parallel Intelligence (DTPI), Beijing, China, 19–21 July 2021; pp. 332–335. [Google Scholar]
  32. Sha, Z.; Li, Z.; Yu, N.; Zhang, Y. DE-FAKE: Detection and attribution of fake images generated by text-to-image generation models. arXiv 2022, arXiv:2210.06998. [Google Scholar]
Figure 1. Top row: Real images from the dataset. Bottom row: AI-generated fake images. (a) Real 1 (b) Real 2 (c) Real 3 (d) Fake 1 (e) Fake 2 (f) Fake 3.
Figure 1. Top row: Real images from the dataset. Bottom row: AI-generated fake images. (a) Real 1 (b) Real 2 (c) Real 3 (d) Fake 1 (e) Fake 2 (f) Fake 3.
Applsci 15 07421 g001
Figure 2. Architecture flowchart of the proposed SE-ResNet50 hybrid model.
Figure 2. Architecture flowchart of the proposed SE-ResNet50 hybrid model.
Applsci 15 07421 g002
Figure 3. ROC curve comparison—(a) VGGNet, (b) DenseNet, (c) SE-ResNet-50.
Figure 3. ROC curve comparison—(a) VGGNet, (b) DenseNet, (c) SE-ResNet-50.
Applsci 15 07421 g003
Figure 4. Confusion Matrix Comparison—(Left) ResNet-50, (Right) SE-ResNet50.
Figure 4. Confusion Matrix Comparison—(Left) ResNet-50, (Right) SE-ResNet50.
Applsci 15 07421 g004
Figure 5. Training and validation trends for SE-ResNet-50 across 10 epochs. Metrics include accuracy, precision, recall, and loss.
Figure 5. Training and validation trends for SE-ResNet-50 across 10 epochs. Metrics include accuracy, precision, recall, and loss.
Applsci 15 07421 g005
Table 1. Training hyperparameters for baseline and hybrid models.
Table 1. Training hyperparameters for baseline and hybrid models.
ModelLearning RateBatch SizeEpochs
ResNet-500.00013210
VGGNet-160.00013210
Hybrid (SE-ResNet50)0.00013210
Table 2. Performance metrics of baseline models.
Table 2. Performance metrics of baseline models.
ModelAccuracyPrecisionRecallF1-ScoreROC-AUC
ResNet-50 96.12 % ± 0.25 % 97.04 % ± 0.50 % 88.94 % ± 0.50 % 92.82 % ± 0.50 % 0.9862
VGGNet 94.06 % ± 0.42 % 96.60 % ± 0.42 % 85.43 % ± 0.42 % 90.04 % ± 0.42 % 0.98
DenseNet 95.98 % ± 0.26 % 96.25 % ± 0.26 % 87.80 % ± 0.26 % 91.07 % ± 0.27 % 0.98
Table 3. Confusion matrix for ResNet-50 on CIFAKE dataset.
Table 3. Confusion matrix for ResNet-50 on CIFAKE dataset.
Predicted RealPredicted AI-Generated
Real9729271
AI-Generated11068894
Table 4. Comparison with prior works on AI-generated image detection.
Table 4. Comparison with prior works on AI-generated image detection.
ModelDatasetAccuracy (%)AUC
ResNet-50 (baseline)CIFAKE93.460.962
VGGNet-16CIFAKE95.980.980
DenseNet-121CIFAKE96.060.980
Fine-tuned ViTCIFAKE92.00-
Proposed SE-ResNet-50CIFAKE96.120.9862
Table 5. McNemar contingency table (SE-ResNet50 vs. DenseNet).
Table 5. McNemar contingency table (SE-ResNet50 vs. DenseNet).
DenseNet CorrectDenseNet Wrong
SE-ResNet50 Correct8780918
SE-ResNet50 Wrong189113
Table 6. Performance metrics of SE-ResNet50 on the CIFAKE dataset.
Table 6. Performance metrics of SE-ResNet50 on the CIFAKE dataset.
MetricSE-ResNet50 (%)
Accuracy96.12 ± 0.25
Precision97.04 ± 0.50
Recall88.94 ± 0.50
F1-Score92.82 ± 0.50
ROC-AUC0.9862
Table 7. Confusion matrix comparison: ResNet-50 vs. SE-ResNet50.
Table 7. Confusion matrix comparison: ResNet-50 vs. SE-ResNet50.
ModelPredicted RealPredicted AI-Generated
ResNet-509541 (Real)459 (Real)
328 (AI)9672 (AI)
SE-ResNet508894 (Real)1106 (Real)
271 (AI)9729 (AI)
Table 8. Training efficiency comparison (per epoch).
Table 8. Training efficiency comparison (per epoch).
ModelTime per Epoch (s)GPU Utilization (%)
ResNet-5028537
VGGNet30542
DenseNet64656
SE-ResNet5031848
Table 9. Detailed results of the hybrid model (SE-ResNet50).
Table 9. Detailed results of the hybrid model (SE-ResNet50).
ModelAccuracy (%)Training Time (s)GPU Utilization (%)
Hybrid (ResNet + SE Attention)96.12 ± 0.2531848
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gunukula, A.R.; Das Gupta, H.; Sheng, V.S. Detecting AI-Generated Images Using a Hybrid ResNet-SE Attention Model. Appl. Sci. 2025, 15, 7421. https://doi.org/10.3390/app15137421

AMA Style

Gunukula AR, Das Gupta H, Sheng VS. Detecting AI-Generated Images Using a Hybrid ResNet-SE Attention Model. Applied Sciences. 2025; 15(13):7421. https://doi.org/10.3390/app15137421

Chicago/Turabian Style

Gunukula, Abhilash Reddy, Himel Das Gupta, and Victor S. Sheng. 2025. "Detecting AI-Generated Images Using a Hybrid ResNet-SE Attention Model" Applied Sciences 15, no. 13: 7421. https://doi.org/10.3390/app15137421

APA Style

Gunukula, A. R., Das Gupta, H., & Sheng, V. S. (2025). Detecting AI-Generated Images Using a Hybrid ResNet-SE Attention Model. Applied Sciences, 15(13), 7421. https://doi.org/10.3390/app15137421

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop