Next Article in Journal
Symmetry-Guided Surrogate-Assisted NSGA-II for Multi-Objective Optimization of Renewable Energy Systems
Previous Article in Journal
PSO-Driven Scalable Dual-Adaptive PV Array Reconfiguration Under Partial Shading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lightweight Financial Fraud Detection Using a Symmetrical GAN-CNN Fusion Architecture

1
School of Business, Sias University, No.168 Renmin Road, Xinzheng 451150, China
2
School of Software, Jiangxi Normal University, Nanchang 330022, China
3
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(8), 1366; https://doi.org/10.3390/sym17081366
Submission received: 23 June 2025 / Revised: 16 August 2025 / Accepted: 18 August 2025 / Published: 21 August 2025
(This article belongs to the Section Computer)

Abstract

With the rapid development of information technology and the deep integration of the Internet platform, the scale and form of financial transactions continue to grow and expand, significantly improving users’ payment experience and life efficiency. However, financial transactions bring us convenience but also expose many security risks, such as money laundering activities, forged checks, and other financial fraud that occurs frequently, seriously threatening the stability and security of the financial system. Due to the imbalance between the proportion of normal and abnormal transactions in the data, most of the existing deep learning-based methods still have obvious deficiencies in learning small numbers sample classes, context modeling, and computational complexity control. To address these deficiencies, this paper proposes a symmetrical structure-based GAN-CNN model for lightweight financial fraud detection. The symmetrical structure can improve the feature extraction and fusion ability and enhance the model’s recognition effect for complex fraud patterns. Synthetic fraud samples are generated based on a GAN to alleviate category imbalance. Multi-scale convolution and attention mechanisms are designed to extract local and global transaction features, and adaptive aggregation and context encoding modules are introduced to improve computational efficiency. We conducted numerous replicate experiments on two public datasets, YelpChi and Amazon. The results showed that on the Amazon dataset with a 50% training ratio, compared with the CNN-GAN model, the accuracy of our model was improved by 1.64%, and the number of parameters was reduced by approximately 88.4%. Compared with the hybrid CNN-LSTM–attention model under the same setting, the accuracy was improved by 0.70%, and the number of parameters was reduced by approximately 87.6%. The symmetry-based lightweight architecture proposed in this work is novel in terms of structural design, and the experimental results show that it is both efficient and accurate in detecting imbalanced transactions.

1. Introduction

With the rapid development of information technology and the widespread use of Internet platforms, the number and types of financial transactions have increased dramatically, providing great convenience for users’ lives. Financial transactions provide users with a good sense of transaction experience; however, at the same time, they also provide a breeding ground for financial fraud, such as by malicious actors through technical loopholes or illegal means, money laundering, counterfeiting checks, and other financial fraud, posing a serious threat to the security of financial institutions and the stability of the global financial system. According to data from the International Monetary Fund (IMF), globally, illegal money laundering activities account for about 2% to 5% of gross national product each year, and this proportion is on the rise year by year [1]. According to the Global Financial Crime Report 2024 published by Nasdaq, financial fraud brought nearly 500 billion dollars of losses to the world in 2023, and the frequent occurrence of financial fraud cases has become a major challenge for the financial industry. Therefore, strengthening financial fraud detection has become a prominent research focus.
One of the distinguishing features of financial fraud is the problem of category imbalance. Since the number of normal financial transactions is much larger than the number of abnormal fraudulent transactions, this poses a greater challenge for financial fraud detection. Financial fraud detection methods are categorized into two types: (1) traditional financial fraud detection methods, and (2) deep learning financial fraud detection methods. Traditional financial fraud detection methods mainly include random forest (RF) [2], decision tree (DT) [3], and logistic regression (LR) [4] methods. For example, Lin [2] analyzed a dataset of 150 bank transactions using a random forest technique to evaluate its effectiveness in identifying financial frauds, and the experimental results showed the high efficiency of this method in detecting fraudulent transactions. Devi and Kavitha [3] proposed a decision tree-based method for classifying credit card transactions as normal or suspicious and used multiple metrics for their evaluation, with high accuracy. Liu [4] demonstrated the use of logistic regression in telecom bank card fraud detection, with good results. Ali et al. [5] emphasized the efficiency of support vector machines (SVMs) in dealing with category-imbalanced datasets. Wu et al. [6] concluded that the simple Bayesian algorithm (NB) with its simple structure and high computational efficiency is suitable for the task of credit assessment. The NB algorithm is a simple and computationally efficient algorithm for credit assessment and it can be applied to the credit assessment task.
Despite the effectiveness of the above methods, there is a high risk of misclassification when dealing with complex patterns, and such models have limitations in dealing with complex data. For example, the selection of features needs to be based on specialized domain knowledge. In order to overcome the shortcomings of the above methods, scholars have proposed deep learning-based financial fraud detection methods. Thilagavathi et al. [7] proposed combining graph neural networks (GNNs) with anomaly detection to identify complex financial fraud such as account supervision, identity theft, etc. Li et al. [8] used graph convolutional networks (GCNs) for fraud detection in cryptocurrency trading networks. The evolutionary properties of the network were considered, combined with local structure and equilibrium theory to guide training. Experimental results show that the model is well validated. Nalayini et al. [9] proposed a CNN model based on a smart matrix algorithm for credit card fraud detection using large real-time datasets. Wu et al. [10] applied generative adversarial networks (GANs) in combination with deep learning for financial fraud detection to improve precision and recall through data augmentation and enhance financial transaction security. Fletcher et al. [11] proposed a financial fraud detection model based on a self-attention mechanism, which improves recognition accuracy by capturing remote dependencies and global information. Zhao et al. [12] proposed a credit card fraud detection method based on self-attention generative adversarial networks (SAGANs), which distinguishes the key features through a self-attention mechanism that generates data reflecting real fraudulent behavior, which significantly improves the detection accuracy and recall rate.
Although existing deep learning models in financial fraud detection have made some progress, there are still some problems. Unbalanced data categories can be a problem. Most deep learning relies on a large number of training samples, and when the training samples are insufficient, the training process is biased towards normal transaction data and ignores a few fraudulent behaviors, resulting in low recall. Furthermore, capture of contextual feature information may be inadequate. Existing methods have difficulty adequately capturing key information such as time series and money flows in dynamic transaction networks, which affects detection accuracy. Low computational efficiency can occur with a large number of parameters. Most of the existing deep learning models suffer from large numbers of parameters, high computational complexity, and high feature dimensionality.
To solve the above problems, this study proposes a new symmetrical structure-based GAN-CNN model for detecting financial fraud. In this model, based on a lightweight symmetrical design, fraud samples are generated to alleviate the category imbalance problem, and temporal features and graph structure features are extracted to enhance the model’s ability to model contextual information.
The contributions of this study are reflected in the following aspects. To solve the category imbalance problem, GAN is used to generate data samples to balance the ratio of normal and fraudulent transactions in the dataset, thus enhancing the model’s ability to learn from a few fraudulent samples. To enhance the ability to capture contextual feature information, CNN is introduced to comprehensively consider the temporal features and graph structure features in the transaction data, to enhance the model’s ability to express the transaction relationship and to accurately recognize complex fraudulent behaviors. To reduce the complexity of model computation, the design of the lightweight symmetrical multi-scale attention mechanism and adaptive aggregation module effectively reduces the number of parameters and feature dimensions of the model, further improving the computational performance.

2. Related Work

2.1. Traditional Financial Transaction Fraud Detection Methods

Traditional financial transaction detection methods include methods such as plain Bayes, logistic regression, random forest, and decision tree. For example, Gupta et al. [13] investigated how to utilize the plain Bayes algorithm for credit card fraud detection, especially in the face of highly unbalanced data, and found that plain Bayes can effectively identify credit card fraud. Du et al. [14] proposed a transaction fraud detection method based on random forests, and the experimental results showed significant advantages. In order to cope with the data imbalance problem, Wan and Zhao [15] proposed an improved logistic regression model with a synthetic minority over-sampling technique (SMOTE) and cross-validation to optimize the model parameters, which significantly improved the model performance, especially in terms of accuracy and AUC value. Similarly, Aburbeian and Ashqar [16] proposed an enhanced version of a random forest classifier with SMOTE technique to address the data imbalance problem, and after optimizing the hyper-parameters, the method demonstrated stronger credit card fraud detection capabilities, especially in complex data environments.
In terms of algorithm comparison, Afriyie et al. [17] compared three algorithms, logistic regression, random forest, and decision tree, and the results showed that random forest performed the best in terms of accuracy and AUC value, making it the best detection algorithm by their recommendation. Li and Geng [18] used the BLS algorithm to deal with imbalanced credit data, and compared four algorithms’ performance, namely logistic regression, simple Bayes, support vector machine, and random forest; The results show that the Random Forest classifier is most suitable for classifying the credit dataset, and in the future, Random Forest can be used as a base classifier to develop an integrated model to further improve the prediction accuracy and performance. The assessment of the four algorithms showed that random forest performed best in terms of accuracy, AUC value, and other evaluation indexes. In addition, Zhu et al. [19] pointed out that logistic regression and plain Bayes are suitable for dealing with structurally simple classification tasks, support vector machines are adapted to high-dimensional small-sample problems, and random forests perform particularly well in terms of classification accuracy and model robustness. Kumar et al. [20] proposed a machine learning algorithm based on support vector machines (SVMs) for improving the accuracy of credit card fraud detection, effectively making up for the shortcomings of traditional methods in recognizing complex transaction fraud. Obeng et al. [21] analyzed the effect of multiple machine learning algorithms for fraud detection in financial transactions and found that the eXtreme Gradient Boosting (XGBoost) and random forest models performed well.

2.2. Deep Learning Based Fraud Methods for Financial Transactions

Deep learning methods play an important role in financial fraud detection. GNN and GCN enhance relationship modeling through graph structure and node features, GAN alleviates the data imbalance problem, and CNN extracts spatiotemporal features to improve detection accuracy; several studies have verified their effectiveness. For example, the ASA-GCN model proposed by Ahmad [22] enhances fraud detection performance through adaptive sampling and aggregation mechanisms. The GNN-CL model proposed by Cheng et al. [23] combines GNNs, CNNs, and LSTMs, which enhances the ability to handle complex transaction patterns. The FinFD-GCN model proposed by Yadegar and Rahmani [24] optimizes fraud detection in credit card transactions through graph convolutional networks. The spatiotemporal graph neural network proposed by Khosrav et al. [25] effectively improves the detection efficiency. The two-channel GAN model by Wei and Lee [26] significantly improves the accuracy and robustness of fraud detection in financial transactions.
The FraudGNN-RL model proposed by Zhang et al. [27] combines graph neural networks with reinforcement learning, which enhances the model’s ability to cope with dynamically changing fraud patterns. The credit card fraud detection model based on dilated convolutional neural networks (DCNNs) with a sampling technique designed by Karthika and Senthilselvi [28] improves the detection accuracy. The hybrid SMOTE-GAN technique proposed by Cheah et al. [29] effectively solves the data imbalance problem and improves the model performance. Strelcenia and Prakoonwit [30] investigated the application of generative adversarial networks (GANs) in credit card fraud detection. Li et al. [31] proposed the TA-Struc2Vec algorithm to detect credit card fraud by combining the topological features and transaction amount features of financial transaction networks to improve the efficiency of Internet financial fraud detection. Duan et al. [32] proposed a CaT-GNN model combining causal inference and time graph neural network to improve the robustness of credit card fraud detection.

2.3. Attention Mechanism Based Financial Transaction Fraud Detection Method

Financial fraud detection based on attention mechanisms can extract key features, recognize complex patterns, learn long-distance dependencies, apply dynamic fraud behaviors, and perform superiorly in different financial scenarios. For example, Zhao et al. [12] constructed a self-attentive generative adversarial network (GAN) combined with a graph attention technique to recognize fraudulent transactions, enhancing the model’s adaptive ability and data security. Wei and Lee [26] used a dual-channel graph attention network (GAT) combined with GBDT-DGAN to improve the accuracy of fraud detection and introduced blockchain technology to safeguard data privacy. Xie et al. [33] proposed a spatiotemporal gating network (STGN) to optimize credit card fraud detection through temporal attention and location-aware mechanisms, highlighting the spatiotemporal characteristics of transaction behaviors. Yu et al. [34] explored the application of the transformer model in credit card fraud detection and compared it with a variety of traditional models, demonstrating the transformer’s ability to capture complex transaction patterns. Ghosh et al. [35] combined a multilayer perceptron (MLP), feed-forward neural network (FNN), and attention LSTM in the Bitcoin network, which effectively enhanced fraud detection ability for Bitcoin transactions. Li et al. [36] designed a method based on sliding windows and attention LSTM for Internet financial credit assessment, which improved the accuracy of risk prediction.

3. Method

3.1. Overview

This paper proposes a lightweight financial fraud detection model based on GAN-CNN, which adopts a symmetric branch architecture. As shown in Figure 1, the model includes key stages such as data augmentation, feature modeling, feature fusion, context modeling, and classification decision-making, forming a broader detection process. The model first uses a generative adversarial network (GAN) to synthesize diverse fraud samples, alleviating class imbalance and expanding the distribution of minority class samples to achieve data balancing, thereby providing stable training conditions for subsequent feature learning. A multi-scale attention mechanism simultaneously models local and global transaction features at different perceptual scales, with channel attention and spatial attention working together to enhance the representation accuracy of key patterns. The adaptive aggregation module screens high-correlation transaction relationships based on similarity calculations and random discard strategies, suppressing noise interference and enhancing the ability to identify key fraud patterns. The feature optimization fusion network (FRFN) combines deep separable convolutions and residual connections to integrate multi-level, multi-scale features, improving the sensitivity of complex fraud pattern recognition and model robustness. The context encoding module integrates time-frequency features and account out-degree features to characterize account behavior patterns in dynamic transaction networks, supplementing temporal and structural information to strengthen discriminative evidence. Boosting and bagging ensemble learning fuses outputs from various stages to achieve efficient and precise fraud detection for large-scale high-frequency transaction data.

3.2. Sample Generation

As shown in Figure 2, the GAN generates fraudulent transaction samples by learning real transaction data, balancing the data distribution, and preventing the model from over-biasing towards normal transactions [37]. The generated fraud data is mixed with real data and key nodes and edges are selected by a subgraph generator to enhance the diversity of fraud patterns, while the hidden graph structure information is used to optimize the model [38]. Different connection methods (e.g., dynamic connection) are used to capture complex fraudulent relationships in the transaction network and improve the generalization ability of the detection system. During the training process, the loss function is optimized to ensure the similarity in distribution between the generated data and the real fraudulent transaction data. The loss function is shown below:
LOSS GAN   = m = 1 M n = 1 M x m x n 2 2
where x m and x n denote the feature vectors of the two transaction samples respectively, this loss function measures the Euclidean distance between the samples, ensuring that the generated data points maintain the structural properties of the fraudulent transactions in the feature space, thus improving the detection model’s ability to recognize fraudulent patterns. The method effectively mitigates the category imbalance problem and improves the accuracy and robustness of financial fraudulent transaction detection by using GAN to generate synthetic fraudulent transactions, optimizing the graph structure information, and dynamically constructing the transaction network.

3.3. Adaptive Aggregation Module

As shown in Figure 3, the adaptive aggregation module extracts key features in the transaction network through similarity computation, random discard, and weighted aggregation to improve the accuracy of financial fraud detection. Specifically, node similarity m u , v is first computed to measure account transaction associations, and nodes and their neighboring features are processed by feature mapping M v . Before entering the aggregation phase, the model employs similarity-driven stochastic discarding to remove neurons with a certain probability α to reduce irrelevant transaction interference and prevent model overfitting. Subsequently, the adaptive aggregator weights and fuses the feature information to improve the recognition of abnormal transaction patterns (e.g., money laundering, ring trading, and high-frequency abnormal trading). Finally, the model is regularized using the Dropout mechanism combined with the SeLU activation function, calculated as follows:
h θ i 0 = σ dropout x i , α · ω 1
where dropout x i , α randomly deletes neurons with probability α to reduce the model’s dependence on specific transaction patterns and improve the generalization ability; ω 1 is used to linearly transform the features and optimize the feature expression; σ · is the activation function; and SeLU is used here to enhance the network’s nonlinear expression ability and improve the stability of training. Through this mechanism, the model can efficiently screen key transaction features, reduce noise interference, and improve the stability and accuracy of financial fraud detection.

3.4. Attention Mechanisms

As shown in Figure 4, the model introduces a lightweight multi-scale attention mechanism for modeling inter-account transaction associations and enhancing the characterization of financial fraud patterns. The input data is first subjected to layer normalization (LN) to stabilize the feature distribution and enhance the training effect, followed by the generation of query vectors (Q), key vectors (K), and value vectors (V). Q is processed in three parts: ReLU activation to enhance the nonlinear representation, linear transformation to adapt to diverse trading patterns, and attention computation to focus on key trading relationships and highlight abnormal behavioral features. Q and K are input into the lightweight attention module to generate attention weights to enhance the trading features that are highly correlated with the target account. Q and K are each subjected to a 3 × 3 deep separable convolution ( DWConv 3 × 3 ) and 1 × 1 standard convolution ( Conv 1 × 1 ), as shown in Table 1. Under the condition that the number of inputs, outputs, and channels is 1024, the depthwise separable convolution significantly reduces the number of parameters compared with the ordinary convolution. Specifically, for example, the 3 × 3 and 5 × 5 convolution kernels are reduced by 18.9 M and 52.4 M, respectively, reflecting higher computational efficiency and lightweight advantages. Subsequently, after combining the attention information, the dimensionality is adjusted by linear transformation (Linear) to fit the classification task. While maintaining the model’s high efficiency, the mechanism effectively integrates local and global transaction features, improving the detection capability of complex fraudulent behaviors such as abnormal fund flows, ring trading, and high-frequency trading, and making the identification results more accurate and stable.

3.5. Feature-Response Fusion (FRFN)

As shown in Figure 5, in financial fraud transaction behavior detection, the FRFN module is used to integrate different transaction features to improve detection accuracy. Specifically, first, the input matrix x H × W × C is layer normalized (LN) to stabilize the data distribution and prevent the gradient problem. Then, parallel convolution (Parallel Conv 3 × 3 ) extracts trading patterns at different scales to enhance the model’s perception of long and short-term trading behaviors. The data is processed linearly and reshaped to match the input structure of the depthwise separable convolutional layer ( DWConv 3 × 3 ).   DWConv 3 × 3 extracts localized features to reduce computational complexity and maintain spatial features. Then, the data is flattened (Flatten), and the original features are fused with the processed features using residual concatenation, followed by batch normalization (BN) to stabilize the feature distribution. Further, another layer of DWConv 3 × 3 enhances feature extraction to identify hidden fraud. Finally, the data is nonlinearly mapped by the SeLU activation function to improve the model’s robustness. The whole FRFN structure accurately detects abnormal transaction patterns through the steps of convolution, normalization, and nonlinear transformation to improve financial fraud detection performance.

3.6. Text Encoding

To capture the fraudster’s contextual behavior more accurately, we introduce a context encoding module, which combines time-frequency features without-degree features in order to generate context encodings that effectively characterize each node. As shown in Figure 6. The encoding size is controlled by hyperparameters and fused with hidden node embeddings in subsequent stages to finally output an optimized Z V representation of the node.

3.6.1. Time-Frequency Coding

To describe the trading behavior of a particular entity in a time series, its set of transactions is defined as follows:
E V out = V , u 1 , t 1 ) , V , u 2 , t 2 , ( V , u T , t T )
where V denotes a fixed entity identifier (e.g., user or account) and u i represents the type of transaction or operation that occurred at time t i for that entity. All transactions are arranged in chronological order to construct a time series. To quantify the time intervals of these transactions, their average time intervals are calculated as follows:
t V = t 2 t 1 + t 3 t 2 + + t T t T 1 T 1
This formula calculates the time difference between neighboring transaction events and averages them to represent the time interval characteristics of the transaction. We perform a further feature transformation on the time intervals to obtain the time-frequency encoding (TFE), as follows:
T E t v i = SeLU W TE t v
where W TE is the weight matrix and SeLU is used as a nonlinear activation function to transform the temporal features to obtain a feature representation applicable to the model inputs.

3.6.2. Out-Degree Coding

To extract the out-degree information, the out-degree encoding is computed for the out-degree feature associated with each index θ i , as follows:
d e θ i = SeLU W Ode d θ i out
where SeLU (scaled exponential linear unit) is the nonlinear activation function, and W Ode is the weight matrix, which is responsible for linearly transforming the input out-degree feature d θ i out and nonlinearly mapping it through SeLU to optimize the out-degree feature d e θ i .

3.6.3. Time-Frequency and Out-Degree Coding Fusion

In financial transaction fraud detection, the model extracts transaction behavior features by combining time-frequency features and out-degree coding to identify anomalous transaction patterns. First, the model fuses the time-frequency information T E t V i and out-degree feature Ode V i to compute the contextual coding C E v as follows:
C E v = 1 β T E t v i + β Ode v i
Subsequently, the model splices the context-encoded C E v with the multilevel features h 0 ^ and performs a nonlinear mapping via the weight transformation W Z and the S e L U activation function in order to obtain the optimized transaction features Z V as follows:
Z V = SeLU W Z h 0 ^ & C E v
The initial feature representation h 0 ^ is calculated by the feature refined fusion network (FRFN), integrating transaction data, historical behavioral characteristics, and account association information at multiple levels, as follows:
h 0 ^ = FRFN h v 1 & h v 2 & & h v k   &   h r & X v
Through this computational process, the model is able to learn the behavioral patterns of trading accounts to effectively detect anomalous transactions, improving the accuracy of fraud detection and the security of the financial system.

4. Experiment

4.1. Datasets

The YelpChi and Amazon datasets are standard datasets commonly used in the field of financial fraud detection and are widely used for the training and evaluation of graph neural network models. As shown in Table 2, the YelpChi dataset [39] contained 45,954 nodes and 3,846,979 edges, mainly covering hotel and restaurant review data, in which the proportion of fraudulent reviews was 14.53%. The dataset defines three types of relationships: R-U-R denotes connections between reviews posted by the same user, R-S-R denotes relationships between reviews with the same rating, and R-T-R denotes connections between reviews posted in the same period. The Amazon dataset [1] contained 11,944 nodes and 4,398,392 edges covering reviews of products in the category of music equipment. The percentage of fraudulent reviews was 6.87%. The relationship types include U-P-U for connections between users reviewing the same product, U-S-U for connections between users giving the same ratings in the same period, and U-V-U for relationships between users calculated based on the similarity of review texts.

4.2. Experimental Setup and Metrics

4.2.1. Experimental Setup and Realization

All experiments were conducted under the same training parameters, which are shown in Table 3. Referring to existing models [40,41,42,43,44,45,46,47,48,49,50,51,52,53,54], we adopted different training ratio settings for the YelpChi and Amazon datasets. Specifically, the experiments were conducted at 20% and 50% training ratios for the YelpChi dataset, while the training ratio for the Amazon dataset was set at 50%. The experiments were implemented based on the PyTorch framework, and parameters such as optimizer, learning rate, and convolutional layer structure were kept the same. To reduce the effect of randomness, we repeated each set of experiments independently 10 times; the mean and standard deviation are reported. During training, the learning rate was set to 10−3, the weight decay coefficient to 5 × 10−4, and the momentum parameter to 0.73 to control training stability and prevent overfitting. Node embeddings were extracted across multiple layers based on the FRFN module structure, utilizing deep separable convolutions and residual connections to enhance representational capacity. The entire model was trained for 185 epochs, with each epoch using a batch size of 16 for iterative optimization. To improve the model’s generalization performance, we introduced the dropout mechanism in the multi-scale attention mechanism and adaptive aggregation module, with a probability of 0.5. For the time-frequency and out-degree features in the context encoding module, the SeLU activation function was used for nonlinear mapping to enhance the representation of dynamic transaction behavior. Additionally, we set differentiated hyperparameters for tuning on the YelpChi and Amazon datasets to adapt to different graph structures and fraud rates. YelpChi used a saturation parameter of 1.7, while Amazon used a subclassification score of 64, ensuring the model’s stability and effectiveness across multiple scenarios.

4.2.2. Metrics

This paper uses overall accuracy (OA), confusion matrix (CM), number of model parameters, gigaflops per second (GMACs), precision, recall, and F1 as evaluation metrics. Overall accuracy measures the model’s overall classification performance; the confusion matrix visually displays the prediction results for each category and the degree of confusion; GMACs indicates the number of multiply-accumulate operations the model performs per second during inference, reflecting computational complexity; the number of parameters measures the model’s size and structural complexity; precision evaluates the proportion of samples correctly classified as positive by the model; recall indicates the proportion of all true positive samples that are successfully identified; F1 is the harmonic mean of precision and recall, comprehensively reflecting the trade-off between accuracy and coverage. The relevant formulas for these metrics are as follows:
OA = TP + TN TP + TN + FP + FN
Precision = TP TP + FP
Recall = TP TP + FN
F 1 = 2 PrecisionRecall Precision + Recall = 2 TP 2 TP + FP + FN  
where true positives (TPs) are the number of fraudulent transactions correctly identified by the model, false positives (FPs) are the number of legitimate transactions incorrectly identified as fraudulent, false negatives (FNs) are the number of fraudulent transactions not identified by the model, and true negatives (TNs) are the number of legitimate transactions correctly identified by the model as non-fraudulent.

4.3. Experimental Results

4.3.1. Experimental Results on YelpChi and Amazon Datasets

  • Comparison of classification performance with different training ratios
As shown in Table 4, the proposed Symmetrical GAN-CNN model achieved an overall accuracy of 94.97% and 97.15% on the YelpChi dataset and 96.37% on the Amazon dataset under a 20% and 50% training ratio, respectively. Compared with GraphConsis, CARE-GNN, and HHLN-GNN models, the accuracy at the 20% training ratio improved by 3.75%, 3.67%, and 0.15% on the YelpChi dataset, respectively; at the 50% training ratio, it improved by 2.31%, 2.63%, and 1.18% on YelpChi, and on Amazon, it improved by 1.53%, 1.74%, and 1.45%. In addition, the accuracy of the model on the Amazon dataset was also superior to that of GCN+SMOTE, CNN+SMOTE, CNN-GAN, and hybrid CNN-LSTM–attention models, with improvements of 1.58%, 1.69%, 1.64%, and 0.70%, respectively. This result shows that symmetrical GAN-CNN maintains excellent performance under both data-scarce and sufficient conditions and remains stable, especially when the training samples are small. This is mainly due to the introduction of the GAN sample generation module, which effectively mitigates the category imbalance, avoids model bias towards majority class transactions, and improves the learning ability for fraudulent transactions (minority class samples).
2.
Improvement in the effect of the attention mechanism on accuracy
As shown in Table 4, the accuracy of the benchmark attention model GAT on the YelpChi dataset at a 20% training ratio was 91.28%, and on the Amazon dataset it was 93.52% at a a 50% training ratio, which was a significant improvement in accuracy compared with the XGBoost model without the introduction of the attention mechanism, indicating that the attention mechanism helped to enhance the model’s ability to focus on key features. The Symmetrical GAN-CNN model proposed in this paper further improved the detection performance based on the fusion of channel attention and spatial attention mechanisms, multi-scale feature extraction, and a depth-separable convolutional structure. The accuracy on the YelpChi and Amazon datasets reached 94.97% and 96.37%, respectively, improvements of 3.69% and 2.85% over the GAT model, 3.75% and 1.53% over the GraphConsis model, 0.15% and 1.45% over the HHLN-GNN model, and up to 9.50% and 7.58% compared with the XGBoost model. At the 50% training ratio, symmetrical GAN-CNN achieved 97.15% and 96.37% accuracy on the YelpChi and Amazon datasets, outperforming all kinds of comparative models, with improvements of 2.38% and 2.85% compared with the GAT model, 2.31% and 1.53% compared with the GraphConsis model, 1.18% and 1.53% compared with the HHLN-GNN model, In addition, compared with the GCN model and the model method that introduced SMOTE technology, the performance of model in this paper improved by 1.30% and 1.39%, respectively, on the YelpChi dataset; compared with the CNN-GAN and hybrid CNN-LSTM–attention models, it also improved by 0.45% and 0.22%, respectively. These results show that the proposed lightweight multi-scale attention mechanism has significant advantages in improving detection accuracy, in which channel attention strengthens the focus on key features such as fund flow direction and frequency, and spatial attention and DWConv structure synergistically model the local and global transaction relationship so that the model has a stronger discriminative ability in complex fraud identification.
3.
Enhanced Recognition of Anomalous Structures
As shown in Table 4, the overall accuracy of the symmetrical GAN-CNN improved on both the YelpChi and Amazon datasets compared with GraphSAINT, FdGars, HHLN-GNN, GCN+SMOTE, CNN+SMOTE, CNN-GAN, and hybrid CNN-LSTM–attention models, with the maximum improvement of 1.45% on the Amazon dataset. The model extracts transaction features at different scales by introducing depth-separable convolutions, where the convolution is used to capture local anomalous behaviors, and the convolution focuses on modeling the relationship between long-distance money flows, thus enhancing the ability to express complex transaction structures. Meanwhile, the context coding module, which combines time-frequency and out-degree information, effectively improves the model’s ability to model dynamic trading behavior. In the feature fusion stage, the model introduces the batch normalization (BN) operation to standardize the feature distribution, improve the training stability, and reduce the feature dimension and computational cost while improving the accuracy of identifying abnormal fund flows, ring trades, and high-frequency frauds.

4.3.2. Model Compactness, Computational Efficiency, and Minority Class Recognition Capability Analysis

This paper selected GCN, SGC, GAT, FAGCN, FdGars, GraphSAINT, HHILN-GNN, GCN+SMOTE, CNN+SMOTE, CNN-GAN, and hybrid CNN-LSTM–attention as comparison models, and comprehensively evaluated their performance in terms of classification accuracy, parameter count, GMACs, precision, recall, and F1. As shown in Table 5, the detailed information of different models and the model proposed in this paper is listed under the YelpChi dataset with a training ratio of 20%. From this, we can obtain the following observations:
  • Parameter scale and model compactness analysis: While most models achieve a classification accuracy of approximately 90%, this is typically accompanied by a large model parameter scale. For example, HHLN-GNN achieves an accuracy of 94.82% on the YelpChi dataset, with a parameter count of 5.5 M, while the symmetrical GAN-CNN model proposed in this paper maintains a slightly higher accuracy of 94.97% with a parameter count of only 4.92 M. Additionally, the classification accuracy of Symmetrical GAN-CNN is 94.97%, which is higher than that of CNN-GAN (94.52%) and hybrid CNN-LSTM–attention (94.75%). Furthermore, the GCN+SMOTE and CNN+SMOTE models have parameter counts of 36.4 M and 49.23 M, respectively, which are significantly higher than the model proposed in this paper. Overall, symmetrical GAN-CNN demonstrates superior performance in parameter count control, showcasing stronger model compression capabilities and deployment adaptability.
  • Comparison of computational overhead under the GMACs metric: Compared with HHLN-GNN, the proposed model reduces computational overhead by 0.18 G, while it is slightly higher than GraphSAINT (+0.17 G). In addition, compared with GCN+SMOTE, CNN+SMOTE, CNN-GAN, and hybrid CNN-LSTM–attention, the symmetrical GAN-CNN reduces GMACs by 62%, 78%, 76%, and 67%, respectively, demonstrating its powerful lightweight advantages and inference efficiency. This performance is particularly critical in financial fraud detection tasks, effectively supporting real-time processing and providing a rapid response to large-scale, high-frequency transaction data, offering greater application value.
  • Comprehensive evaluation of minority class recognition capabilities based on Precision, recall, and F1: To more comprehensively evaluate the model’s performance in minority class identification tasks, this paper introduces three additional metrics for supplementary analysis—precision, recall, and F1. On the YelpChi dataset, the symmetrical GAN-CNN model outperformed mainstream comparison models in recall, with improvements of 14.6%, 6.3%, 5.7%, and 5.6% compared with GraphConsis, CARE-GNN, CNN-GAN, and hybrid CNN-LSTM–attention models, respectively, significantly enhancing its ability to identify genuine fraudulent transactions. Additionally, compared with GCN+SMOTE and CNN+SMOTE models that employ oversampling strategies, the symmetrical GAN-CNN model achieved improvements of 6.2% and 8.4% in recall and 4.1% and 4.7% in F1, further validating its detection advantages in imbalanced data scenarios. Overall, symmetrical GAN-CNN outperformed the CNN-GAN and hybrid CNN-LSTM–attention models in F1 metrics, demonstrating stronger classification robustness and minority class sensitivity, making it suitable for practical applications such as identifying high-risk transactions.

4.4. Ablation Experiments

To systematically evaluate the impact of each key module in the proposed model on the classification performance, this study included progressive ablation experiments using the YelpChi dataset with a training ratio of 20% as an example. Starting from the base model, the experiment sequentially introduces the category imbalance processing module, the attention mechanism, and the context encoding module, and progressively analyzes the magnitude of the enhancement of the classification accuracy by each module to quantify its practical effects on sample enhancement, feature extraction and behavior modeling.
  • Impact of category imbalance processing on experimental results
As shown in Table 6, in the base model, all the enhancement modules were removed and only the most basic convolutional structure with node input features were retained; the classification accuracy was 90.31%. A class imbalance processing module based on a generative adversarial network (GAN) was introduced to generate synthetic fraudulent transaction samples and extend the size of the few class samples in the training set. The module learns the distribution characteristics of real fraud samples through adversarial training, thus improving its ability to discern the minority class. Experimental results show that the introduction of this module improved the accuracy rate from 90.31% to 94.97%, with an improvement of 4.66 percentage points, which verifies its significant role in mitigating the skewness of the category distribution and improving the robustness of the model.
2.
Impact of Attention Mechanism on Experimental Results
Based on the introduction of category imbalance processing, the lightweight multi-scale attention mechanism module was further integrated. This module is modeled through the synergy of channel attention and spatial attention so that the model can dynamically pay attention to the nodes and edges in the transaction graph that have strong correlation with fraudulent behaviors, which enhances the ability to perceive both local and global anomalous patterns in the structure of the transaction graph. The experimental results are shown in Table 7; 91.25% to 94.97% is an increase of 3.72 percentage points, indicates that the mechanism has a strong effect in improving feature discrimination and capturing potential risk paths.
3.
Impact of Context Encoding Module on Experimental Results
Based on the previous two modules, the context encoding module is further introduced to fully explore the trading behavior characteristics of accounts in the time dimension. This module integrates contextual information such as transaction frequency, active period, and out-degree of nodes, constructs temporal embedding, and participates in the generation process of the final node representation, to enhance the model’s ability to model dynamic behavioral patterns. The experimental results are shown in Table 8; the accuracy rate increases from 92.31% to 94.97%, an increase of 2.66 percentage points, indicating that the introduction of contextual information can provide more comprehensive behavioral descriptions for the model and further enhance its comprehensive discriminative performance.
4.
The Impact of Symmetry on Experimental Results
Experimental results from the YelpChi dataset demonstrate that the symmetrical structures significantly enhance the model’s performance. As shown in Table 9, incorporating the symmetry module into the GAN-CNN detection framework improved the model accuracy from 92.16% to 94.97%, an increase of 2.81 percentage points, indicating that symmetric structures play a crucial role in enhancing feature extraction efficiency and pattern recognition capabilities. Additionally, combining the symmetry module with a weight sharing strategy further optimizes the network structure, reducing parameter redundancy while maintaining stable model performance. Without the symmetry and weight-sharing modules, the accuracy rates were 92.16% and 93.78%, respectively, both lower than the 94.97% achieved with the complete structure. This validates the significant value of symmetry design in constructing lightweight and efficient financial fraud detection models.

5. Conclusions

In this paper, a lightweight financial fraud detection model based on a symmetrical GAN-CNN structure is proposed to effectively deal with the problems of category imbalance, insufficient contextual feature extraction, and high computational complexity. The model includes a GAN to generate fraud samples to alleviate sample imbalance, a CNN to extract temporal and graph structure features, and an integrated lightweight multi-scale attention mechanism and adaptive aggregation module to improve feature expression capability and computational efficiency. The experiments were conducted on YelpChi and Amazon datasets, and the results showed that the model outperformed mainstream methods such as FdGars, GraphSAINT, and HHLN-GNN in terms of accuracy, parameter size, and GMACs, with good detection performance and application potential.
Although this achieved some success, it may have certain shortcomings in terms of accuracy when it comes to data compression and extreme class imbalance. In the future, we will combine multimodal data and sample generation to solve the above problems and increase the transparency of the research model.

Author Contributions

Conceptualization, Y.Y., C.X. and G.T.; methodology, Y.Y. and C.X.; software, C.X.; validation, Y.Y., C.X. and G.T.; formal analysis, Y.Y. and C.X.; investigation, G.T.; resources, C.X.; data curation, Y.Y.; writing—original draft preparation, Y.Y.; writing—review and editing, Y.Y. and C.X.; visualization, Y.Y., C.X. and G.T.; supervision, C.X.; project administration, Y.Y., C.X. and G.T.; funding acquisition, Y.Y. and G.T. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Natural Science Foundation of China (42261068), the Natural Science Foundation of Jiangxi Province (20242BAB25112), Demonstration Course on Curriculum Civics and Politics in Undergraduate Colleges and Universities in Henan Province in 2024“ International Trade Practice” (185), the Soft Science Research Project in Henan Province in 2025 (252400410103), Henan Province Online First-Class Undergraduate Course “International Trade Theory and Practice” (67), New Round of Key Discipline “Applied Economics” in Henan Province (340), and Funding Program for Specialty Construction of Privately Run General Colleges and Universities (Economics) in Henan Province in 2023.

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

AR-GCNAdaptive receptive field graph convolutional network
BNBatch normalization
CatBoostCategorical boosting
CMConfusion matrix
CNNConvolutional neural network
ConvConvolution
DWConvDepthwise separable convolution
FAGCNFrequency adaptive graph convolutional network
FNFalse negative
FPFalse positive
FRFNFeature refined fusion network
GANGenerative adversarial network
GATGraph attention network
GCNGraph convolutional network
LNLayer normalization
OAOverall aAccuracy
SeLUScaled exponential linear unit
SGCSimplifying graph convolutional network
SMOTESynthetic minority over-sampling technique
TFETemporal frequency encoding
TNTrue negative
TPTrue positive
XGBoostExtreme Gradient Boosting

References

  1. Tong, G.; Shen, J. Financial Transaction Fraud Detector Based on Imbalance Learning and Graph Neural Network. Appl. Soft Comput. 2023, 149, 18387–18399. [Google Scholar] [CrossRef]
  2. Lin, A.K. The AI Revolution in Financial Services: Emerging Methods for Fraud Detection and Prevention. J. Galaksi. 2024, 1, 43–51. [Google Scholar] [CrossRef]
  3. Devi, J.V.; Kavitha, K.S. Fraud detection in credit card transactions by using classification algorithms. In Proceedings of the 2017 International Conference on Current Trends in Computer, Electrical, Electronics and Communication (CTCEEC), Mysore, India, 8–9 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 125–131. [Google Scholar] [CrossRef]
  4. Liu, G. Leveraging Machine Learning for Telecom Banking Card Fraud Detection: A Comparative Analysis of Logistic Regression, Random Forest, and XGBoost Models. Comput. Artif. Intell. 2024, 1, 13–27. [Google Scholar] [CrossRef]
  5. Ali, A.; Abd Razak, S.; Othman, S.H.; Eisa, T.A.E.; Al-Dhaqm, A.; Nasser, M.; Elha, T.; Elshafie, H.; Saif, A. Financial Fraud Detection Based on Machine Learning: A Systematic Literature Review. Appl. Sci. 2022, 12, 9637. [Google Scholar] [CrossRef]
  6. Wu, G.; He, X.; Li, M. Credit Evaluation Based on Improved Naive Bayesian Model. Adv. Appl. Math. 2019, 8, 1410–1417. [Google Scholar] [CrossRef]
  7. Thilagavathi, M.; Saranyadevi, R.; Vijayakumar, N.; Selvi, K.; Anitha, L.; Sudharson, K. AI-Driven Fraud Detection in Financial Transactions with Graph Neural Networks and Anomaly Detection. In Proceedings of the 2024 IEEE International Conference on Science Technology Engineering and Management (ICSTEM), Coimbatore, India, 26–27 April 2024; pp. 1–6. [Google Scholar] [CrossRef]
  8. Li, S.; Zhou, J.; Mo, C.; Li, J.; Tso, G.K.; Tian, Y. Motif-Aware Temporal GCN for Fraud Detection in Signed Cryptocurrency Trust Networks. arXiv 2022, arXiv:2211.13123. [Google Scholar] [CrossRef]
  9. Nalayini, C.M.; Katiravan, J.; Sathyabama, A.R.; Rajasuganya, P.V.; Abirami, K. Identification and Detection of Credit Card Frauds Using CNN. In Proceedings of the International Conference on Computers, Management & Mathematical Sciences, Nirjuli, India, 29–30 July 2022; Springer International Publishing: Cham, Switzerland, 2022; pp. 267–280. [Google Scholar] [CrossRef]
  10. Wu, R. Financial Fraud Detection System Combining Generative Adversarial Networks and Deep Learning. In Proceedings of the 2024 International Conference on Industrial IoT, Big Data and Supply Chain (IIoTBDSC), Wuhan, China, 20–22 September 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 105–110. [Google Scholar] [CrossRef]
  11. Fletcher, G.; Shi, T. Financial Fraud Detection with Self-Attention Mechanism: A Comparative Study. J. Comput. Sci. Softw. Appl. 2025, 5, 10–18. [Google Scholar] [CrossRef]
  12. Zhao, C.; Sun, X.; Wu, M.; Kang, L. Advancing Financial Fraud Detection: Self-Attention Generative Adversarial Networks for Precise and Effective Identification. Finance Res. Lett. 2024, 60, 104843. [Google Scholar] [CrossRef]
  13. Gupta, A.; Lohani, M.C.; Manchanda, M. Financial Fraud Detection Using Naive Bayes Algorithm in Highly Imbalance Data Set. J. Discret. Math. Sci. Cryptogr. 2021, 24, 1559–1572. [Google Scholar] [CrossRef]
  14. Shaohui, D.; Qiu, G.; Mai, H.; Yu, H. Customer Transaction Fraud Detection Using Random Forest. In Proceedings of the 2021 IEEE International Conference on Consumer Electronics and Computer Engineering (ICCECE), Guangzhou, China, 15–17 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 144–147. [Google Scholar] [CrossRef]
  15. Wang, T.; Zhao, Y. Credit Card Fraud Detection Using Logistic Regression. In Proceedings of the 2022 International Conference on Big Data, Information and Computer Network (BDICN), Sanya, China, 20–22 January 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 301–305. [Google Scholar] [CrossRef]
  16. Aburbeian, A.M.; Ashqar, H.I. Credit Card Fraud Detection Using Enhanced Random Forest Classifier for Imbalanced Data. In Proceedings of the International Conference on Advances in Computing Research, Orlando, FL, USA, 8–10 May 2023; Springer Nature: Cham, Switzerland, 2023; pp. 605–616. [Google Scholar] [CrossRef]
  17. Afriyie, J.K.; Tawiah, K.; Pels, W.A.; Addai-Henne, S.; Dwamena, H.A.; Owiredu, E.O.; Ayeh, A.; Eshun, J. A Supervised Machine Learning Algorithm for Detecting and Predicting Fraud in Credit Card Transactions. Decis. Anal. J. 2023, 6, 100163. [Google Scholar] [CrossRef]
  18. Li, C.; Geng, Y. Research on Credit Scoring Based on BLS-Classifier. Finance 2020, 10, 548–559. [Google Scholar] [CrossRef]
  19. Zhu, X.; Ao, X.; Qin, Z.; Chang, Y.; Liu, Y.; He, Q.; Li, J. Intelligent Financial Fraud Detection Practices in Post-Pandemic Era. Innovation 2021, 2, 18387–18399. [Google Scholar] [CrossRef] [PubMed]
  20. Kumar, S.; Gunjan, V.K.; Ansari, M.D.; Pathak, R. Credit card fraud detection using support vector machine. In Proceedings of the 2nd International Conference on Recent Trends in Machine Learning, IoT, Smart Cities and Applications: ICMISC 2021, Hyderabad, India, 28–29 March 2021; Springer: Singapore, 2021; pp. 27–37. [Google Scholar] [CrossRef]
  21. Obeng, S.; Iyelolu, T.V.; Akinsulire, A.A.; Idemudia, C. Utilizing Machine Learning Algorithms to Prevent Financial Fraud and Ensure Transaction Security. World J. Adv. Res. Rev. 2024, 23, 1972–1980. [Google Scholar] [CrossRef]
  22. Ahmad, N. Deep Learning for Fraud Detection in Financial Transactions: A Novel Approach to Detect Hidden Anomalies. MCS 2024. [Google Scholar]
  23. Cheng, Y.; Guo, J.; Long, S.; Wu, Y.; Sun, M.; Zhang, R. Advanced Financial Fraud Detection Using GNN-CL Model. In Proceedings of the 2024 International Conference on Computers, Information Processing and Advanced Education (CIPAE), Ottawa, ON, Canada, 26–28 August 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 453–460. [Google Scholar] [CrossRef]
  24. Yadegar, M.M.; Rahmani, H. FinFD-GCN: Using Graph Convolutional Networks for Fraud Detection in Financial Data. J. Artif. Intell. Data Min. 2024, 12, 487–495. [Google Scholar] [CrossRef]
  25. Khosravi, S.; Kargari, M.; Teimourpour, B.; Talebi, M. Transaction Fraud Detection via Attentional Spatial–Temporal GNN. J. Supercomput. 2025, 81, 537. [Google Scholar] [CrossRef]
  26. Wei, S.; Lee, S. Financial Anti-Fraud Based on Dual-Channel Graph Attention Network. J. Theor. Appl. Electron. Commer. Res. 2024, 19, 297–314. [Google Scholar] [CrossRef]
  27. Zhang, G.; Li, Z.; Huang, J.; Wu, J.; Zhou, C.; Yang, J.; Gao, J. eFraudCom: An E-Commerce Fraud Detection System via Competitive Graph Neural Networks. ACM Trans. Inf. Syst. (TOIS) 2022, 40, 1–29. [Google Scholar] [CrossRef]
  28. Karthika, J.; Senthilselvi, A. Smart Credit Card Fraud Detection System Based on Dilated Convolutional Neural Network with Sampling Technique. Multimed. Tools Appl. 2023, 82, 31691–31708. [Google Scholar] [CrossRef]
  29. Cheah, P.C.Y.; Yang, Y.; Lee, B.G. Enhancing Financial Fraud Detection Through Addressing Class Imbalance Using Hybrid SMOTE-GAN Techniques. Int. J. Financ. Stud. 2023, 11, 110. [Google Scholar] [CrossRef]
  30. Strelcenia, E.; Prakoonwit, S. A Survey on GAN Techniques for Data Augmentation to Address the Imbalanced Data Issues in Credit Card Fraud Detection. Mach. Learn. Knowl. Extr. 2023, 5, 304–329. [Google Scholar] [CrossRef]
  31. Li, R.; Liu, Z.; Ma, Y.; Yang, D.; Sun, S. Internet Financial Fraud Detection Based on Graph Learning. IEEE Trans. Comput. Soc. Syst. 2022, 10, 1394–1401. [Google Scholar] [CrossRef]
  32. Duan, Y.; Zhang, G.; Wang, S.; Peng, X.; Ziqi, W.; Mao, J.; Wu, H.; Jiang, X.; Wang, K. Cat-GNN: Enhancing Credit Card Fraud Detection via Causal Temporal Graph Neural Networks. arXiv 2024, arXiv:2402.14708. [Google Scholar]
  33. Xie, Y.; Liu, G.; Zhou, M.; Wei, L.; Zhu, H.; Zhou, R.; Cao, L. A Spatial–Temporal Gated Network for Credit Card Fraud Detection by Learning Transactional Representations. IEEE Trans. Autom. Sci. Eng. 2023, 21, 6978–6991. [Google Scholar] [CrossRef]
  34. Yu, C.; Xu, Y.; Cao, J.; Zhang, Y.; Jin, Y.; Zhu, M. Credit Card Fraud Detection Using Advanced Transformer Model. In Proceedings of the 2024 IEEE International Conference on Metaverse Computing, Networking, and Applications (MetaCom), Hong Kong, China, 12–14 August 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 343–350. [Google Scholar] [CrossRef]
  35. Ghosh, C.; Chowdhury, A.; Das, N.; Sadhukhan, B. Enhancing Financial Fraud Detection in Bitcoin Networks Using Ensemble Deep Learning. In Proceedings of the 2023 IEEE International Conference on Blockchain and Distributed Systems Security (ICBDS), New Raipur, India, 6–8 October 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar] [CrossRef]
  36. Li, M.; Zhang, Z.; Lu, M.; Jia, X.; Liu, R.; Zhou, X.; Zhang, Y. Internet Financial Credit Risk Assessment with Sliding Window and Attention Mechanism LSTM Model. Teh. Vjesn. 2023, 30, 1–7. [Google Scholar] [CrossRef]
  37. Liu, H.L.; Cao, S.J.; Xu, J.Y.; Chen, S.H. Anti-Fraud Research Advances on Digital Credit Payment. J. Front. Comput. Sci. Technol. 2023, 17, 2300–2324. [Google Scholar] [CrossRef]
  38. Hong, B.; Lu, P.; Chen, R.; Lin, K.; Yang, F. Health Insurance Fraud Detection via Multiview Heterogeneous Information Networks with Augmented Graph Structure Learning. IEEE Trans. Comput. Soc. Syst. 2024, 17, 1–21. [Google Scholar] [CrossRef]
  39. Lou, C.; Wang, Y.; Li, J.; Qian, Y.; Li, X. Graph Neural Network for Fraud Detection via Context Encoding and Adaptive Aggregation. Expert Syst. Appl. 2025, 261, 125473. [Google Scholar] [CrossRef]
  40. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  41. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar] [CrossRef]
  42. Wu, F.; Souza, A.; Zhang, T.; Fifty, C.; Yu, T.; Weinberger, K.Q. Simplifying Graph Convolutional Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6861–6871. [Google Scholar]
  43. Ma, X.; Li, Z.; Xu, L.; Song, G.; Li, Y.; Shi, C. Learning Discrete Adaptive Receptive Fields for Graph Convolutional Networks. Sci. China Inf. Sci. 2023, 66, 222101. [Google Scholar] [CrossRef]
  44. Bo, D.; Wang, X.; Shi, C.; Shen, H. Beyond Low-Frequency Information in Graph Convolutional Networks. Proc. AAAI Conf. Artif. Intell. 2021, 35, 3950–3957. [Google Scholar] [CrossRef]
  45. Hiremath, A.C.; Arya, A.; Sriranga, L.; Reddy, K.V.S.R.; Nikhil, M. Ensemble of Graph Neural Networks for Enhanced Financial Fraud Detection. In Proceedings of the 2024 IEEE 9th International Conference for Convergence in Technology (I2CT), Pune, India, 5 April 2024; pp. 1–8. [Google Scholar] [CrossRef]
  46. Ning, Z.; Wang, Z.; Zhang, R.; Xu, P.; Liu, K.; Wang, P.; Ju, W.; Wang, P.; Zhou, Y.; Cambria, E. Deep Cut-Informed Graph Embedding and Clustering. arXiv 2024, arXiv:2503.06635. [Google Scholar] [CrossRef]
  47. Kanezashi, H.; Suzumura, T.; Liu, X.; Hirofuchi, T. Ethereum Fraud Detection with Heterogeneous Graph Neural Networks. arXiv 2022, arXiv:2203.12363. [Google Scholar] [CrossRef]
  48. Zeng, H.; Zhou, H.; Srivastava, A.; Kannan, R.; Prasanna, V. GraphSAINT: Graph Sampling-Based Inductive Learning Method. arXiv 2019, arXiv:1907.04931. [Google Scholar] [CrossRef]
  49. Chen, Z.; Zhu, M.; Du, J.W. Multi-View Graph Neural Network for Fraud Detection Algorithm. J. Commun./Tongxin Xuebao 2022, 43, 225–232. [Google Scholar] [CrossRef]
  50. Dou, Y.; Liu, Z.; Sun, L.; Deng, Y.; Peng, H.; Yu, P.S. Enhancing Graph Neural Network-Based Fraud Detectors against Camouflaged Fraudsters. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Beijing, China, 3–7 November 2020; pp. 315–324. [Google Scholar] [CrossRef]
  51. Singh, A.; Tanwar, S. Revolutionizing Online Transaction Safety with CNN and GAN-Based Fraud Detection Strategies. In Proceedings of the 2024 Asia Pacific Conference on Innovation in Technology (APCIT), Mysore, India, 26–27 July 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–4. [Google Scholar] [CrossRef]
  52. Akour, I.; Mohamed, N.; Salloum, S. Hybrid CNN-LSTM with Attention Mechanism for Robust Credit Card Fraud Detection. IEEE Access 2025, 13, 114056–114068. [Google Scholar] [CrossRef]
  53. Gupta, V.; Mishra, N.; Dash, Y.; Kumar, U.; Abraham, A. Graph Convolutional Network-Driven Adaptive Learning Framework for Fraud Detection in Complex Transactional Cryptonetworks. In Proceedings of the 2025 3rd International Conference on Communication, Security, and Artificial Intelligence (ICCSAI), Greater Noida, India, 4–6 April 2025; IEEE: Piscataway, NJ, USA, 2025; Volume 3, pp. 685–689. [Google Scholar] [CrossRef]
  54. Markova, M. A Comparison of Deep Neural Network and Convolutional Neural Network for Credit Card Fraud Detection. In Proceedings of the International Conference on New Trends in the Applications of Differential Equations in Sciences, Cham, Switzerland, 7–10 July 2024; Springer Nature Switzerland: Cham, Switzerland, 2024; pp. 405–418. [Google Scholar]
Figure 1. Symmetrical GAN-CNN Fusion Architecture.
Figure 1. Symmetrical GAN-CNN Fusion Architecture.
Symmetry 17 01366 g001
Figure 2. Category Imbalance Processing Module.
Figure 2. Category Imbalance Processing Module.
Symmetry 17 01366 g002
Figure 3. Adaptive aggregation module.
Figure 3. Adaptive aggregation module.
Symmetry 17 01366 g003
Figure 4. Lightweight multi-scale attention mechanism.
Figure 4. Lightweight multi-scale attention mechanism.
Symmetry 17 01366 g004
Figure 5. Feature interaction fusion.
Figure 5. Feature interaction fusion.
Symmetry 17 01366 g005
Figure 6. Context encoding module.
Figure 6. Context encoding module.
Symmetry 17 01366 g006
Table 1. Parameter Comparison between Ordinary Convolution and Depthwise Separable Convolution.
Table 1. Parameter Comparison between Ordinary Convolution and Depthwise Separable Convolution.
MethodsKernel SizeInput ChannelOutput ChannelLayerParametersTotal (M)
Ordinary3 × 310241024 Conv 1 1024 × 1024 × 3 × 3 = 9,437,184≈28.3 M
Conv 2 1024 × 1024 × 3 × 3 = 9,437,184
Conv 3 1024 × 1024 × 3 × 3 = 9,437,184
5 × 510241024 Conv 1 1024 × 1024 × 5 × 5 = 26,214,400≈78.6 M
Conv 2 1024 × 1024 × 5 × 5 = 26,214,400
Conv 3 1024 × 1024 × 5 × 5 = 26,214,400
Depthwise Separable3 × 310241024 Conv 1 1024 × 1024 × 3 × 3 = 9,437,184≈9.4 M
Conv 2
Conv 3
5 × 510241024 Conv 1 1024 × 1024 × 5 × 5 = 26,214,400≈26.2 M
Conv 2
Conv 3
Table 2. Datasets.
Table 2. Datasets.
DatasetYelpChiAmazon
Nodes45,95411,944
edges3,846,9794,398,392
(fraud%)14.5%10%
RelationR-U-R, R-S-R, R-T-RU-P-U, U-S-U, U-V-U
Table 3. Setting of experimental environment and other parameters.
Table 3. Setting of experimental environment and other parameters.
ProjectContent
CPUIntel Core i7-4700, 2.70 GHz × 12
Memory32 GB
Operating systemCentOS 7.8 64-bit
Hard disk1TB
GPUNvidia Titan-X × 2
Python3.7.2
PyTorch1.4.0
CUDA10.0
Learning rate10–3
Momentum0.73
Weight decay5 × 10–4
Batch16
Saturation1.7
Subdivisions64
Table 4. Classification accuracy (%) of each model on YelpChi vs. Amazon datasets with different training ratios.
Table 4. Classification accuracy (%) of each model on YelpChi vs. Amazon datasets with different training ratios.
MethodYelpChiAmazon
20%50%50%
XGBoost [40]85.47 ± 0.2887.92 ± 0.2988.79 ± 0.41
CatBoost [40]92.11 ± 0.2195.09 ± 0.2293.49 ± 0.29
GCN [41]87.42 ± 0.2791.04 ± 0.2194.81 ± 0.22
SGC [42]85.95 ± 0.2290.11 ± 0.2596.72 ± 0.12
GAT [41]91.28 ± 0.1594.77 ± 0.1293.52 ± 0.34
AR-GCN [43]91.82 ± 0.1595.02 ± 0.1996.42 ± 0.39
FAGCN [44]91.44 ± 0.1494.82 ± 0.1792.87 ± 0.28
FdGars [45]92.15 ± 0.1994.92 ± 0.1496.44 ± 0.52
GEM [46]91.79 ± 0.1992.95 ± 0.2493.96 ± 0.34
GraphSAGE [47]88.94 ± 0.2191.45 ± 0.2791.46 ± 0.27
GraphSAINT [48]92.48 ± 0.1595.77 ± 0.2196.77 ± 0.31
GraphConsis [49]91.22 ± 0.2294.84 ± 0.0994.84 ± 0.09
CARE-GNN [50]91.47 ± 0.2894.52 ± 0.1594.63 ± 0.16
HHLN-GNN [1]94.82 ± 0.1495.97 ± 0.2194.92 ± 0.18
GCN+SMOTE [51]93.67 ± 0.1596.86 ± 0.1294.79 ± 0.15
CNN+SMOTE [52]93.58 ± 0.1796.78 ± 0.1394.68 ± 0.22
CNN-GAN [53]94.52 ± 0.2294.52±0.2294.73±0.16
Hybrid
CNN-LSTM–attention [54]
94.75 ± 0.1396.89 ± 0.1595.67 ± 0.23
Proposed94.97 ± 0.2297.15 ± 0.2596.37 ± 0.23
Table 5. Model size evaluation.
Table 5. Model size evaluation.
MethodAcc (%)Parameters (M)GMACs
(G)
PrecisionRecallF1
XGBoost [40]85.471.20.01
CatBoost [40]92.111.50.02
GCN [41]87.422.80.20.52110.5533
SGC [42]85.9520.25
GAT [41]91.282.50.50.53860.4649
AR-GCN [43]91.824.20.55
FAGCN [44]91.444.50.5
FdGars [45]92.154.80.55
GEM [46]91.7950.7
GraphSAGE [47]88.942.20.40.52660.5471
GraphSAINT [48]92.482.80.45
GraphConsis [49]91.2240.480.62080.6070
CARE-GNN [50]91.475.10.750.70380.6138
HHLN-GNN [1]94.825.50.80.9680.7230.812
GCN+SMOTE [51]93.6736.41.630.9450.7050.782
CNN+SMOTE [52]93.5849.232.790.9370.6830.776
CNN-GAN [53]94.5242.372.560.9520.7100.791
Hybrid
CNN-LSTM–attention [54]
94.7539.651.870.9610.7110.807
Proposed94.974.920.620.9710.7670.823
Table 6. Effect of category imbalance treatment on experimental results on YelpChi dataset.
Table 6. Effect of category imbalance treatment on experimental results on YelpChi dataset.
Module ConfigurationAccuracy (%)
Baseline (total removal)90.31
+Category imbalance processing (GAN sample generation)94.97
Table 7. Effect of Attention Mechanisms on Experimental Results on the YelpChi Dataset.
Table 7. Effect of Attention Mechanisms on Experimental Results on the YelpChi Dataset.
Module ConfigurationAccuracy (%)
+Category imbalance processing (GAN sample generation)91.25
+Category Imbalance Processing + Attention Mechanism94.97
Table 8. Effect of context encoding module on experimental results on YelpChi dataset.
Table 8. Effect of context encoding module on experimental results on YelpChi dataset.
Module ConfigurationAccuracy (%)
+Category Imbalance Handling + Attention Mechanism92.31
+Category Imbalance Handling + Attention Mechanism + Context Encoding94.97
Table 9. The impact of symmetry on experimental results from the YelpChi dataset.
Table 9. The impact of symmetry on experimental results from the YelpChi dataset.
Module ConfigurationAccuracy (%)
With symmetry module94.97
Without symmetry module92.16
Use weight sharing94.97
Remove weight sharing93.78
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Xu, C.; Tian, G. Lightweight Financial Fraud Detection Using a Symmetrical GAN-CNN Fusion Architecture. Symmetry 2025, 17, 1366. https://doi.org/10.3390/sym17081366

AMA Style

Yang Y, Xu C, Tian G. Lightweight Financial Fraud Detection Using a Symmetrical GAN-CNN Fusion Architecture. Symmetry. 2025; 17(8):1366. https://doi.org/10.3390/sym17081366

Chicago/Turabian Style

Yang, Yiwen, Chengjun Xu, and Guisheng Tian. 2025. "Lightweight Financial Fraud Detection Using a Symmetrical GAN-CNN Fusion Architecture" Symmetry 17, no. 8: 1366. https://doi.org/10.3390/sym17081366

APA Style

Yang, Y., Xu, C., & Tian, G. (2025). Lightweight Financial Fraud Detection Using a Symmetrical GAN-CNN Fusion Architecture. Symmetry, 17(8), 1366. https://doi.org/10.3390/sym17081366

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop