Next Article in Journal
Killing Vector Fields of Invariant Metrics on Five-Dimensional Solvable Lie Groups
Previous Article in Journal
Two-Stage Analysis for Supply Chain Disruptions Considering the Trade-Off Between Profit Maximization and Adaptability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Financial Fraud Detection via Causal Intervention and Multi-View Contrastive Learning on Dynamic Hypergraphs

Department of Information Technology, Uppsala University, 752 37 Uppsala, Sweden
Mathematics 2025, 13(24), 4018; https://doi.org/10.3390/math13244018
Submission received: 2 December 2025 / Revised: 13 December 2025 / Accepted: 14 December 2025 / Published: 17 December 2025

Abstract

Financial fraud detection is critical to modern economic security, yet remains challenging due to collusive group behavior, temporal drift, and severe class imbalance. Most existing graph neural network (GNN) detectors rely on pairwise edges and correlation-driven learning, which limits their ability to represent high-order group interactions and makes them vulnerable to spurious environmental cues (e.g., hubs or temporal bursts) that correlate with labels but are not necessarily causal. We propose Causal-DHG, a dynamic hypergraph framework that integrates hypergraph modeling, causal intervention, and multi-view contrastive learning. First, we construct label-agnostic hyperedges from publicly available metadata to capture high-order group structures. Second, a Multi-Head Spatio-Temporal Hypergraph Attention encoder models group-wise dependencies and their temporal evolution. Third, a Causal Disentanglement Module decomposes representations into causal and environment-related factors using HSIC regularization, and a dictionary-based backdoor adjustment approximates the interventional prediction P ( Y d o ( C ) ) to suppress spurious correlations. Finally, we employ self-supervised multi-view contrastive learning with mild hypergraph augmentations to leverage unlabeled data and stabilize training. Experiments on YelpChi, Amazon, and DGraph-Fin show consistent gains in AUC/F1 over strong baselines such as CARE-GNN and PC-GNN, together with improved robustness under feature and structural perturbations.

1. Introduction

The rapid growth of digital financial services has reshaped global markets, enabling instant credit approvals, online lending, and cross-border payments, but also enlarging the attack surface for fraud. Cases now range from synthetic identity lending and collusive rating manipulation to contact-circle abuse and credit default schemes, with annual losses estimated in the hundreds of billions of dollars [1]. Modern financial fraud typically involves three coupled characteristics: group collusion, where coordinated rings operate across accounts, reviews, or contact profiles; dynamic drift, as fraud patterns respond to deployed models and induce temporal distribution shifts [2]; and camouflage, where fraudulent accounts mimic benign behavior statistics to evade detection [3]. These factors make purely rule-based and i.i.d. learning approaches inadequate.
Graph neural networks (GNNs) are a natural tool for fraud detection because transaction logs, review graphs, and emergency-contact networks are inherently relational. By modeling entities (e.g., users or reviews) as nodes and interactions as edges, GNN-based methods can exploit both local and global structures. Fraud-oriented architectures such as SemiGNN, GeniePath, CARE-GNN, and PC-GNN [3,4,5,6] address heterogeneity, class imbalance, and camouflage and have shown strong performance on public benchmarks and industrial platforms.
However, GNN-based fraud detectors still face two core limitations in realistic deployments. First, they are built on simple graphs, where edges represent pairwise relations. Real fraud often appears as high-order group patterns: multiple suspicious reviews tied to the same user and product, clusters of users sharing contact behavior within short time windows, or dense circles of mutually referencing accounts. Naively decomposing such structures into pairwise edges causes the “clique expansion” problem in hypergraph theory [7], inflating edges and obscuring the semantics of joint co-participation. Second, most models are correlation-driven and suffer from a causal gap. Deep networks tend to overfit environmental context—anonymous profile statistics, temporal bursts, or hub nodes—that correlate with labels in the training environment but do not reflect genuine fraudulent intent [8,9,10]. When the environment changes, performance can drop sharply.
Figure 1 illustrates this causal gap. In practice, fraud labels are often statistically associated with environmental properties that are not causal mechanisms. For example, users attached to a particular hub or concentrated in certain time windows may be disproportionately fraudulent in the training set due to historical enforcement or platform policies. A correlation-driven GNN treats such patterns as strong predictors; once benign users exhibit similar structures, the model degrades. Bridging this gap requires representations that separate intrinsic signals of fraudulent intent from an environment-dependent context.
We propose Causal-DHG, a dynamic hypergraph neural network that integrates causal intervention and contrastive learning for robust financial fraud detection. The framework is instantiated on three widely used public datasets (YelpChi, Amazon, DGraph-Fin). Our main contributions are as follows:
  • A causality-aware dynamic hypergraph framework that captures high-order group interactions in realistic financial graphs (review networks and contact networks), with hypergraphs constructed directly from observed structures and attributes in YelpChi, Amazon, and DGraph-Fin.
  • A Causal Disentanglement Module that decomposes node representations into intrinsic causal and environment-related factors, enforces their independence via HSIC, and applies a dictionary-based backdoor adjustment mechanism to approximate the interventional distribution P ( Y d o ( C ) ) and suppress spurious environmental correlations.
  • A multi-view hypergraph contrastive learning scheme that exploits unlabeled data under label scarcity, using edge dropping and feature masking together with an InfoNCE-style objective defined on disentangled representations.
  • A theoretical characterization of the causal graph underlying Causal-DHG, explaining how HSIC-based disentanglement and dictionary-based backdoor adjustment approximate environment-marginalized prediction and analyze robustness and complexity under environment shifts.
  • Extensive experiments on YelpChi, Amazon, and DGraph-Fin show consistent improvements over strong baselines such as CARE-GNN and PC-GNN (up to about four percentage points in AUC) and demonstrate that the causal module substantially mitigates performance degradation under feature perturbations and structural noise.

2. Related Work

2.1. Graph Neural Networks for Fraud Detection

The relational nature of financial data makes GNNs a natural tool for fraud detection. Early work applied general-purpose graph models, such as GraphSAGE [11], to transaction and review graphs by treating entities as nodes and interactions as edges. These methods aggregate neighbor information through message passing and provide strong baselines, but they are not specifically tailored to severe class imbalance, structured camouflage, or heterogeneous relations.
Several GNN variants have been developed for fraud detection. SemiGNN [4] uses hierarchical attention to integrate user, product, and review views; GeniePath [5] learns adaptive receptive fields to filter noisy neighbors; CARE-GNN [3] introduces camouflage-resistant neighbor selection driven by reinforcement learning; and PC-GNN [6] combines pick-and-choose sampling with graph augmentations to handle imbalanced, camouflaged fraud. These models demonstrate the benefit of domain-specific GNN design but remain pairwise and correlation-driven, without explicit treatment of high-order interactions or causal mechanisms, making them sensitive to distribution shifts [10].
Beyond GNN-based architectures, matrix factorization and subspace learning techniques have also been explored for detecting abnormal behavior in structured data. For example, graph-regularized orthogonal non-negative matrix factorization with Itakura–Saito divergence has been applied to fault detection problems on graph-structured signals [12], providing a complementary perspective to graph neural models.

2.2. Hypergraph Learning and Dynamic Modeling

Hypergraphs generalize graphs by allowing a hyperedge to connect an arbitrary number of nodes, providing a natural representation for group-level interactions [13]. HGNN [7] extends spectral convolutions to hypergraphs via a hypergraph Laplacian, and subsequent work has improved efficiency and flexibility through low-rank approximations and attention mechanisms. Dynamic hypergraph models, such as DHGNN [14], further incorporate temporal evolution by updating node and hyperedge states over time. In parallel, temporal graph models like EvolveGCN and TGN [15,16] have shown that explicitly modeling time improves prediction on evolving networks.
These ideas have been explored in adjacent areas such as session-based recommendation, where hyperedges group items within a session [17]. However, applications of dynamic hypergraphs to financial fraud remain limited, and existing models typically treat learning as purely correlation-based. They capture high-order and temporal structures but do not aim to separate causal and environment-related factors or to implement principled intervention.

2.3. Causal Inference on Graphs

Causal inference provides a framework for distinguishing correlation from causation [9] and has recently attracted interest in representation learning and graph modeling [18]. On graphs, several works propose stable or invariant objectives to mitigate biases introduced by structural properties and environment shifts [10,19]. Disentangled graph representation methods, such as DisenGCN and related models, seek to factorize node embeddings into relatively independent components, improving interpretability and, in some cases, robustness.
Most existing approaches, however, either operate on static pairwise graphs or target graph-level tasks. They do not jointly model dynamic high-order interactions and node-level fraud labels, and they rarely provide an explicit mechanism to approximate the interventional distribution P ( Y d o ( C ) ) . In contrast, Causal-DHG is built on dynamic hypergraphs, explicitly disentangles intrinsic causal and environment-related factors, and combines dictionary-based backdoor adjustment with contrastive learning to improve robustness under temporal and structural shifts in fraud detection.

3. Methodology: The Causal-DHG Framework

In this section, we present the proposed Causal-DHG framework in detail. We first describe the unified dynamic hypergraph construction that is carefully instantiated for the three public datasets. We then introduce the Multi-Head Spatio-Temporal Hypergraph Attention (MH-DHA) encoder, followed by the Causal Disentanglement and Backdoor Adjustment module and the multi-view contrastive objective (Figure 2).

3.1. Dynamic Hypergraph Construction

We start from the released graphs 𝒢 = ( 𝒱 , E ) with node features X R | 𝒱 | × d x . In YelpChi, nodes are review instances and edges encode shared user/product and similarity relations; in Amazon, nodes are user accounts with edges induced by overlapping reviewing behaviors; in DGraph-Fin, nodes are users linked by directed emergency-contact edges annotated with timestamps and contact types. Node features are the provided handcrafted statistics (32/25/17 dimensions for YelpChi/Amazon/DGraph-Fin), and labels indicate fraudulent vs. normal entities.
To capture temporal evolution, we discretize interactions into T time windows. For YelpChi and Amazon, which are commonly used as static benchmarks, we treat the provided graphs as single snapshots ( T = 1 ) unless otherwise stated. For DGraph-Fin, edge timestamps are grouped into T non-overlapping intervals (e.g., by quantiles or fixed-size buckets), yielding a sequence of time-indexed graphs { 𝒢 1 , , 𝒢 T } .
For each window t, we build a hypergraph H t = ( 𝒱 , E t H ) whose hyperedges connect nodes that participate in observed group patterns, using only publicly available structures and attributes:
  • YelpChi (review nodes).
    (i) User-centric: For each user u, a hyperedge e u connects all reviews written by u. (ii) Product-centric: For each product p, a hyperedge e p connects all reviews of p in the window. (iii) Similarity-based: For each review, we connect it with its top-k structurally or textually similar neighbors (given by the released similarity edges) into a hyperedge.
  • Amazon (user nodes). (i) Co-item: For each item p, users who reviewed p in a window form a hyperedge. (ii) Temporal co-activity: For each ( p , t ) , users reviewing p within the same temporal bin form a hyperedge. (iii) Pattern-based: Users with similar rating or activity statistics (computed from released features) are grouped into hyperedges.
  • DGraph-Fin (user nodes with contact edges u v ). (i) Shared-contact: For each contact user v and window t, a hyperedge connects all u such that u v occurs in t. (ii) Contact-pattern: For each contact type c and window t, users with at least one outgoing edge of type c in t form a hyperedge. (iii) Local structure: For each user, we build hyperedges over densely connected 1-hop ego neighborhoods to summarize tight contact circles.
The resulting dynamic hypergraphs { H t } t = 1 T are then fed into the hypergraph encoder to obtain node representations H t R | 𝒱 | × d h .
To avoid any label-dependent bias or artificial homophily, all hypergraph construction rules are strictly label-agnostic: hyperedges are defined only from publicly released identifiers and metadata that are also available at inference time (such as user and item IDs, emergency-contact edges, contact-type codes, and discretized timestamps), without using fraud labels or manually engineered “suspicious” patterns. Although the concrete grouping rules differ across YelpChi, Amazon, and DGraph-Fin due to their heterogeneous graph schemas (review-centric graphs versus user–contact networks), they follow the same principle of connecting nodes that share observable context in the original benchmarks. Structural statistics of the three benchmarks are summarized in Figure 3.

3.2. Multi-Head Spatio-Temporal Hypergraph Attention

We now describe the MH-DHA encoder, which processes the sequence of dynamic hypergraphs { H t } t = 1 T to produce spatio-temporal node representations.

3.2.1. Spatial Hypergraph Attention

For each time window t, we consider the hypergraph H t = ( 𝒱 , E t H ) with incidence matrix B t { 0 , 1 } | 𝒱 |   ×   | E t H | . Let h v ( l , t ) and h e ( l , t ) denote node and hyperedge embeddings at layer l and time t.
We adopt a two-stage node–hyperedge attention mechanism with K heads. For clarity, we omit the time index t when not ambiguous.
Node-to-Hyperedge Aggregation
For each hyperedge e and node v e , we compute attention scores:
α v e ( k ) = exp σ a 1 ( k ) W 1 ( k ) h v ( l ) W 1 ( k ) h e ( l 1 ) u e exp σ a 1 ( k ) W 1 ( k ) h u ( l ) W 1 ( k ) h e ( l 1 ) ,
where · denotes concatenation, σ ( · ) is a nonlinearity, and  ( W 1 ( k ) , a 1 ( k ) ) are trainable parameters for head k. The hyperedge embedding is
h e ( l ) = k = 1 K ELU v e α v e ( k ) W 1 ( k ) h v ( l ) ,
where  denotes concatenation across heads.
Hyperedge-to-Node Aggregation
Similarly, for each node v and incident hyperedge e E v , we compute
β e v ( k ) = exp σ a 2 ( k ) W 2 ( k ) h e ( l ) W 2 ( k ) h v ( l ) f E v exp σ a 2 ( k ) W 2 ( k ) h f ( l ) W 2 ( k ) h v ( l ) ,
and update the node embedding as
h v ( l + 1 ) = k = 1 K ELU e E v β e v ( k ) W 2 ( k ) h e ( l ) .
After L layers of hypergraph attention, we obtain the spatial representations S t R | 𝒱 | × d s for each time t.

3.2.2. Temporal Aggregation

To capture temporal evolution, we feed the sequence { S 1 , , S T } into a Gated Recurrent Unit (GRU) at the node level. For each node v, its embedding at time t is denoted by s v ( t ) , and the GRU updates:
h v ( t ) = GRU s v ( t ) , h v ( t 1 ) ,
H init = H ( T ) = [ h v ( T ) ] v 𝒱 .
When T = 1 , the GRU reduces to a simple transformation, and  H init coincides with S 1 up to a linear mapping.

3.3. Causal Disentanglement and Backdoor Adjustment

We now describe the causal module that operates on H init to obtain robust node representations (Figure 4).

3.3.1. Disentangled Representations

We assume that H init encodes both intrinsic causal factors C (e.g., stable behavioral patterns associated with fraudulent intent) and environment-related factors E (e.g., structural hubs, temporal bursts, or global usage shifts). We use two MLP projectors:
Z c = MLP c ( H init ) , Z e = MLP e ( H init ) .
To encourage Z c and Z e to capture distinct information, we penalize their statistical dependence via the Hilbert–Schmidt Independence Criterion (HSIC) [20]:
L dis = HSIC ( Z c , Z e ) = 1 ( n 1 ) 2 Tr ( K c H K e H ) ,
where K c and K e are kernel matrices computed from Z c and Z e , H is the centering matrix, and n is the batch size.

3.3.2. Dictionary-Based Backdoor Adjustment

We model the environment-related factors via a learnable dictionary M = [ m 1 , , m K ] R K × d e . Intuitively, each dictionary atom m k corresponds to a prototypical environment pattern, such as a certain combination of node degree, temporal activity, and hyperedge participation statistics derived from the publicly available data.
Following Pearl’s backdoor adjustment [9], the interventional distribution P ( Y d o ( C ) ) can be approximated by averaging over environment patterns:
P ( Y d o ( C ) ) k = 1 K P ( Y C , E = m k ) P ( E = m k ) .
In practice, we implement this via a scaled dot-product attention over the dictionary. Let Q , K , V be linear projections applied to Z c and M :
Q = Z c W Q ,
K M = M W K ,
V M = M W V .
The attention weights over the dictionary are
A = Softmax Q K M d ,
and the aggregated environmental contribution is
E adj = A V M .
We then form the robust representation:
Z * = [ Z c E adj ] ,
which is used for both supervised fraud classification and contrastive learning.

3.4. Multi-View Contrastive Learning

To exploit unlabeled nodes and enhance robustness under feature/structure perturbations, we adopt a self-supervised multi-view contrastive learning strategy.

3.4.1. Hypergraph Augmentations

We generate two augmented hypergraph views H ( 1 ) and H ( 2 ) by applying stochastic perturbations that are independent of fraud labels:
  • Hyperedge dropping. Each hyperedge is removed with probability p edge , simulating missing or noisy group interactions.
  • Feature masking. Each feature dimension is masked with probability p feat , simulating incomplete or corrupted features (e.g., missing profile entries).
Passing each view through the MH-DHA encoder and causal module yields robust representations Z * ( 1 ) and Z * ( 2 ) . A small MLP projection head g ( · ) maps them into z ( 1 ) and z ( 2 ) for contrastive learning.

3.4.2. Contrastive Objective

For a mini-batch of N nodes, we define the InfoNCE loss for each node i:
( i ) = log exp sim ( z i ( 1 ) , z i ( 2 ) ) / τ j = 1 N I [ j i ] exp sim ( z i ( 1 ) , z j ( 1 ) ) / τ + j = 1 N exp sim ( z i ( 1 ) , z j ( 2 ) ) / τ ,
where sim ( u , v ) = u v / ( u v ) is cosine similarity, I [ · ] is the indicator function, and  τ is a temperature hyperparameter. The total contrastive loss is
L cl = 1 2 N i = 1 N ( i ) + ( i ) ,
where ( i ) is defined symmetrically by swapping the two views.

3.5. Supervised Objective and Joint Training

For labeled nodes 𝒱 L , we apply a linear classifier on top of Z * and use binary cross-entropy loss as the supervised objective:
L sup = 1 | 𝒱 L | v 𝒱 L y v log y ^ v + ( 1 y v ) log ( 1 y ^ v ) .
The total training loss combines all components:
L total = L sup + λ 1 L dis + λ 2 L cl ,
where λ 1 and λ 2 control the contributions of the disentanglement and contrastive terms.

4. Theoretical Analysis

This section provides a theoretical characterization of Causal-DHG. We first formalize a structural causal model (SCM) on dynamic hypergraphs and the associated disentangled representations, then analyze the dictionary-based backdoor approximation and robustness under environment shift, and finally summarize the computational complexity.

4.1. Causal Formulation and Disentangled Representations

We consider node features X R d x derived from publicly available statistics (e.g., review behavior and profile attributes), a sequence of hypergraph incidence structures A H = { A t H } t = 1 T constructed from observed interactions, intrinsic causal factors C encoding user intent and stable behavioral mechanisms related to fraud, environment-related factors E capturing structural and temporal context (e.g., local degree patterns, hyperedge participation, time windows), and the fraud label Y { 0 , 1 } . We model these variables via the SCM
C = f c ( X , A H , U c ) ,
E = f e ( X , A H , U e ) ,
Y = g ( C , E , U y ) ,
where U c , U e , U y are mutually independent exogenous noise variables. The associated causal graph contains directed edges
X C , A H C , X E , A H E , C Y , E Y .
Our goal is to approximate the interventional distribution
P ( Y d o ( C = c ) ) = e P ( Y C = c , E = e ) P ( E = e ) ,
which blocks spurious environment-driven pathways. We adopt the following assumptions.
Assumption 1
(Environment shift and invariance). Between training and deployment, the conditional mechanism P ( Y C , E ) remains invariant, while the environment distribution P ( E ) and its dependence on ( X , A H ) may change:
P train ( Y C , E ) = P test ( Y C , E ) , P train ( E X , A H ) P test ( E X , A H ) in general .
Assumption 2
(Representation sufficiency). There exist encoders h c and h e such that the learned representations
Z c = h c ( X , A H ) , Z e = h e ( X , A H )
are sufficient statistics for C and E in the sense that
P ( Y C , E ) = P ( Y Z c , Z e ) , P ( E C ) = P ( Z e Z c ) ,
up to invertible transformations in the representation space.
Causal-DHG instantiates these representations as
Z c = MLP c ( H init ) , Z e = MLP e ( H init ) ,
where H init is the output of the dynamic hypergraph encoder. Under Assumptions 1 and 2, approximating P ( Y d o ( C ) ) reduces to learning Z c that is predictive for Y but insensitive to shifts in P ( E ) , while Z e captures environment-specific variability that can be marginalized out.

4.2. HSIC-Based Disentanglement

Causal-DHG enforces independence between Z c and Z e via an HSIC-based regularizer. Let k c and k e be characteristic kernels on the domains of Z c and Z e , respectively, and let HSIC k c , k e ( Z c , Z e ) denote the corresponding empirical HSIC [20]. The disentanglement loss is
L dis = HSIC k c , k e ( Z c , Z e ) .
Proposition 1
(HSIC and independence). If k c and k e are characteristic kernels, then HSIC k c , k e ( Z c , Z e ) = 0 if and only if Z c and Z e are statistically independent. Minimizing (24) therefore drives the learned causal and environmental representations toward independence.
Proof sketch. 
For characteristic kernels, the kernel mean embedding of a distribution is injective. HSIC can be written as the squared Hilbert–Schmidt norm of the cross-covariance operator between the embeddings of Z c and Z e [20]. This norm is zero if and only if the cross-covariance operator is zero, which is equivalent to independence. The empirical HSIC is a consistent estimator of the population quantity, so minimizing it encourages independence in the large-sample limit.    □
Combined with Assumption 2, Proposition 1 implies that Z c can align with intrinsic causal factors, while Z e captures orthogonal environment-related patterns. In practice, we use mini-batch estimates of HSIC; although exact independence is unattainable in finite samples, prior work [10,19,21,22] shows that HSIC penalties are effective in suppressing spurious dependencies.

4.3. Dictionary-Based Backdoor Approximation and Robustness

The dictionary-based module aims to approximate the backdoor-adjusted predictor associated with (23) at the representation level. Let M = [ m 1 , , m K ] be the environmental dictionary, and let α k ( Z c ) be the attention weight assigned to atom m k given Z c . Let v k denote the value projection of m k . The adjusted environment representation is
E adj ( Z c ) = k = 1 K α k ( Z c ) v k .
We require mild coverage and calibration conditions on the dictionary.
Assumption 3
(Dictionary coverage and calibration). The atoms { m k } k = 1 K form an ε-cover of the support of E; i.e., for any environment realization e there exists some k with e m k ε , and the attention weights satisfy α k ( Z c ) P ( E = m k Z c ) for all k.
Denote by h θ the classifier acting on [ Z c E adj ( Z c ) ] , and assume it is Lipschitz in its environment-related argument.
Proposition 2
(Variational backdoor approximation). Under Assumptions 1–3 and Lipschitz continuity of h θ in its second argument, the dictionary-based predictor
Y ^ ( Z c ) = h θ Z c , E adj ( Z c )
provides a finite-dimensional variational approximation to the interventional predictor based on P ( Y d o ( C ) ) . The approximation error is bounded by a constant that scales with the dictionary resolution ε.
Proof sketch. 
By Assumption 2, P ( Y C , E ) = P ( Y Z c , E ) , and the ideal backdoor-adjusted predictor based on Z c is
f * ( Z c ) = e P ( Y Z c , E = e ) P ( E = e ) .
Assumption 3 allows approximating (27) by a finite mixture over { m k } :
f * ( Z c ) k = 1 K P ( Y Z c , E = m k ) P ( E = m k ) ,
with an error controlled by ε and the Lipschitz constant of P ( Y Z c , E ) in E. If  α k ( Z c ) P ( E = m k Z c ) , then E adj ( Z c ) in (25) approximates a suitable conditional expectation of an environment embedding, and  h θ in (26) can approximate the mapping from ( Z c , E [ g E ( E ) Z c ] ) to f * ( Z c ) . Lipschitz continuity ensures that the induced error is bounded by a function of ε .    □
Beyond approximating P ( Y d o ( C ) ) , this construction improves robustness under environment shift. Let ( y ^ , y ) be a bounded loss and define
R ( h ; P ) = E ( X , A H , Y ) P ( h ( X , A H ) , Y ) ,
the expected risk of predictor h under distribution P. Consider a correlation-based baseline h corr ( X , A H ) and the Causal-DHG predictor h causal ( X , A H ) based on backdoor-adjusted representations. Let d ( · , · ) be an integral probability metric (e.g., a Wasserstein distance) on environment distributions.
Assumption 4
(Lipschitz dependence on environment). There exist constants L corr and L causal such that for any fixed C and any e 1 , e 2 ,
| ( h corr ( C , e 1 ) , Y ) ( h corr ( C , e 2 ) , Y ) | L corr e 1 e 2 , | ( h causal ( C , e 1 ) , Y ) ( h causal ( C , e 2 ) , Y ) | L causal e 1 e 2 .
Due to HSIC-based disentanglement and backdoor adjustment, we expect L causal L corr .
Lemma 1
(Risk sensitivity to environment shift). Let P train and P test satisfy Assumption 1. Under Assumption 4, there exist constants K corr , K causal , such that
| R ( h corr ; P test ) R ( h corr ; P train ) | K corr d P train ( E ) , P test ( E ) , | R ( h causal ; P test ) R ( h causal ; P train ) | K causal d P train ( E ) , P test ( E ) ,
with K causal / K corr L causal / L corr .
Proof sketch. 
For a fixed predictor, the change in risk when moving from P train to P test that share the same P ( Y C , E ) is controlled by the change in P ( E ) . Under Lipschitz dependence on E, this difference can be bounded by constant times d ( P train ( E ) , P test ( E ) ) . The constants K corr , K causal scale with L corr , L causal , respectively. Since Causal-DHG explicitly suppresses spurious dependence on E, one obtains L causal L corr and thus a tighter bound.    □
Lemma 1 formalizes the robustness improvement observed in our feature perturbation experiments: for a fixed magnitude of environment shift d ( P train ( E ) , P test ( E ) ) , the risk variation of Causal-DHG is controlled by a smaller constant factor than that of correlation-based baselines.

4.4. Complexity Guarantees

We finally summarize the computational complexity of Causal-DHG. Let | 𝒱 | be the number of nodes, | E H | the number of hyperedges, and  | I | the number of node–hyperedge incidences (i.e., non-zero entries in the incidence matrices). Denote by L the number of MH-DHA layers, T the number of time windows, d the hidden dimension, K the dictionary size, and B the batch size.
Theorem 1
(Time and space complexity). Assume that (i) the maximum number of hyperedges incident to any node is bounded by C h , and (ii) the maximum size of any hyperedge is bounded by C s . Then a single training epoch of Causal-DHG has time complexity
O L T | I | d + B 2 + B K d ,
and space complexity
O | 𝒱 | d + | E H | d + K d .
In particular, since | I | = O ( | 𝒱 | C h ) = O ( | E H | C s ) with C h , C s treated as constants, the time complexity is linear in | 𝒱 | + | E H | up to logarithmic and constant factors.
Proof sketch. 
Each MH-DHA layer performs node-to-hyperedge and hyperedge-to-node attention. Both directions cost O ( | I | d ) operations, leading to O ( L T | I | d ) over L layers and T time windows. The GRU-based temporal aggregation scales as O ( L T | 𝒱 | d 2 ) and is dominated by the incidence term in sparse hypergraphs. The HSIC term involves batch-wise kernel matrices with cost O ( B 2 ) per batch, and the dictionary-based module performs attention over K atoms per node, giving O ( B K d ) per batch. Space is dominated by storing node and hyperedge embeddings, O ( | 𝒱 | d + | E H | d ) , and the dictionary O ( K d ) . Under bounded C h and C s , the total number of incidences | I | scales linearly with | 𝒱 | and | E H | , completing the argument.    □
Theorem 1 shows that Causal-DHG retains linear-time scalability in the size of the dynamic hypergraph, making it suitable for large-scale fraud detection scenarios such as DGraph-Fin, while providing theoretically grounded robustness benefits through causal modeling.

5. Experimental Section

5.1. Experimental Setup

5.1.1. Datasets

We evaluate Causal-DHG on three widely used public fraud detection datasets (Table 1).
YelpChi
This dataset contains review nodes with manually crafted 32-dimensional features capturing textual and behavioral statistics. The binary label indicates whether a review is fraudulent. The released graph includes multiple review–review relations capturing common user and item context. We construct hyperedges as described in Section 3, using user-centric, product-centric, and similarity-based groups. In our experiments, we use the benchmark version provided by the FraudYelpDataset interface in the Deep Graph Library (DGL). (See the DGL documentation at FraudYelpDataset (https://www.dgl.ai/dgl_docs/generated/dgl.data.FraudYelpDataset.html (accessed on 1 December 2025))).
Amazon
FraudAmazonDataset represents users as nodes with 25-dimensional features summarizing their reviewing behaviors, including activity patterns and rating statistics. Labels denote fraudulent or benign users. We build hyperedges based on co-reviewing the same items, temporal co-activity, and behavior-pattern similarity, all derived from available metadata and edges. We use the processed benchmark provided by the FraudAmazonDataset interface in DGL. (See the DGL documentation at FraudAmazonDataset (https://www.dgl.ai/dgl_docs/generated/dgl.data.FraudAmazonDataset.html (accessed on 1 December 2025))).
DGraph-Fin
DGraph-Fin is a large credit-related user contact network from Finvolution, with 17-dimensional anonymous node features. Directed edges represent emergency-contact relations with discrete timestamps and contact types. We construct hyperedges that group users sharing the same contact in a time window, users with similar contact-type patterns, and dense local contact circles. The dataset and leaderboard are publicly available through the DGraph platform (https://dgraph.xinye.com (accessed on 1 December 2025)).

5.1.2. Baselines

We compare Causal-DHG with a diverse set of baselines:
  • GCN [23]: A standard spectral graph convolutional network.
  • GAT [24]: A graph attention network.
  • GraphSAGE [11]: A neighborhood sampling-based GNN.
  • H-GNN [7]: A hypergraph neural network applied to static hypergraphs constructed from the same grouping strategy as our method but without temporal or causal modules.
  • EvolveGCN [15]: Temporal GNN modeling evolving graphs.
  • CARE-GNN [3]: A fraud-specialized GNN with camouflage-resistant neighbor selection.
  • PC-GNN [6]: A GNN for imbalanced and camouflaged fraud detection with pick-and-choose sampling.

5.1.3. Evaluation Protocol

We follow common practice and report Area Under the ROC Curve (AUC) and macro-averaged F1 (F1-Macro). For each dataset, we randomly split nodes into training, validation, and test sets with ratios of 60 % / 20 % / 20 % , preserving the label distribution. All results are averaged over five runs with different random seeds. We additionally report precision–recall AUC (PR-AUC/Average Precision) in Appendix A.1 Table A1 to complement ROC-AUC under class imbalance.

5.1.4. Implementation Details

We implement Causal-DHG in PyTorch 2.6.0 with appropriate graph libraries. The node embedding dimension is set to 64, with L = 2 hypergraph attention layers and K = 4 attention heads by default. The GRU hidden dimension equals 64. The environmental dictionary size is K = 32 . We use the Adam optimizer with a learning rate of 10 3 , weight decay 5 × 10 5 , and batch sizes of 256 (YelpChi), 256 (Amazon), and 1024 (DGraph-Fin). For temporal modeling, YelpChi and Amazon are treated as single-snapshot benchmarks ( T = 1 ) due to the released benchmark setting, while for DGraph-Fin, we discretize timestamps into T = 7 non-overlapping windows (see Appendix A.3 Table A3 for a window-granularity sweep). The loss weights are λ 1 = 0.1 and λ 2 = 0.5 . Hyperedge dropping probability p edge and feature masking probability p feat are set to 0.2 and 0.1 , respectively, and the contrastive temperature τ is fixed at 0.5 . These values are chosen in the commonly used range of 0.1 0.2 for graph contrastive perturbations, in line with prior work such as GraphCL [25]. HSIC is estimated on each mini-batch; across five runs, we did not observe training divergence or instability attributable to HSIC estimation. Experiments were conducted on an NVIDIA A100 GPU (40GB) with 64 GB RAM and an Intel Xeon CPU.

5.2. Overall Performance

Table 2 reports the main results on the three datasets. Causal-DHG consistently outperforms all baselines in terms of both AUC and F1-Macro, while maintaining moderate absolute gains (generally within four percentage points) over the strongest methods.
On YelpChi, Causal-DHG achieves an AUC of 0.9045 and F1-Macro of 0.7820, surpassing PC-GNN by approximately 2.3 and 3.7 percentage points, respectively. On Amazon, the gains are around 1.9 and 3.3 percentage points in AUC and F1-Macro. On the more challenging DGraph-Fin dataset, Causal-DHG improves AUC from 0.7620 to 0.7985 and F1-Macro from 0.6350 to 0.6870. These improvements are consistent yet moderate, aligning with practical expectations for incremental gains on strong baselines. We further report PR-AUC (Average Precision) in Appendix A.1 Table A1. Sensitivity analyses on hyperedge size and temporal window granularity, as well as robustness to structural perturbations, are provided in Appendix A Table A2, Table A3 and Table A4.

5.3. Robustness to Feature Perturbations

To evaluate robustness against environment-related noise, we inject controlled perturbations into node features that serve as proxies for environmental statistics. Concretely, for each dataset, we first compute simple per-dimension statistics (such as variance and correlation with local context indicators like node degree and temporal activity counts). We then select a small subset of feature dimensions whose values exhibit strong dependence on these context indicators and treat them as environment-sensitive proxies. For each node, we add Gaussian noise ϵ N ( 0 , σ 2 ) to this subset of dimensions and measure the AUC degradation as σ increases.
As shown in Table 3, PC-GNN suffers significant AUC drops as noise increases, while Causal-DHG degrades much more slowly. At the highest noise level ( σ = 2.0 ), Causal-DHG retains an AUC of 0.8520, compared to 0.7150 for PC-GNN. This suggests that disentangling causal and environment-related factors and performing backdoor adjustment indeed enhances robustness against perturbations that mimic environmental shifts.

5.4. Ablation Study

We conduct ablation studies on YelpChi to assess the contribution of each component: hypergraph modeling, causal disentanglement, and contrastive learning (Table 4).
Removing the causal module (disentanglement and backdoor adjustment) causes the largest drop in performance, with AUC decreasing from 0.9045 to 0.8720 and F1-Macro from 0.7820 to 0.7310. This confirms that explicitly modeling and marginalizing out environment-related factors is critical for robust fraud detection. Replacing hypergraphs with pairwise graphs also reduces performance, validating the importance of high-order interaction modeling. Finally, discarding contrastive learning yields a moderate performance loss, indicating that leveraging unlabeled data improves representation quality under label scarcity.

5.5. Representation Analysis

To further understand the effect of the causal module, we visualize node embeddings via t-SNE on YelpChi.
As shown in Figure 5, the embeddings produced by Causal-DHG exhibit clearer separation between fraudulent and benign nodes compared to PC-GNN, consistent with the quantitative improvements in AUC and F1-Macro.

5.6. Hyperparameter Sensitivity and Efficiency

We study the sensitivity of Causal-DHG to the dictionary size K and analyze its training and inference efficiency (Figure 6).
Table 5 summarizes the efficiency comparison on DGraph-Fin. Causal-DHG requires slightly more training time per epoch than PC-GNN due to hypergraph attention and the causal module, but the inference time remains comparable and suitable for large-scale deployment.

6. Conclusions

This paper proposed Causal-DHG, a dynamic hypergraph neural network with causal intervention for robust financial fraud detection. We constructed dynamic hypergraphs that explicitly capture high-order interaction patterns using only observed entities and relations. On top of these structures, we designed a Multi-Head Spatio-Temporal Hypergraph Attention encoder and a Causal Disentanglement Module with dictionary-based backdoor adjustment to separate and marginalize environment-related factors. A multi-view contrastive objective was further introduced to leverage unlabeled nodes and enhance robustness under feature and structural perturbations.
Extensive experiments demonstrated that Causal-DHG achieves consistent performance gains over strong baselines while providing substantially improved robustness against environment-related noise. Ablation and representation analyses confirmed that both hypergraph modeling and causal intervention play essential roles. These findings suggest that integrating causal reasoning into hypergraph-based fraud detection is a promising step toward trustworthy AI in financial applications.
Future work includes extending the framework to multi-task settings (e.g., jointly detecting fraud and predicting credit risk), incorporating more fine-grained temporal encoders, and exploring causal discovery techniques to learn structural causal relations directly from data.

Funding

The author received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.

Data Availability Statement

The data presented in this study are openly available in [DGraph-Fin] [https://dgraph.xinye.com (accessed on 1 December 2025)].

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A. Additional Experiments and Analyses

Appendix A.1. PR-AUC (Average Precision)

Table A1. PR-AUC (Average Precision) on YelpChi, Amazon, and DGraph-Fin (mean ± std over five runs). Note: DGraph-Fin values are lower due to extreme class imbalance (approx. 1%).
Table A1. PR-AUC (Average Precision) on YelpChi, Amazon, and DGraph-Fin (mean ± std over five runs). Note: DGraph-Fin values are lower due to extreme class imbalance (approx. 1%).
MethodYelpChiAmazonDGraph-Fin
CARE-GNN0.7124 ± 0.0110.7850 ± 0.0090.2215 ± 0.013
PC-GNN0.7560 ± 0.0090.8120 ± 0.0080.2450 ± 0.011
Causal-DHG0.7950 ± 0.0080.8410 ± 0.0060.2940 ± 0.009

Appendix A.2. Sensitivity to Hyperedge Size (Top-ksim in Similarity Hyperedges)

Table A2. Sensitivity of Causal-DHG to the hyperedge size controlled by the top-ksim neighbors in the similarity-based hyperedge construction on YelpChi. We report AUC (mean over five runs).
Table A2. Sensitivity of Causal-DHG to the hyperedge size controlled by the top-ksim neighbors in the similarity-based hyperedge construction on YelpChi. We report AUC (mean over five runs).
ksim510152030
AUC0.89500.90200.90450.90350.9030

Appendix A.3. Sensitivity to the Number of Time Windows T (DGraph-Fin)

Table A3. Effect of time-window granularity T on DGraph-Fin. Edge timestamps are discretized into T non-overlapping windows; we report AUC (mean over five runs).
Table A3. Effect of time-window granularity T on DGraph-Fin. Edge timestamps are discretized into T non-overlapping windows; we report AUC (mean over five runs).
T135710
AUC0.76100.78500.79200.79850.7820

Appendix A.4. Robustness to Structural Perturbations

Table A4. Robustness to structural perturbations on YelpChi. We randomly drop a fraction r of edges for PC-GNN and a fraction r of hypergraph incidences for Causal-DHG at test time, with r { 0 , 0.1 , 0.2 } . We report AUC (mean over five runs).
Table A4. Robustness to structural perturbations on YelpChi. We randomly drop a fraction r of edges for PC-GNN and a fraction r of hypergraph incidences for Causal-DHG at test time, with r { 0 , 0.1 , 0.2 } . We report AUC (mean over five runs).
Method r = 0 r = 0.1 r = 0.2
PC-GNN0.88100.85400.8210
Causal-DHG0.90450.89500.8820

References

  1. The Nilson Report. Card Fraud Losses Dip to $28.58 Billion in 2020; The Nilson Report: Santa Barbara, CA, USA, 2020. [Google Scholar]
  2. Xu, D.; Ruan, C.; Korpeoglu, E.; Kumar, S.; Achan, K. Inductive representation learning on temporal graphs. arXiv 2020, arXiv:2002.07962. [Google Scholar] [CrossRef]
  3. Dou, Y.; Liu, Z.; Sun, L.; Deng, Y.; Peng, H.; Yu, P.S. Enhancing graph neural network-based fraud detectors against camouflaged fraudsters. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Virtual Event, 19–23 October 2020; pp. 315–324. [Google Scholar] [CrossRef]
  4. Wang, D.; Lin, J.; Cui, P.; Jia, Q.; Wang, Z.; Fang, Y.; Yu, Q.; Zhou, J.; Yang, S.; Qi, Y. A semi-supervised graph attentive network for financial fraud detection. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; IEEE: New York, NY, USA, 2019; pp. 598–607. [Google Scholar] [CrossRef]
  5. Liu, Z.; Chen, C.; Li, L.; Zhou, J.; Li, X.; Song, L.; Qi, Y. Geniepath: Graph neural networks with adaptive receptive paths. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4424–4431. [Google Scholar] [CrossRef]
  6. Liu, Y.; Ao, X.; Qin, Z.; Chi, J.; Feng, J.; Yang, H.; He, Q. Pick and choose: A GNN-based imbalanced learning approach for fraud detection. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 12–23 April 2021; pp. 3168–3177. [Google Scholar] [CrossRef]
  7. Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y. Hypergraph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 3558–3565. [Google Scholar] [CrossRef]
  8. Bengio, Y.; Deleu, T.; Rahaman, N.; Ke, R.; Lachapelle, S.; Bilaniuk, O.; Goyal, A.; Pal, C. A Meta-Transfer Objective for Learning to Disentangle Causal Mechanisms. In Proceedings of the International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar] [CrossRef]
  9. Pearl, J. Causality, 2nd ed.; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  10. Wu, Q.; Zhang, H.; Yan, J.; Wipf, D. Handling distribution shifts on graphs: An invariance perspective. arXiv 2022, arXiv:2202.02466. [Google Scholar] [CrossRef]
  11. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar] [CrossRef]
  12. Liu, Y.; Wu, J.; Zhang, J.; Leung, M.F. Graph-Regularized Orthogonal Non-Negative Matrix Factorization with Itakura–Saito (IS) Divergence for Fault Detection. Mathematics 2025, 13, 2343. [Google Scholar] [CrossRef]
  13. Bretto, A. Hypergraph Theory: An Introduction; Springer: Cham, Switzerland, 2013; Volume 1, pp. 209–216. [Google Scholar]
  14. Jiang, J.; Wei, Y.; Feng, Y.; Cao, J.; Gao, Y. Dynamic hypergraph neural networks. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 2635–2641. [Google Scholar] [CrossRef]
  15. Pareja, A.; Domeniconi, G.; Chen, J.; Ma, T.; Suzumura, T.; Kanezashi, H.; Kaler, T.; Schardl, T.; Leiserson, C. Evolvegcn: Evolving graph convolutional networks for dynamic graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 5363–5370. [Google Scholar] [CrossRef]
  16. Rossi, E.; Chamberlain, B.; Frasca, F.; Eynard, D.; Monti, F.; Bronstein, M. Temporal graph networks for deep learning on dynamic graphs. arXiv 2020, arXiv:2006.10637. [Google Scholar] [CrossRef]
  17. Xia, X.; Yin, H.; Yu, J.; Wang, Q.; Cui, L.; Zhang, X. Self-supervised hypergraph convolutional networks for session-based recommendation. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 2–9 February 2021; Volume 35, pp. 4503–4511. [Google Scholar] [CrossRef]
  18. Schölkopf, B.; Locatello, F.; Bauer, S.; Ke, N.R.; Kalchbrenner, N.; Goyal, A.; Bengio, Y. Toward causal representation learning. Proc. IEEE 2021, 109, 612–634. [Google Scholar] [CrossRef]
  19. Kuang, K.; Xiong, R.; Cui, P.; Athey, S.; Li, B. Stable Prediction with Model Misspecification and Agnostic Distribution Shift. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar] [CrossRef]
  20. Gretton, A.; Bousquet, O.; Smola, A.; Schölkopf, B. Measuring Statistical Dependence with Hilbert-Schmidt Norms. In Proceedings of the Algorithmic Learning Theory (ALT), Singapore, 8–11 October 2005. [Google Scholar] [CrossRef]
  21. Bae, I.; Jeon, H.G. Disentangled multi-relational graph convolutional network for pedestrian trajectory prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 2–9 February 2021; Volume 35, pp. 911–919. [Google Scholar] [CrossRef]
  22. Yang, D.; Zha, D.; Kurokawa, R.; Zhao, T.; Wang, H.; Zou, N. Factorizable Graph Convolutional Networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Virtual Event, 6–12 December 2020. [Google Scholar] [CrossRef]
  23. Kipf, T. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar] [CrossRef]
  24. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. Stat 2017, 1050, 10-48550. [Google Scholar] [CrossRef]
  25. You, Y.; Chen, T.; Sui, Y.; Chen, T.; Wang, Z.; Shen, Y. Graph contrastive learning with augmentations. Adv. Neural Inf. Process. Syst. 2020, 33, 5812–5823. [Google Scholar] [CrossRef]
Figure 1. Motivating comparison between correlation-based GNNs and Causal-DHG. (Left): Standard GNNs aggregate neighbor information indiscriminately and tend to rely on environment-driven structural patterns, such as high-degree hubs or temporally concentrated bursts, which may be spuriously correlated with fraud labels. (Right): The proposed Causal-DHG disentangles intrinsic causal factors from environment-related factors and performs backdoor-style intervention to block spurious environmental paths, yielding representations that are more robust under distribution shifts.
Figure 1. Motivating comparison between correlation-based GNNs and Causal-DHG. (Left): Standard GNNs aggregate neighbor information indiscriminately and tend to rely on environment-driven structural patterns, such as high-degree hubs or temporally concentrated bursts, which may be spuriously correlated with fraud labels. (Right): The proposed Causal-DHG disentangles intrinsic causal factors from environment-related factors and performs backdoor-style intervention to block spurious environmental paths, yielding representations that are more robust under distribution shifts.
Mathematics 13 04018 g001
Figure 2. Overall Causal-DHG architecture. Raw graphs are converted into dynamic hypergraphs and encoded by the MH-DHA module to obtain initial node representations H init . The Causal Disentanglement Module projects H init into causal factors Z c and environmental factors Z e , enforces their independence (HSIC), and performs dictionary-based backdoor adjustment to produce an environment-marginalized representation Z * . The model is optimized jointly with supervised loss L sup and contrastive loss L cl .
Figure 2. Overall Causal-DHG architecture. Raw graphs are converted into dynamic hypergraphs and encoded by the MH-DHA module to obtain initial node representations H init . The Causal Disentanglement Module projects H init into causal factors Z c and environmental factors Z e , enforces their independence (HSIC), and performs dictionary-based backdoor adjustment to produce an environment-marginalized representation Z * . The model is optimized jointly with supervised loss L sup and contrastive loss L cl .
Mathematics 13 04018 g002
Figure 3. Structural statistics of YelpChi, Amazon, and DGraph-Fin. (Top row): Node-degree CDFs P ( K k ) versus degree k (log–log scale), showing a long-tail with a small number of super-nodes. (Bottom row): Hyperedge-size CDFs P ( S s ) versus hyperedge size s (log–log scale), indicating the existence of dense groups. These heavy-tailed patterns motivate explicit hypergraph modeling of high-order relations rather than relying only on pairwise edges.
Figure 3. Structural statistics of YelpChi, Amazon, and DGraph-Fin. (Top row): Node-degree CDFs P ( K k ) versus degree k (log–log scale), showing a long-tail with a small number of super-nodes. (Bottom row): Hyperedge-size CDFs P ( S s ) versus hyperedge size s (log–log scale), indicating the existence of dense groups. These heavy-tailed patterns motivate explicit hypergraph modeling of high-order relations rather than relying only on pairwise edges.
Mathematics 13 04018 g003
Figure 4. Causal Disentanglement Module. Node representations H init are projected into a causal subspace Z c and an environment-related subspace Z e via two MLPs. HSIC regularization encourages independence between Z c and Z e . A learnable environmental dictionary aggregates global environment prototypes, and a backdoor-style adjustment uses attention over the dictionary to marginalize out spurious environmental factors, yielding robust representations Z * for classification and contrastive learning.
Figure 4. Causal Disentanglement Module. Node representations H init are projected into a causal subspace Z c and an environment-related subspace Z e via two MLPs. HSIC regularization encourages independence between Z c and Z e . A learnable environmental dictionary aggregates global environment prototypes, and a backdoor-style adjustment uses attention over the dictionary to marginalize out spurious environmental factors, yielding robust representations Z * for classification and contrastive learning.
Mathematics 13 04018 g004
Figure 5. t-SNE visualization of node embeddings on YelpChi. (a) PC-GNN embeddings show considerable overlap between normal and fraudulent nodes. (b) Causal-DHG embeddings result in more clearly separated clusters, illustrating that the causal module helps to suppress environment-driven noise and emphasize intrinsic fraudulent patterns.
Figure 5. t-SNE visualization of node embeddings on YelpChi. (a) PC-GNN embeddings show considerable overlap between normal and fraudulent nodes. (b) Causal-DHG embeddings result in more clearly separated clusters, illustrating that the causal module helps to suppress environment-driven noise and emphasize intrinsic fraudulent patterns.
Mathematics 13 04018 g005
Figure 6. Model analysis. (a) AUC on YelpChi as a function of dictionary size K; performance improves as K increases from 8 to 32 and stabilizes thereafter. (b) Training time per epoch on DGraph-Fin for PC-GNN and Causal-DHG, showing that Causal-DHG incurs only a moderate overhead relative to PC-GNN while providing stronger robustness. The arrow indicates performance trend only and does not encode additional information.
Figure 6. Model analysis. (a) AUC on YelpChi as a function of dictionary size K; performance improves as K increases from 8 to 32 and stabilizes thereafter. (b) Training time per epoch on DGraph-Fin for PC-GNN and Causal-DHG, showing that Causal-DHG incurs only a moderate overhead relative to PC-GNN while providing stronger robustness. The arrow indicates performance trend only and does not encode additional information.
Mathematics 13 04018 g006
Table 1. Statistics of the datasets used in the experiments. Hyperedge counts are induced by our construction and scale with the number of users/products/contacts and time windows.
Table 1. Statistics of the datasets used in the experiments. Hyperedge counts are induced by our construction and scale with the number of users/products/contacts and time windows.
DatasetNodesEdgesNode Feature DimFraud Ratio
YelpChi45,954 (reviews)3,846,9793214.5%
Amazon11,944 (users)4,398,392256.8%
DGraph-Fin3,700,550 (users)4,300,999171.04%
Table 2. Overall performance on YelpChi, Amazon, and DGraph-Fin. We report AUC and F1-Macro (mean ± standard deviation over five runs). The best result is in bold, and the second best is underlined.
Table 2. Overall performance on YelpChi, Amazon, and DGraph-Fin. We report AUC and F1-Macro (mean ± standard deviation over five runs). The best result is in bold, and the second best is underlined.
MethodYelpChiAmazonDGraph-Fin
AUC F1-Macro AUC F1-Macro AUC F1-Macro
GCN 0.7420 ± 0.010 0.5830 ± 0.018 0.8120 ± 0.012 0.6340 ± 0.019 0.6810 ± 0.017 0.5120 ± 0.021
GAT 0.7640 ± 0.011 0.6120 ± 0.020 0.8250 ± 0.013 0.6510 ± 0.018 0.6950 ± 0.015 0.5340 ± 0.020
GraphSAGE 0.7510 ± 0.012 0.5940 ± 0.017 0.8200 ± 0.011 0.6420 ± 0.016 0.7020 ± 0.016 0.5280 ± 0.019
H-GNN 0.8520 ± 0.010 0.7100 ± 0.019 0.9010 ± 0.009 0.8300 ± 0.017 0.7310 ± 0.018 0.5980 ± 0.022
EvolveGCN 0.8150 ± 0.015 0.6720 ± 0.021 0.8840 ± 0.015 0.8010 ± 0.020 0.7250 ± 0.014 0.5820 ± 0.021
CARE-GNN 0.8650 ± 0.009 0.7210 ± 0.016 0.9150 ± 0.008 0.8450 ± 0.014 0.7450 ± 0.017 0.6120 ± 0.019
PC-GNN 0.8810 ± 0.008 0.7450 ± 0.015 0.9320 ± 0.007 0.8610 ± 0.013 0.7620 ± 0.010 0.6350 ± 0.017
Causal-DHG 0 . 9045 ± 0 . 007 0 . 7820 ± 0 . 012 0 . 9510 ± 0 . 006 0 . 8940 ± 0 . 010 0 . 7985 ± 0 . 009 0 . 6870 ± 0 . 014
Table 3. Robustness evaluation on YelpChi under increasing feature noise levels σ . We report AUC; higher is better.
Table 3. Robustness evaluation on YelpChi under increasing feature noise levels σ . We report AUC; higher is better.
Method σ = 0 σ = 0.5 σ = 1.0 σ = 2.0
PC-GNN0.88100.84500.79200.7150
Causal-DHG0.90450.89110.87500.8520
Table 4. Ablation study on YelpChi. We remove one component at a time and report AUC and F1-Macro.
Table 4. Ablation study on YelpChi. We remove one component at a time and report AUC and F1-Macro.
VariantAUCF1-Macro
Causal-DHG (full)0.90450.7820
   w/o Causal Module0.87200.7310
   w/o Hypergraph (pairwise only)0.88500.7540
   w/o Contrastive Learning0.89100.7620
Table 5. Training and inference efficiency on DGraph-Fin. Inference time is measured per 10,000 nodes.
Table 5. Training and inference efficiency on DGraph-Fin. Inference time is measured per 10,000 nodes.
MethodTrain Time/Epoch (s)Inference Time (ms)
GraphSAGE15.245
CARE-GNN128.5210
PC-GNN145.0235
Causal-DHG168.2260
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luo, X. Robust Financial Fraud Detection via Causal Intervention and Multi-View Contrastive Learning on Dynamic Hypergraphs. Mathematics 2025, 13, 4018. https://doi.org/10.3390/math13244018

AMA Style

Luo X. Robust Financial Fraud Detection via Causal Intervention and Multi-View Contrastive Learning on Dynamic Hypergraphs. Mathematics. 2025; 13(24):4018. https://doi.org/10.3390/math13244018

Chicago/Turabian Style

Luo, Xiong. 2025. "Robust Financial Fraud Detection via Causal Intervention and Multi-View Contrastive Learning on Dynamic Hypergraphs" Mathematics 13, no. 24: 4018. https://doi.org/10.3390/math13244018

APA Style

Luo, X. (2025). Robust Financial Fraud Detection via Causal Intervention and Multi-View Contrastive Learning on Dynamic Hypergraphs. Mathematics, 13(24), 4018. https://doi.org/10.3390/math13244018

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop