Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (50)

Search Parameters:
Keywords = imbalanced graph

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 2096 KiB  
Article
A Hybrid Approach Using Graph Neural Networks and LSTM for Attack Vector Reconstruction
by Yelizaveta Vitulyova, Tetiana Babenko, Kateryna Kolesnikova, Nikolay Kiktev and Olga Abramkina
Computers 2025, 14(8), 301; https://doi.org/10.3390/computers14080301 - 24 Jul 2025
Abstract
The escalating complexity of cyberattacks necessitates advanced strategies for their detection and mitigation. This study presents a hybrid model that integrates Graph Neural Networks (GNNs) with Long Short-Term Memory (LSTM) networks to reconstruct and predict attack vectors in cybersecurity. GNNs are employed to [...] Read more.
The escalating complexity of cyberattacks necessitates advanced strategies for their detection and mitigation. This study presents a hybrid model that integrates Graph Neural Networks (GNNs) with Long Short-Term Memory (LSTM) networks to reconstruct and predict attack vectors in cybersecurity. GNNs are employed to analyze the structural relationships within the MITRE ATT&CK framework, while LSTM networks are utilized to model the temporal dynamics of attack sequences, effectively capturing the evolution of cyber threats. The combined approach harnesses the complementary strengths of these methods to deliver precise, interpretable, and adaptable solutions for addressing cybersecurity challenges. Experimental evaluation on the CICIDS2017 dataset reveals the model’s strong performance, achieving an Area Under the Curve (AUC) of 0.99 on both balanced and imbalanced test sets, an F1-score of 0.85 for technique prediction, and a Mean Squared Error (MSE) of 0.05 for risk assessment. These findings underscore the model’s capability to accurately reconstruct attack paths and forecast future techniques, offering a promising avenue for strengthening proactive defense mechanisms against evolving cyber threats. Full article
(This article belongs to the Section ICT Infrastructures for Cybersecurity)
Show Figures

Figure 1

17 pages, 1192 KiB  
Article
A Power Monitor System Cybersecurity Alarm-Tracing Method Based on Knowledge Graph and GCNN
by Tianhao Ma, Juan Yu, Binquan Wang, Maosheng Gao, Zhifang Yang, Yajie Li and Mao Fan
Appl. Sci. 2025, 15(15), 8188; https://doi.org/10.3390/app15158188 - 23 Jul 2025
Viewed by 46
Abstract
Ensuring cybersecurity in power monitoring systems is of paramount importance to maintain the operational safety and stability of modern power grids. With the rapid expansion of grid infrastructure and increasing sophistication of cyber threats, existing manual alarm-tracing methods face significant challenges in handling [...] Read more.
Ensuring cybersecurity in power monitoring systems is of paramount importance to maintain the operational safety and stability of modern power grids. With the rapid expansion of grid infrastructure and increasing sophistication of cyber threats, existing manual alarm-tracing methods face significant challenges in handling the massive volume of security alerts, leading to delayed responses and potential system vulnerabilities. Current approaches often lack the capability to effectively model complex relationships among alerts and are hindered by imbalanced data distributions, which degrade tracing accuracy. To this end, this paper proposes a power monitor system cybersecurity alarm-tracing method based on the knowledge graph (KG) and graph convolutional neural networks (GCNN). Specifically, a cybersecurity KG is constituted based on the historical alert, accurately representing the entities and relationships in massive alerts. Then, a GCNN with attention mechanisms is applied to sufficiently extract the topological features along alarms in KG so that it can precisely and effectively trace the massive alarms. Most importantly, to mitigate the influence of imbalanced alarms for tracing, a specialized data process and model ensemble strategy by adaptively weighted imbalance sample is proposed. Finally, based on 70,000 alarm information from a regional power grid, by applying the method proposed in this paper, an alarm traceability accuracy rate of 96.59% was achieved. Moreover, compared with the traditional manual method, the traceability efficiency was improved by more than 80%. Full article
(This article belongs to the Special Issue Design, Optimization and Control Strategy of Smart Grids)
Show Figures

Figure 1

24 pages, 2469 KiB  
Article
Generative and Contrastive Self-Supervised Learning for Virulence Factor Identification Based on Protein–Protein Interaction Networks
by Yalin Yao, Hao Chen, Jianxin Wang and Yeru Wang
Microorganisms 2025, 13(7), 1635; https://doi.org/10.3390/microorganisms13071635 - 10 Jul 2025
Viewed by 334
Abstract
Virulence factors (VFs), produced by pathogens, facilitate pathogenic microorganisms to invade, colonize, and damage the host cells. Accurate VF identification advances pathogenic mechanism understanding and provides novel anti-virulence targets. Existing models primarily utilize protein sequence features while overlooking the systematic protein–protein interaction (PPI) [...] Read more.
Virulence factors (VFs), produced by pathogens, facilitate pathogenic microorganisms to invade, colonize, and damage the host cells. Accurate VF identification advances pathogenic mechanism understanding and provides novel anti-virulence targets. Existing models primarily utilize protein sequence features while overlooking the systematic protein–protein interaction (PPI) information, despite pathogenesis typically resulting from coordinated protein–protein actions. Moreover, a severe imbalance exists between virulence and non-virulence proteins, which causes existing models trained on balanced datasets by sampling to fail in incorporating proteins’ inherent distributional characteristics, thus restricting generalization to real-world imbalanced data. To address these challenges, we propose a novel Generative and Contrastive self-supervised learning framework for Virulence Factor identification (GC-VF) that transforms VF identification into an imbalanced node classification task on graphs generated from PPI networks. The framework encompasses two core modules: the generative attribute reconstruction module learns attribute space representations via feature reconstruction, capturing intrinsic data patterns and reducing noise; the local contrastive learning module employs node-level contrastive learning to precisely capture local features and contextual information, avoiding global aggregation losses while ensuring node representations truly reflect inherent characteristics. Comprehensive benchmark experiments demonstrate that GC-VF outperforms baseline methods on naturally imbalanced datasets, exhibiting higher accuracy and stability, as well as providing a potential solution for accurate VF identification. Full article
(This article belongs to the Section Molecular Microbiology and Immunology)
Show Figures

Figure 1

24 pages, 6164 KiB  
Article
Transformer–GCN Fusion Framework for Mineral Prospectivity Mapping: A Geospatial Deep Learning Approach
by Le Gao, Gnanachandrasamy Gopalakrishnan, Adel Nasri, Youhong Li, Yuying Zhang, Xiaoying Ou and Kele Xia
Minerals 2025, 15(7), 711; https://doi.org/10.3390/min15070711 - 3 Jul 2025
Viewed by 408
Abstract
Mineral prospectivity mapping (MPM) is a pivotal technique in geoscientific mineral resource exploration. To address three critical challenges in current deep convolutional neural network applications for geoscientific mineral resource prediction—(1) model bias induced by imbalanced distribution of ore deposit samples, (2) deficiency in [...] Read more.
Mineral prospectivity mapping (MPM) is a pivotal technique in geoscientific mineral resource exploration. To address three critical challenges in current deep convolutional neural network applications for geoscientific mineral resource prediction—(1) model bias induced by imbalanced distribution of ore deposit samples, (2) deficiency in global feature extraction due to excessive reliance on local spatial correlations, and (3) diminished discriminative capability caused by feature smoothing in deep networks—this study innovatively proposes a T-GCN model integrating Transformer with graph convolutional neural networks (GCNs). The model achieves breakthrough performance through three key technological innovations: firstly, constructing a global perceptual field via Transformer’s self-attention mechanism to effectively capture long-range geological relationships; secondly, combining GCNs’ advantages in topological feature extraction to realize multi-scale feature fusion; and thirdly, designing a feature enhancement module to mitigate deep network degradation. In practical application to the PangXD ore district, the T-GCN model achieved a prediction accuracy of 97.27%, representing a 3.76 percentage point improvement over the best comparative model, and successfully identified five prospective mineralization zones, demonstrating its superior performance and application value under complex geological conditions. Full article
Show Figures

Figure 1

29 pages, 6210 KiB  
Article
GT-STAFG: Graph Transformer with Spatiotemporal Attention Fusion Gate for Epileptic Seizure Detection in Imbalanced EEG Data
by Mohamed Sami Nafea and Zool Hilmi Ismail
AI 2025, 6(6), 120; https://doi.org/10.3390/ai6060120 - 9 Jun 2025
Viewed by 774
Abstract
Background: Electroencephalography (EEG) assists clinicians in diagnosing epileptic seizures by recording brain electrical activity. Existing models process spatiotemporal features inefficiently either through cascaded spatiotemporal architectures or static functional connectivity, limiting their ability to capture deeper spatial–temporal correlations. Objectives: To address these limitations, we [...] Read more.
Background: Electroencephalography (EEG) assists clinicians in diagnosing epileptic seizures by recording brain electrical activity. Existing models process spatiotemporal features inefficiently either through cascaded spatiotemporal architectures or static functional connectivity, limiting their ability to capture deeper spatial–temporal correlations. Objectives: To address these limitations, we propose a Graph Transformer with Spatiotemporal Attention Fusion Gate (GT-STAFG). Methods: We analyzed 18-channel EEG data sampled at 200 Hz, transformed into the frequency domain, and segmented into 30- second windows. The graph transformer exploits dynamic graph data, while STAFG leverages self-attention and gating mechanisms to capture complex interactions by augmenting graph features with both spatial and temporal information. The clinical significance of extracted features was validated using the Integrated Gradients attribution method, emphasizing the clinical relevance of the proposed model. Results: GT-STAFG achieves the highest area under the precision–recall curve (AUPRC) scores of 0.605 on the TUSZ dataset and 0.498 on the CHB-MIT dataset, surpassing baseline models and demonstrating strong cross-patient generalization on imbalanced datasets. We applied transfer learning to leverage knowledge from the TUSZ dataset when analyzing the CHB-MIT dataset, yielding an average improvement of 8.3 percentage points in AUPRC. Conclusions: Our approach has the potential to enhance patient outcomes and optimize healthcare utilization. Full article
Show Figures

Figure 1

21 pages, 1572 KiB  
Article
OWNC: Open-World Node Classification on Graphs with a Dual-Embedding Interaction Framework
by Yuli Chen and Chun Wang
Mathematics 2025, 13(9), 1475; https://doi.org/10.3390/math13091475 - 30 Apr 2025
Viewed by 352
Abstract
Traditional node classification is typically conducted in a closed-world setting, where all labels are known during training, enabling graph neural network methods to achieve high performance. However, in real-world scenarios, the constant emergence of new categories and updates to existing labels can result [...] Read more.
Traditional node classification is typically conducted in a closed-world setting, where all labels are known during training, enabling graph neural network methods to achieve high performance. However, in real-world scenarios, the constant emergence of new categories and updates to existing labels can result in some nodes no longer fitting into any known category, rendering closed-world classification methods inadequate. Thus, open-world classification becomes essential for graph data. Due to the inherent diversity of graph data in the open-world setting, it is common for the number of nodes with different labels to be imbalanced, yet current models are ineffective at handling such imbalance. Additionally, when there are too many or too few nodes from unseen classes, classification performance typically declines. Motivated by these observations, we propose a solution to address the challenges of open-world node classification and introduce a model named OWNC. This model incorporates a dual-embedding interaction training framework and a generator–discriminator architecture. The dual-embedding interaction training framework reduces label loss and enhances the distinction between known and unseen samples, while the generator–discriminator structure enhances the model’s ability to identify nodes from unseen classes. Experimental results on three benchmark datasets demonstrate the superior performance of our model compared to various baseline algorithms, while ablation studies validate the underlying mechanisms and robustness of our approach. Full article
Show Figures

Figure 1

20 pages, 1902 KiB  
Article
Distantly Supervised Relation Extraction Method Based on Multi-Level Hierarchical Attention
by Zhaoxin Xuan, Hejing Zhao, Xin Li and Ziqi Chen
Information 2025, 16(5), 364; https://doi.org/10.3390/info16050364 - 29 Apr 2025
Viewed by 412
Abstract
Distantly Supervised Relation Extraction (DSRE) aims to automatically identify semantic relationships within large text corpora by aligning with external knowledge bases. Despite the success of current methods in automating data annotation, they introduce two main challenges: label noise and data long-tail distribution. Label [...] Read more.
Distantly Supervised Relation Extraction (DSRE) aims to automatically identify semantic relationships within large text corpora by aligning with external knowledge bases. Despite the success of current methods in automating data annotation, they introduce two main challenges: label noise and data long-tail distribution. Label noise results in inaccurate annotations, which can undermine the quality of relation extraction. The long-tail problem, on the other hand, leads to an imbalanced model that struggles to extract less frequent, long-tail relations. In this paper, we introduce a novel relation extraction framework based on multi-level hierarchical attention. This approach utilizes Graph Attention Networks (GATs) to model the hierarchical structure of the relations, capturing the semantic dependencies between relation types and generating relation embeddings that reflect the overall hierarchical framework. To improve the classification process, we incorporate a multi-level classification structure guided by hierarchical attention, which enhances the accuracy of both head and tail relation extraction. A local probability constraint is introduced to ensure coherence across the classification levels, fostering knowledge transfer from frequent to less frequent relations. Experimental evaluations on the New York Times (NYT) dataset demonstrate that our method outperforms existing baselines, particularly in the context of long-tail relation extraction, offering a comprehensive solution to the challenges of DSRE. Full article
Show Figures

Figure 1

32 pages, 6398 KiB  
Article
Big Data-Driven Distributed Machine Learning for Scalable Credit Card Fraud Detection Using PySpark, XGBoost, and CatBoost
by Leonidas Theodorakopoulos, Alexandra Theodoropoulou, Anastasios Tsimakis and Constantinos Halkiopoulos
Electronics 2025, 14(9), 1754; https://doi.org/10.3390/electronics14091754 - 25 Apr 2025
Cited by 1 | Viewed by 2541
Abstract
This study presents an optimization for a distributed machine learning framework to achieve credit card fraud detection scalability. Due to the growth in fraudulent activities, this research implements the PySpark-based processing of large-scale transaction datasets, integrating advanced machine learning models: Logistic Regression, Decision [...] Read more.
This study presents an optimization for a distributed machine learning framework to achieve credit card fraud detection scalability. Due to the growth in fraudulent activities, this research implements the PySpark-based processing of large-scale transaction datasets, integrating advanced machine learning models: Logistic Regression, Decision Trees, Random Forests, XGBoost, and CatBoost. These have been evaluated in terms of scalability, accuracy, and handling imbalanced datasets. Key findings: Among the most promising models for complex and imbalanced data, XGBoost and CatBoost promise close-to-ideal accuracy rates in fraudulent transaction detection. PySpark will be instrumental in scaling these systems to enable them to perform distributed processing, real-time analysis, and adaptive learning. This study further discusses challenges like overfitting, data access, and real-time implementation with potential solutions such as ensemble methods, intelligent sampling, and graph-based approaches. Future directions are underlined by deploying these frameworks in live transaction environments, leveraging continuous learning mechanisms, and integrating advanced anomaly detection techniques to handle evolving fraud patterns. The present research demonstrates the importance of distributed machine learning frameworks for developing robust, scalable, and efficient fraud detection systems, considering their significant impact on financial security and the overall financial ecosystem. Full article
(This article belongs to the Special Issue New Advances in Cloud Computing and Its Latest Applications)
Show Figures

Figure 1

22 pages, 4154 KiB  
Article
Distributed Secure Economic Dispatch Strategy Based on Robust Graph Theory and W-MSR Algorithm
by Jian Le, Jing Wang, Hongke Lang and Weihao Wang
Sensors 2025, 25(8), 2551; https://doi.org/10.3390/s25082551 - 17 Apr 2025
Viewed by 413
Abstract
The traditional consensus-based distributed economic dispatch strategy may lose system convergency and cause imbalanced power when facing an information attack on the individual power generation unit; thus, it is unable to achieve the dispatching goal. Taking into consideration several kinds of attack behaviors [...] Read more.
The traditional consensus-based distributed economic dispatch strategy may lose system convergency and cause imbalanced power when facing an information attack on the individual power generation unit; thus, it is unable to achieve the dispatching goal. Taking into consideration several kinds of attack behaviors that may exist in a distributed control system, this paper develops models of node attacks from the two aspects of action mode and deployment scope, and analyzes the influence of attack behaviors on the distributed economic dispatch system. Based on the idea of the W-MSR algorithm that deletes the information received from nodes that may be attacked, a distributed security consensus-based economic dispatch strategy is synthetized with the incremental cost of the power generation unit as the consensus variable. Based on the graph robustness index, this paper gives its conditions along with its proof that the communication network topology of the system should be satisfied when adopting the W-MSR algorithm. The simulation results of the IEEE-39 bus distribution network show that the strategy proposed in this paper can effectively counter various information attacks, enhancing both the security and economic efficiency of the distributed economic dispatch system. In addition, the (F + 1, F + 1)-robust graph is a necessary and sufficient condition to achieve the consensus of the dispatch strategy. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

23 pages, 1178 KiB  
Article
A Novel Methodology to Develop Mining Stope Stability Graphs on Imbalanced Datasets Using Probabilistic Approaches
by Lucas de Almeida Gama Paixao, William Pratt Rogers and Erisvaldo Bitencourt de Jesus
Mining 2025, 5(2), 24; https://doi.org/10.3390/mining5020024 - 30 Mar 2025
Viewed by 453
Abstract
Predicting and analyzing the stability of underground stopes is critical for ensuring worker safety, reducing dilution, and maintaining operational efficiency in mining. Traditional stability graphs are widely used but often criticized for oversimplifying the stability phenomenon and relying on subjective classifications. Additionally, the [...] Read more.
Predicting and analyzing the stability of underground stopes is critical for ensuring worker safety, reducing dilution, and maintaining operational efficiency in mining. Traditional stability graphs are widely used but often criticized for oversimplifying the stability phenomenon and relying on subjective classifications. Additionally, the imbalanced nature of stope stability datasets poses challenges for traditional machine learning and statistical models, which often bias predictions toward the majority class. This study proposes a novel methodology for developing site-specific stability graphs using probabilistic modeling and machine learning techniques, addressing the limitations of traditional graphs and the challenges of imbalanced datasets. The approach includes rebalancing of the dataset using the Synthetic Minority Over-Sampling Technique (SMOTE) and feature selection using permutation importance to identify key features that impact instability, using those to construct a bi-dimensional stability graph that provides both improved performance and interpretability. The results indicate that the proposed graph outperforms traditional stability graphs, particularly in identifying unstable stopes, even under highly imbalanced data conditions, highlighting the importance of operational and geometric variables in stope stability, providing actionable insights for mine planners. Conclusively, this study demonstrates the potential for integrating modern probabilistic techniques into mining geotechnics, paving the way for more accurate and adaptive stability assessment tools. Future work includes extending the methodology to multi-mine datasets and exploring dynamic stability graph frameworks. Full article
Show Figures

Figure 1

16 pages, 5234 KiB  
Article
Edge and Node Enhancement Graph Convolutional Network: Imbalanced Graph Node Classification Method Based on Edge-Node Collaborative Enhancement
by Jiadong Tian, Jiali Lin and Dagang Li
Mathematics 2025, 13(7), 1038; https://doi.org/10.3390/math13071038 - 22 Mar 2025
Viewed by 596
Abstract
In addressing the issue of node classification with imbalanced data distribution, traditional models exhibit significant limitations. Conventional improvement methods, such as node replication or weight adjustment, often focus solely on nodes, neglecting connection relationships. However, numerous studies have demonstrated that optimizing edge distribution [...] Read more.
In addressing the issue of node classification with imbalanced data distribution, traditional models exhibit significant limitations. Conventional improvement methods, such as node replication or weight adjustment, often focus solely on nodes, neglecting connection relationships. However, numerous studies have demonstrated that optimizing edge distribution can improve the quality of node embeddings. In this paper, we propose the Edge and Node Collaborative Enhancement method (ENE-GCN). This method identifies potentially associated node pairs by similarity measures and constructs a hybrid adjacency matrix, which enlarges the fitting space of node embedding. Subsequently, an adversarial generation strategy is employed to augment the minority class nodes, thereby constructing a balanced sample set. Compared to existing methods, our approach achieves collaborative enhancement of both edges and nodes in a concise manner, improving embedding quality and balancing the training scenario. Experimental comparisons on four public graph datasets reveal that, compared to baseline methods, our proposed method achieves notable improvements in Recall and AUC metrics, particularly in sparsely connected datasets. Full article
Show Figures

Figure 1

36 pages, 1597 KiB  
Article
Analysis of Approximation Methods of Laplacian Eigenvectors of the Kronecker Product of Graphs
by Marko Miladinović, Milan Bašić and Aleksandar Stamenković
Axioms 2025, 14(3), 192; https://doi.org/10.3390/axioms14030192 - 5 Mar 2025
Viewed by 578
Abstract
This paper analyzes two approximation methods for the Laplacian eigenvectors of the Kronecker product, as recently presented in the literature. We enhance the approximations by comparing the correlation coefficients of the eigenvectors, which indicate how well an arbitrary vector approximates a matrix’s eigenvector. [...] Read more.
This paper analyzes two approximation methods for the Laplacian eigenvectors of the Kronecker product, as recently presented in the literature. We enhance the approximations by comparing the correlation coefficients of the eigenvectors, which indicate how well an arbitrary vector approximates a matrix’s eigenvector. In the first method, some correlation coefficients are explicitly calculable, while others are not. In the second method, only certain coefficients can be estimated with good accuracy, as supported by empirical and theoretical evidence, with the rest remaining incalculable. The primary objective is to evaluate the accuracy of the approximation methods by analyzing and comparing limited sets of coefficients on one hand and the estimation on the other. Therefore, we compute the extreme values of the mentioned sets and theoretically compare them. Our observations indicate that, in most cases, the relationship between the majority of the values in the first set and those in the second set reflects the relationship between the remaining coefficients of both approximations. Moreover, it can be observed that each of the sets generally contains smaller values compared to the values found among the remaining correlation coefficients. Finally, we find that the performance of the two approximation methods is significantly influenced by imbalanced graph structures, exemplified by a class of almost regular graphs discussed in the paper. Full article
(This article belongs to the Section Mathematical Analysis)
Show Figures

Figure 1

17 pages, 1516 KiB  
Article
Improving Identification of Drug-Target Binding Sites Based on Structures of Targets Using Residual Graph Transformer Network
by Shuang-Qing Lv, Xin Zeng, Guang-Peng Su, Wen-Feng Du, Yi Li and Meng-Liang Wen
Biomolecules 2025, 15(2), 221; https://doi.org/10.3390/biom15020221 - 3 Feb 2025
Viewed by 1111
Abstract
Improving identification of drug-target binding sites can significantly aid in drug screening and design, thereby accelerating the drug development process. However, due to challenges such as insufficient fusion of multimodal information from targets and imbalanced datasets, enhancing the performance of drug-target binding sites [...] Read more.
Improving identification of drug-target binding sites can significantly aid in drug screening and design, thereby accelerating the drug development process. However, due to challenges such as insufficient fusion of multimodal information from targets and imbalanced datasets, enhancing the performance of drug-target binding sites prediction models remains exceptionally difficult. Leveraging structures of targets, we proposed a novel deep learning framework, RGTsite, which employed a Residual Graph Transformer Network to improve the identification of drug-target binding sites. First, a residual 1D convolutional neural network (1D-CNN) and the pre-trained model ProtT5 were employed to extract the local and global sequence features from the target, respectively. These features were then combined with the physicochemical properties of amino acid residues to serve as the vertex features in graph. Next, the edge features were incorporated, and the residual graph transformer network (GTN) was applied to extract the more comprehensive vertex features. Finally, a fully connected network was used to classify whether the vertex was a binding site. Experimental results showed that RGTsite outperformed the existing state-of-the-art methods in key evaluation metrics, such as F1-score (F1) and Matthews Correlation Coefficient (MCC), across multiple benchmark datasets. Additionally, we conducted interpretability analysis for RGTsite through the real-world cases, and the results confirmed that RGTsite can effectively identify drug-target binding sites in practical applications. Full article
(This article belongs to the Section Bioinformatics and Systems Biology)
Show Figures

Figure 1

21 pages, 2525 KiB  
Article
A Data-Driven Deep Learning Framework for Prediction of Traffic Crashes at Road Intersections
by Mengxiang Wang, Wang-Chien Lee, Na Liu, Qiang Fu, Fujun Wan and Ge Yu
Appl. Sci. 2025, 15(2), 752; https://doi.org/10.3390/app15020752 - 14 Jan 2025
Cited by 1 | Viewed by 2002
Abstract
Traffic crash prediction (TCP) is a fundamental problem for intelligent transportation systems in smart cities. Improving the accuracy of traffic crash prediction is important for road safety and effective traffic management. Owing to recent advances in artificial neural networks, several new deep-learning models [...] Read more.
Traffic crash prediction (TCP) is a fundamental problem for intelligent transportation systems in smart cities. Improving the accuracy of traffic crash prediction is important for road safety and effective traffic management. Owing to recent advances in artificial neural networks, several new deep-learning models have been proposed for TCP. However, these works mainly focus on accidents in regions, which are typically pre-determined using a grid map. We argue that TCP for roads, especially for crashes at or near road intersections which account for more than 50% of the fatal or injury crashes based on the Federal Highway Administration, has a significant practical and research value and thus deserves more research. In this paper, we formulate TCP at Road Intersections as a classification problem and propose a three-phase data-driven deep learning model, called Road Intersection Traffic Crash Prediction (RoadInTCP), to predict traffic crashes at intersections by exploiting publicly available heterogeneous big data. In Phase I we extract discriminative latent features called topological-relational features (tr-features), of intersections using a neural network model by exploiting topological information of the road network and various relationships amongst nearby intersections. In Phase II, in addition to tr-features which capture some inherent properties of the road network, we also explore additional thematic information in terms of environmental, traffic, weather, risk, and calendar features associated with intersections. In order to incorporate the potential correlation in nearby intersections, we utilize a Graph Convolution Network (GCN) to aggregate features from neighboring intersections based on a message-passing paradigm for TCP. While Phase II serves well as a TCP model, we further explore the signals embedded in the sequential feature changes over time for TCP in Phase III, by exploring RNN or 1DCNN which have known success on sequential data. Additionally, to address the serious issues of imbalanced classes in TCP and large-scale heterogeneous big data, we propose an effective data sampling approach in data preparation to facilitate model training. We evaluate the proposed RoadInTCP model via extensive experiments on a real-world New York City traffic dataset. The experimental results show that the proposed RoadInTCP robustly outperforms existing methods. Full article
Show Figures

Figure 1

31 pages, 6359 KiB  
Article
Adaptive Tip Selection for DAG-Shard-Based Federated Learning with High Concurrency and Fairness
by Ruiqi Xiao, Yun Cao and Bin Xia
Sensors 2025, 25(1), 19; https://doi.org/10.3390/s25010019 - 24 Dec 2024
Viewed by 990
Abstract
To cope with the challenges posed by high-concurrency training tasks involving large models and big data, Directed Acyclic Graph (DAG) and shard were proposed as alternatives to blockchain-based federated learning, aiming to enhance training concurrency. However, there is insufficient research on the specific [...] Read more.
To cope with the challenges posed by high-concurrency training tasks involving large models and big data, Directed Acyclic Graph (DAG) and shard were proposed as alternatives to blockchain-based federated learning, aiming to enhance training concurrency. However, there is insufficient research on the specific consensus designs and the effects of varying shard sizes on federated learning. In this paper, we combine DAG and shard by designing three tip selection consensus algorithms and propose an adaptive algorithm to improve training performance. Additionally, we achieve concurrent control over the scale of the directed acyclic graph’s structure through shard and algorithm adjustments. Finally, we validate the fairness of our model with an incentive mechanism and its robustness under different real-world conditions and demonstrate DAG-Shard-based Federated Learning (DSFL)’s advantages in high concurrency and fairness while adjusting the DAG size through concurrency control. In concurrency, DSFL improves accuracy by 8.19–12.21% and F1 score by 7.27–11.73% compared to DAG-FL. Compared to Blockchain-FL, DSFL shows an accuracy gain of 7.82–11.86% and an F1 score improvement of 8.89–13.27%. Additionally, DSFL outperforms DAG-FL and Chains-FL on both balanced and imbalanced datasets. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Back to TopTop