Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,385)

Search Parameters:
Keywords = hybrid deep networks

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 1793 KB  
Article
Multisource POI-Matching Method Based on Deep Learning and Feature Fusion
by Yazhou Ding, Qi Tian, Yun Han, Cailin Li, Yue Wang and Baoyun Guo
Appl. Sci. 2026, 16(2), 796; https://doi.org/10.3390/app16020796 (registering DOI) - 13 Jan 2026
Abstract
In the fields of geographic information science and location-based services, the fusion of multisource Point-of-Interest (POI) data is of remarkable importance but faces several challenges. Existing matching methods, including those based on single non-spatial attributes, single spatial geometric features, and traditional hybrid methods [...] Read more.
In the fields of geographic information science and location-based services, the fusion of multisource Point-of-Interest (POI) data is of remarkable importance but faces several challenges. Existing matching methods, including those based on single non-spatial attributes, single spatial geometric features, and traditional hybrid methods with fixed rules, suffer from limitations such as reliance on a single feature and inadequate consideration of spatial context. This study takes Dongcheng District, Beijing, as the research area and proposes a POI-matching method based on multi-feature value calculation and a deep neural network (DNN) model. The method comprehensively incorporates multidimensional features such as names, addresses, and spatial distances. Additionally, the approach also incorporates an improved multilevel name association strategy, an address similarity calculation using weighted edit distance, and a spatial distance model that accounts for spatial density and regional functional types. Furthermore, the method utilizes a deep learning model to automatically learn POI entity features and optimize the matching rules. Experimental results show that the precision, recall, and F1 value of the proposed method achieved 97.2%, 97.0%, and 0.971, respectively, notably outperforming traditional methods. Overall, this method provides an efficient and reliable solution for geospatial data integration and POI applications, and offers strong support for GIS optimization, smart city construction, and scientific urban/town planning. However, this method still has room for improvement in terms of data source quality and algorithm optimization. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

23 pages, 1961 KB  
Article
Quantum-Resilient Federated Learning for Multi-Layer Cyber Anomaly Detection in UAV Systems
by Canan Batur Şahin
Sensors 2026, 26(2), 509; https://doi.org/10.3390/s26020509 - 12 Jan 2026
Abstract
Unmanned Aerial Vehicles (UAVs) are increasingly used in civilian and military applications, making their communication and control systems targets for cyber attacks. The emerging threat of quantum computing amplifies these risks. Quantum computers could break the classical cryptographic schemes used in current UAV [...] Read more.
Unmanned Aerial Vehicles (UAVs) are increasingly used in civilian and military applications, making their communication and control systems targets for cyber attacks. The emerging threat of quantum computing amplifies these risks. Quantum computers could break the classical cryptographic schemes used in current UAV networks. This situation underscores the need for quantum-resilient, privacy-preserving security frameworks. This paper proposes a quantum-resilient federated learning framework for multi-layer cyber anomaly detection in UAV systems. The framework combines a hybrid deep learning architecture. A Variational Autoencoder (VAE) performs unsupervised anomaly detection. A neural network classifier enables multi-class attack categorization. To protect sensitive UAV data, model training is conducted using federated learning with differential privacy. Robustness against malicious participants is ensured through Byzantine-robust aggregation. Additionally, CRYSTALS-Dilithium post-quantum digital signatures are employed to authenticate model updates and provide long-term cryptographic security. Researchers evaluated the proposed framework on a real UAV attack dataset containing GPS spoofing, GPS jamming, denial-of-service, and simulated attack scenarios. Experimental results show the system achieves 98.67% detection accuracy with only 6.8% computational overhead compared to classical cryptographic approaches, while maintaining high robustness under Byzantine attacks. The main contributions of this study are: (1) a hybrid VAE–classifier architecture enabling both zero-day anomaly detection and precise attack classification, (2) the integration of Byzantine-robust and privacy-preserving federated learning for UAV security, and (3) a practical post-quantum security design validated on real UAV communication data. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

29 pages, 2009 KB  
Article
GCN-Embedding Swin–Unet for Forest Remote Sensing Image Semantic Segmentation
by Pingbo Liu, Gui Zhang and Jianzhong Li
Remote Sens. 2026, 18(2), 242; https://doi.org/10.3390/rs18020242 - 12 Jan 2026
Abstract
Forest resources are among the most important ecosystems on the earth. The semantic segmentation and accurate positioning of ground objects in forest remote sensing (RS) imagery are crucial to the emergency treatment of forest natural disasters, especially forest fires. Currently, most existing methods [...] Read more.
Forest resources are among the most important ecosystems on the earth. The semantic segmentation and accurate positioning of ground objects in forest remote sensing (RS) imagery are crucial to the emergency treatment of forest natural disasters, especially forest fires. Currently, most existing methods for image semantic segmentation are built upon convolutional neural networks (CNNs). Nevertheless, these techniques face difficulties in directly accessing global contextual information and accurately detecting geometric transformations within the image’s target regions. This limitation stems from the inherent locality of convolution operations, which are restricted to processing data structured in Euclidean space and confined to square-shaped regions. Inspired by the graph convolution network (GCN) with robust capabilities in processing irregular and complex targets, as well as Swin Transformers renowned for exceptional global context modeling, we present a hybrid semantic segmentation framework for forest RS imagery termed GSwin–Unet. This framework embeds the GCN model into Swin–Unet architecture to address the issue of low semantic segmentation accuracy of RS imagery in forest scenarios, which is caused by the complex texture features, diverse shapes, and unclear boundaries of land objects. GSwin–Unet features a parallel dual-encoder architecture of GCN and Swin Transformer. First, we integrate the Zero-DCE (Zero-Reference Deep Curve Estimation) algorithm into GSwin–Unet to enhance forest RS image feature representation. Second, a feature aggregation module (FAM) is proposed to bridge the dual encoders by fusing GCN-derived local aggregated features with Swin Transformer-extracted features. Our study demonstrates that, compared with the baseline models TransUnet, Swin–Unet, Unet, and DeepLab V3+, the GSwin–Unet achieves improvements of 7.07%, 5.12%, 8.94%, and 2.69% in the mean Intersection over Union (MIoU) and 3.19%, 1.72%, 4.3%, and 3.69% in the average F1 score (Ave.F1), respectively, on the RGB forest RS dataset. On the NIRGB forest RS dataset, the improvements in MIoU are 5.75%, 3.38%, 6.79%, and 2.44%, and the improvements in Ave.F1 are 4.02%, 2.38%, 4.72%, and 1.67%, respectively. Meanwhile, GSwin–Unet shows excellent adaptability on the selected GID dataset with high forest coverage, where the MIoU and Ave.F1 reach 72.92% and 84.3%, respectively. Full article
22 pages, 3736 KB  
Article
Optimized Hybrid Deep Learning Framework for Reliable Multi-Horizon Photovoltaic Power Forecasting in Smart Grids
by Bilali Boureima Cisse, Ghamgeen Izat Rashed, Ansumana Badjan, Hussain Haider, Hashim Ali I. Gony and Ali Md Ershad
Electricity 2026, 7(1), 4; https://doi.org/10.3390/electricity7010004 - 12 Jan 2026
Abstract
Accurate short-term forecasting of photovoltaic (PV) output is critical to managing the variability of PV generation and ensuring reliable grid operation with high renewable integration. We propose an enhanced hybrid deep learning framework that combines Temporal Convolutional Networks (TCNs), Gated Recurrent Units (GRUs), [...] Read more.
Accurate short-term forecasting of photovoltaic (PV) output is critical to managing the variability of PV generation and ensuring reliable grid operation with high renewable integration. We propose an enhanced hybrid deep learning framework that combines Temporal Convolutional Networks (TCNs), Gated Recurrent Units (GRUs), and Random Forests (RFs) in an optimized weighted ensemble strategy. This approach leverages the complementary strengths of each component: TCNs capture long-range temporal dependencies via dilated causal convolutions; GRUs model sequential weather-driven dynamics; and RFs enhance robustness to outliers and nonlinear relationships. The model was evaluated on high-resolution operational data from the Yulara solar plant in Australia, forecasting horizons from 5 min to 1 h. Results show that the TCN-GRU-RF model consistently outperforms conventional benchmarks, achieving R2 = 0.9807 (MAE = 0.0136; RMSE = 0.0300) at 5 min and R2 = 0.9047 (RMSE = 0.0652) at 1 h horizons. Notably, the degradation in R2 across forecasting horizons was limited to 7.7%, significantly lower than the typical 10–15% range observed in the literature, highlighting the model’s scalability and resilience. These validated results indicate that the proposed approach provides a robust, scalable forecasting solution that enhances grid reliability and supports the integration of distributed renewable energy sources. Full article
15 pages, 3033 KB  
Article
Comparative Study of Different Algorithms for Human Motion Direction Prediction Based on Multimodal Data
by Hongyu Zhao, Yichi Zhang, Yongtao Chen, Hongkai Zhao, Zhuoran Jiang, Mingwei Cao, Haiqing Yang, Yuhang Ding and Peng Li
Sensors 2026, 26(2), 501; https://doi.org/10.3390/s26020501 - 12 Jan 2026
Abstract
The accurate prediction of human movement direction plays a crucial role in fields such as rehabilitation monitoring, sports science, and intelligent military systems. Based on plantar pressure and inertial sensor data, this study developed a hybrid deep learning model integrating a Convolutional Neural [...] Read more.
The accurate prediction of human movement direction plays a crucial role in fields such as rehabilitation monitoring, sports science, and intelligent military systems. Based on plantar pressure and inertial sensor data, this study developed a hybrid deep learning model integrating a Convolutional Neural Network (CNN) and a Bidirectional Long Short-Term Memory (BiLSTM) network to enable joint spatiotemporal feature learning. Systematic comparative experiments involving four distinct deep learning models—CNN, BiLSTM, CNN-LSTM, and CNN-BiLSTM—were conducted to evaluate their convergence performance and prediction accuracy comprehensively. Results show that the CNN-BiLSTM model outperforms the other three models, achieving the lowest RMSE (0.26) and MAE (0.14) on the test set, with an R2 of 0.86, which indicates superior fitting accuracy and generalization ability. The superior performance of the CNN-BiLSTM model is attributed to its ability to effectively capture local spatial features via CNN and model bidirectional temporal dependencies via BiLSTM, thus demonstrating strong adaptability for complex motion scenarios. This work focuses on the optimization and comparison of deep learning algorithms for spatiotemporal feature extraction, providing a reliable framework for real-time human motion prediction and offering potential applications in intelligent gait analysis, wearable monitoring, and adaptive human–machine interaction. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

18 pages, 1386 KB  
Article
Long-Term and Short-Term Photovoltaic Power Generation Forecasting Using a Multi-Scale Fusion MHA-BiLSTM Model
by Mengkun Li, Letian Sun and Yitian Sun
Energies 2026, 19(2), 363; https://doi.org/10.3390/en19020363 - 12 Jan 2026
Abstract
As the proportion of photovoltaic (PV) power generation continues to increase in power systems, high-precision PV power forecasting has become a critical challenge for smart grid scheduling. Traditional forecasting methods often struggle with accuracy and error propagation, particularly when handling short-term fluctuations and [...] Read more.
As the proportion of photovoltaic (PV) power generation continues to increase in power systems, high-precision PV power forecasting has become a critical challenge for smart grid scheduling. Traditional forecasting methods often struggle with accuracy and error propagation, particularly when handling short-term fluctuations and long-term trends. To address these issues, this paper proposes a multi-time scale forecasting model, MHA-BiLSTM, based on Bidirectional Long Short-Term Memory (BiLSTM) and Multi-Head Attention (MHA). The model combines the short-term dependency modeling ability of BiLSTM with the long-term trend capturing ability of the multi-head attention mechanism, effectively addressing both short-term (within 6 h) and long-term (up to 72 h) dependencies in PV power data. The experimental results on a simulated PV dataset demonstrate that the MHA-BiLSTM model outperforms traditional models such as LSTM, BiLSTM, and Transformer in multiple evaluation metrics (e.g., MSE, RMSE, R2), particularly showing stronger robustness and generalization ability in long-term forecasting tasks. The results prove that MHA-BiLSTM effectively improves the accuracy of both short-term and long-term PV power predictions, providing valuable support for future microgrid scheduling, energy storage optimization, and the development of smart energy systems. Full article
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)
Show Figures

Figure 1

64 pages, 13395 KB  
Review
Low-Cost Malware Detection with Artificial Intelligence on Single Board Computers
by Phil Steadman, Paul Jenkins, Rajkumar Singh Rathore and Chaminda Hewage
Future Internet 2026, 18(1), 46; https://doi.org/10.3390/fi18010046 - 12 Jan 2026
Abstract
The proliferation of Internet of Things (IoT) devices has significantly expanded the threat landscape for malicious software (malware), rendering traditional signature-based detection methods increasingly ineffective in coping with the volume and evolving nature of modern threats. In response, researchers are utilising artificial intelligence [...] Read more.
The proliferation of Internet of Things (IoT) devices has significantly expanded the threat landscape for malicious software (malware), rendering traditional signature-based detection methods increasingly ineffective in coping with the volume and evolving nature of modern threats. In response, researchers are utilising artificial intelligence (AI) for a more dynamic and robust malware detection solution. An innovative approach utilising AI is focusing on image classification techniques to detect malware on resource-constrained Single-Board Computers (SBCs) such as the Raspberry Pi. In this method the conversion of malware binaries into 2D images is examined, which can be analysed by deep learning models such as convolutional neural networks (CNNs) to classify them as benign or malicious. The results show that the image-based approach demonstrates high efficacy, with many studies reporting detection accuracy rates exceeding 98%. That said, there is a significant challenge in deploying these demanding models on devices with limited processing power and memory, in particular those involving of both calculation and time complexity. Overcoming this issue requires critical model optimisation strategies. Successful approaches include the use of a lightweight CNN architecture and federated learning, which may be used to preserve privacy while training models with decentralised data are processed. This hybrid workflow in which models are trained on powerful servers before the learnt algorithms are deployed on SBCs is an emerging field attacting significant interest in the field of cybersecurity. This paper synthesises the current state of the art, performance compromises, and optimisation techniques contributing to the understanding of how AI and image representation can enable effective low-cost malware detection on resource-constrained systems. Full article
Show Figures

Graphical abstract

20 pages, 27179 KB  
Article
Multi-Scale Temporal Learning with EEMD Reconstruction for Non-Stationary Error Forecasting in Current Transformers
by Jian Liu, Chen Hu, Zhenhua Li and Jiuxi Cui
Electronics 2026, 15(2), 325; https://doi.org/10.3390/electronics15020325 - 11 Jan 2026
Viewed by 35
Abstract
Current transformer measurement errors exhibit strong non-stationarity and multi-scale temporal dynamics, which make accurate prediction challenging for conventional deep learning models. This paper presents a hybrid signal processing and temporal learning framework that integrates ensemble empirical mode decomposition (EEMD) with a dual-scale temporal [...] Read more.
Current transformer measurement errors exhibit strong non-stationarity and multi-scale temporal dynamics, which make accurate prediction challenging for conventional deep learning models. This paper presents a hybrid signal processing and temporal learning framework that integrates ensemble empirical mode decomposition (EEMD) with a dual-scale temporal convolutional architecture. EEMD adaptively decomposes the error sequence into intrinsic mode functions, while a Pearson correlation-based selection step removes redundant and noise-dominated components. The refined signal is then processed by a dual-scale temporal convolutional network (TCN) designed with parallel dilated kernels to capture both high-frequency transients and long-range drift patterns. Experimental evaluations on 110 kV substation data confirm that the proposed decomposition-enhanced dual-scale temporal convolutional framework significantly improves generalization and robustness, reducing the root mean square error by 40.9% and the mean absolute error by 37.0% compared with benchmark models. The results demonstrate that combining decomposition-based preprocessing with multi-scale temporal learning effectively enhances the accuracy and stability of non-stationary current transformer error forecasting. Full article
19 pages, 14871 KB  
Article
Deep Q-Network for Maneuver Planning in Beyond-Visual-Range Aerial Pursuit–Evasion with Target Re-Engagement
by Long-Jun Zhu, Kevin W. Tong and Edmond Q. Wu
Aerospace 2026, 13(1), 77; https://doi.org/10.3390/aerospace13010077 - 11 Jan 2026
Viewed by 39
Abstract
Decision-making for maneuvering in the presence of long-range threats is crucial for enhancing the safety and reliability of autonomous aerial platforms operating in beyond-line-of-sight environments. This study employs the Deep Q-Network (DQN) method to investigate maneuvering strategies for simultaneously avoiding incoming high-speed threats [...] Read more.
Decision-making for maneuvering in the presence of long-range threats is crucial for enhancing the safety and reliability of autonomous aerial platforms operating in beyond-line-of-sight environments. This study employs the Deep Q-Network (DQN) method to investigate maneuvering strategies for simultaneously avoiding incoming high-speed threats and re-establishing tracking of a maneuvering target platform. First, kinematic models for the aerial platforms and the approaching interceptor are developed, and a DQN training environment is constructed based on these models. A DQN framework is then designed, integrating scenario-specific state representation, action space, and a hybrid reward structure to enable autonomous strategy learning without prior expert knowledge. The agent is trained within this environment to achieve near-optimal maneuvering decisions, with comparative evaluations against Q-learning and deep deterministic policy gradient (DDPG) baselines. Simulation results demonstrate that the trained model outperforms the baselines on key metrics by effectively avoiding approaching threats, re-establishing robust target tracking, reducing maneuver time, and exhibiting strong generalization across challenging scenarios. This work advances Beyond-Visual-Range (BVR) maneuver planning and provides a foundational methodological framework for future research on complex multi-stage aerial pursuit–evasion problems. Full article
(This article belongs to the Section Aeronautics)
23 pages, 5900 KB  
Article
Hybrid Attention Mechanism Combined with U-Net for Extracting Vascular Branching Points in Intracavitary Images
by Kaiyang Xu, Haibin Wu, Liang Yu and Xin He
Electronics 2026, 15(2), 322; https://doi.org/10.3390/electronics15020322 - 11 Jan 2026
Viewed by 35
Abstract
To address the application requirements of Visual Simultaneous Localization and Mapping (VSLAM) in intracavitary environments and the scarcity of gold-standard datasets for deep learning methods, this study proposes a hybrid attention mechanism combined with U-Net for vascular branch point extraction in endoluminal images [...] Read more.
To address the application requirements of Visual Simultaneous Localization and Mapping (VSLAM) in intracavitary environments and the scarcity of gold-standard datasets for deep learning methods, this study proposes a hybrid attention mechanism combined with U-Net for vascular branch point extraction in endoluminal images (SuperVessel). The network is initialized via transfer learning with pre-trained SuperRetina model parameters and integrated with a vascular feature detection and matching method based on dual branch fusion and structure enhancement, generating a pseudo-gold-standard vascular branch point dataset. The framework employs a dual-decoder architecture, incorporates a dynamic up-sampling module (CBAM-Dysample) to refine local vessel features through hybrid attention mechanisms, designs a Dice-Det loss function weighted by branching features to prioritize vessel junctions, and introduces a dynamically weighted Triplet-Des loss function optimized for descriptor discrimination. Experiments on the Vivo test set demonstrate that the proposed method achieves an average Area Under Curve (AUC) of 0.760, with mean feature points, accuracy, and repeatability scores of 42,795, 0.5294, and 0.46, respectively. Compared to SuperRetina, the method maintains matching stability while exhibiting superior repeatability, feature point density, and robustness in low-texture/deformation scenarios. Ablation studies confirm the CBAM-Dysample module’s efficacy in enhancing feature expression and convergence speed, offering a robust solution for intracavitary SLAM systems. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

28 pages, 5634 KB  
Article
UCA-Net: A Transformer-Based U-Shaped Underwater Enhancement Network with a Compound Attention Mechanism
by Cheng Yu, Jian Zhou, Lin Wang, Guizhen Liu and Zhongjun Ding
Electronics 2026, 15(2), 318; https://doi.org/10.3390/electronics15020318 - 11 Jan 2026
Viewed by 32
Abstract
Images captured underwater frequently suffer from color casts, blurring, and distortion, which are mainly attributable to the unique optical characteristics of water. Although conventional UIE methods rooted in physics are available, their effectiveness is often constrained, particularly in challenging aquatic and illumination conditions. [...] Read more.
Images captured underwater frequently suffer from color casts, blurring, and distortion, which are mainly attributable to the unique optical characteristics of water. Although conventional UIE methods rooted in physics are available, their effectiveness is often constrained, particularly in challenging aquatic and illumination conditions. More recently, deep learning has become a leading paradigm for UIE, recognized for its superior performance and operational efficiency. This paper proposes UCA-Net, a lightweight CNN-Transformer hybrid network. It incorporates multiple attention mechanisms and utilizes composite attention to effectively enhance textures, reduce blur, and correct color. A novel adaptive sparse self-attention module is introduced to jointly restore global color consistency and fine local details. The model employs a U-shaped encoder–decoder architecture with three-stage up- and down-sampling, facilitating multi-scale feature extraction and global context fusion for high-quality enhancement. Experimental results on multiple public datasets demonstrate UCA-Net’s superior performance, achieving a PSNR of 24.75 dB and an SSIM of 0.89 on the UIEB dataset, while maintaining an extremely low computational cost with only 1.44M parameters. Its effectiveness is further validated by improvements in various downstream image tasks. UCA-Net achieves an optimal balance between performance and efficiency, offering a robust and practical solution for underwater vision applications. Full article
14 pages, 2342 KB  
Article
LSTM-Based Absolute Position Estimation of a 2-DOF Planar Delta Robot Using Time-Series Data
by Seunghwan Baek
Sensors 2026, 26(2), 470; https://doi.org/10.3390/s26020470 - 10 Jan 2026
Viewed by 131
Abstract
Accurately estimating the absolute position of robots under external loads is challenging due to nonlinear dynamics, posture-dependent manipulability, and structural sensitivities. This study investigates a data-driven approach for absolute position prediction of a 2-DOF planar delta robot by learning time-series force signals generated [...] Read more.
Accurately estimating the absolute position of robots under external loads is challenging due to nonlinear dynamics, posture-dependent manipulability, and structural sensitivities. This study investigates a data-driven approach for absolute position prediction of a 2-DOF planar delta robot by learning time-series force signals generated during manipulability-driven free motion. Constant torques of opposite directions were applied to the robot without any position or trajectory control, allowing the mechanism to move naturally according to its configuration-dependent manipulability. Reaction forces measured at the end-effector and relative encoder variations were collected across a grid of workspace locations and used to construct a 12-channel time-series input. A hybrid deep learning architecture combining 1D convolutional layers and a bidirectional LSTM network was trained to regress the robot’s absolute X–Y position. Experimental results demonstrate that the predicted trajectories closely match the measured paths in the workspace, yielding overall RMSE values of 3.81 mm(X) and 2.94 mm(Y). Statistical evaluation using RMSE shows that approximately 83.73% of all test sequences achieve an error below 5 mm. The findings confirm that LSTM models can effectively learn posture-dependent dynamic behavior and force-manipulability relationships. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

29 pages, 1099 KB  
Article
Jump Volatility Forecasting for Crude Oil Futures Based on Complex Network and Hybrid CNN–Transformer Model
by Yuqi He, Po Ning and Yuping Song
Mathematics 2026, 14(2), 258; https://doi.org/10.3390/math14020258 - 9 Jan 2026
Viewed by 70
Abstract
The crude oil futures market is highly susceptible to policy changes and international relations, which often trigger abrupt jumps in prices. The existing literature rarely considers jump volatility and the underlying impact mechanisms. This study proposes a hybrid forecasting model integrating a convolutional [...] Read more.
The crude oil futures market is highly susceptible to policy changes and international relations, which often trigger abrupt jumps in prices. The existing literature rarely considers jump volatility and the underlying impact mechanisms. This study proposes a hybrid forecasting model integrating a convolutional neural network (CNN) and self-attention (Transformer) for high-frequency financial data, based on the complex network characteristics between trading information and multi-market financialization indicators. Empirical results demonstrate that incorporating complex network indicators enhances model performance, with the CNN–Transformer model with a complex network achieving the highest predictive accuracy. Furthermore, we verify the model’s effectiveness and robustness in the WTI crude oil market via Diebold–Mariano tests and external event shock. Notably, this study also extends the analytical framework to jump intensity, thereby providing a more accurate and robust jump forecasting model for risk management and trading strategies in the crude oil futures market. Full article
Show Figures

Figure 1

22 pages, 3809 KB  
Article
Research on Remote Sensing Image Object Segmentation Using a Hybrid Multi-Attention Mechanism
by Lei Chen, Changliang Li, Yixuan Gao, Yujie Chang, Siming Jin, Zhipeng Wang, Xiaoping Ma and Limin Jia
Appl. Sci. 2026, 16(2), 695; https://doi.org/10.3390/app16020695 - 9 Jan 2026
Viewed by 86
Abstract
High-resolution remote sensing images are gradually playing an important role in land cover mapping, urban planning, and environmental monitoring tasks. However, current segmentation approaches frequently encounter challenges such as loss of detail and blurred boundaries when processing high-resolution remote sensing imagery, owing to [...] Read more.
High-resolution remote sensing images are gradually playing an important role in land cover mapping, urban planning, and environmental monitoring tasks. However, current segmentation approaches frequently encounter challenges such as loss of detail and blurred boundaries when processing high-resolution remote sensing imagery, owing to their complex backgrounds and dense semantic content. In response to the aforementioned limitations, this study introduces HMA-UNet, a novel segmentation network built upon the UNet framework and enhanced through a hybrid attention strategy. The architecture’s innovation centers on a composite attention block, where a lightweight split fusion attention (LSFA) mechanism and a lightweight channel-spatial attention (LCSA) mechanism are synergistically integrated within a residual learning structure to replace the stacked convolutional structure in UNet, which can improve the utilization of important shallow features and eliminate redundant information interference. Comprehensive experiments on the WHDLD dataset and the DeepGlobe road extraction dataset show that our proposed method achieves effective segmentation in remote sensing images by fully utilizing shallow features and eliminating redundant information interference. The quantitative evaluation results demonstrate the performance of the proposed method across two benchmark datasets. On the WHDLD dataset, the model attains a mean accuracy, IoU, precision, and recall of 72.40%, 60.71%, 75.46%, and 72.41%, respectively. Correspondingly, on the DeepGlobe road extraction dataset, it achieves a mean accuracy of 57.87%, an mIoU of 49.82%, a mean precision of 78.18%, and a mean recall of 57.87%. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

24 pages, 4797 KB  
Article
PRTNet: Combustion State Recognition Model of Municipal Solid Waste Incineration Process Based on Enhanced Res-Transformer and Multi-Scale Feature Guided Aggregation
by Jian Zhang, Junyu Ge and Jian Tang
Sustainability 2026, 18(2), 676; https://doi.org/10.3390/su18020676 - 9 Jan 2026
Viewed by 117
Abstract
Accurate identification of the combustion state in municipal solid waste incineration (MSWI) processes is crucial for achieving efficient, low-emission, and safe operation. However, existing methods often struggle with stable and reliable recognition due to insufficient feature extraction capabilities when confronted with challenges such [...] Read more.
Accurate identification of the combustion state in municipal solid waste incineration (MSWI) processes is crucial for achieving efficient, low-emission, and safe operation. However, existing methods often struggle with stable and reliable recognition due to insufficient feature extraction capabilities when confronted with challenges such as complex flame morphology, blurred boundaries, and significant noise in flame images. To address this, this paper proposes a novel hybrid architecture model named PRTNet, which aims to enhance the accuracy and robustness of combustion state recognition through multi-scale feature enhancement and adaptive fusion mechanisms. First, a local-semantic enhanced residual network is constructed to establish spatial correlations between fine-grained textures and macroscopic combustion patterns. Subsequently, a feature-adaptive fusion Transformer is designed, which models long-range dependencies and high-frequency details in parallel via deformable attention and local convolutions, and achieves adaptive fusion of global and local features through a gating mechanism. Finally, a cross-scale feature guided aggregation module is proposed to fuse shallow detailed information with deep semantic features under dual-attention guidance. Experiments conducted on a flame image dataset from an MSWI plant in Beijing show that PRTNet achieves an accuracy of 96.29% in the combustion state classification task, with precision, recall, and F1-score all exceeding 96%, significantly outperforming numerous mainstream baseline models. Ablation studies further validate the effectiveness and synergistic effects of each module. The proposed method provides a reliable solution for intelligent flame state recognition in complex industrial scenarios, contributing to the advancement of intelligent and sustainable development in municipal solid waste incineration processes. Full article
(This article belongs to the Special Issue Life Cycle and Sustainability Nexus in Solid Waste Management)
Show Figures

Figure 1

Back to TopTop