Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (5,251)

Search Parameters:
Keywords = feature extraction and fusion

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 1666 KB  
Article
Fault Diagnosis Method for Reciprocating Compressors Based on Spatio-Temporal Feature Fusion
by Haibo Xu, Xiaolong Ji, Xiaogang Qin, Weizheng An, Fengli Zhang, Lixiang Duan and Jinjiang Wang
Sensors 2026, 26(3), 798; https://doi.org/10.3390/s26030798 (registering DOI) - 25 Jan 2026
Abstract
Reciprocating compressors, which serve as core equipment in the petrochemical and natural gas transmission sectors, operate under prolonged variable loads and high-frequency impact conditions. Critical components, such as valves and piston rings, are prone to failure. Existing fault diagnosis methods suffer from inadequate [...] Read more.
Reciprocating compressors, which serve as core equipment in the petrochemical and natural gas transmission sectors, operate under prolonged variable loads and high-frequency impact conditions. Critical components, such as valves and piston rings, are prone to failure. Existing fault diagnosis methods suffer from inadequate spatio-temporal feature extraction and neglect spatio-temporal correlations. To address this, this paper proposes a spatio-temporal feature fusion-based fault diagnosis method for reciprocating compressors. This method constructs a spatio-temporal feature fusion model (STFFM) comprising three principal modules: First, a spatio-temporal feature extraction module employing a multi-layered stacked bidirectional gated recurrent unit (BiGRU) with batch normalisation to uncover temporal dependencies in long-term sequence data. A graph structure is constructed via k-nearest neighbours (KNN), and an enhanced graph isomorphism network (GIN) is integrated to capture spatial domain fault information variations. Second, the spatio-temporal bidirectional attention-gated fusion module employs a bidirectional multi-head attention mechanism to enhance temporal and spatial features. It incorporates a cross-modal gated update mechanism and learnable weight parameters to dynamically retain the highly discriminative features. Third, the classification output module enhances the model’s generalisation capability through multi-layer fully connected layers and regularisation design. Research findings demonstrate that this approach effectively integrates spatio-temporal coupled fault features, achieving an average accuracy of 99.14% on an experimental dataset. This provides an effective technical pathway for the precise identification of faults in the critical components of reciprocating compressors. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
28 pages, 5622 KB  
Article
A Multi-Class Bahadur–Lazarsfeld Expansion Framework for Pixel-Level Fusion in Multi-Sensor Land Cover Classification
by Spiros Papadopoulos, Georgia Koukiou and Vassilis Anastassopoulos
Remote Sens. 2026, 18(3), 399; https://doi.org/10.3390/rs18030399 (registering DOI) - 25 Jan 2026
Abstract
In many land cover classification tasks, the limited precision of individual sensors hinders the accurate separation of certain classes, largely due to the complexity of the Earth’s surface morphology. To mitigate these issues, decision fusion methodologies are employed, allowing data from multiple sensors [...] Read more.
In many land cover classification tasks, the limited precision of individual sensors hinders the accurate separation of certain classes, largely due to the complexity of the Earth’s surface morphology. To mitigate these issues, decision fusion methodologies are employed, allowing data from multiple sensors to be synthesized into robust and more conclusive classification outcomes. This study employs fully polarimetric Synthetic Aperture Radar (PolSAR) imagery and leverages the strengths of three decomposition methods, namely Pauli’s, Krogager’s, and Cloude’s, by extracting their respective components for improved detection. From each decomposition method, three scattering components are derived, enabling the extraction of informative features that describe the scattering behavior associated with various land cover types. The extracted scattering features, treated as independent sensors, were used to train three neural network classifiers. The resulting outputs were then considered as local decisions for each land cover type and subsequently fused through a decision fusion rule to generate more complete and accurate classification results. Experimental results demonstrate that the proposed Multi-Class Bahadur–Lazarsfeld Expansion (MC-BLE) fusion significantly enhances classification performance, achieving an overall accuracy (OA) of 95.78% and a Kappa coefficient of 0.94. Compared to individual classification methods, the fusion notably improved per-class accuracy, particularly for complex land cover boundaries. The core innovation of this work is the transformation of the Bahadur–Lazarsfeld Expansion (BLE), originally designed for binary decision fusion into a multi-class framework capable of addressing multiple land cover types, resulting in a more effective and reliable decision fusion strategy. Full article
Show Figures

Figure 1

28 pages, 5166 KB  
Article
Hyperspectral Image Classification Using SIFANet: A Dual-Branch Structure Combining CNN and Transformer
by Yuannan Gui, Lu Xu, Dongping Ming, Yanfei Wei and Ming Huang
Remote Sens. 2026, 18(3), 398; https://doi.org/10.3390/rs18030398 (registering DOI) - 24 Jan 2026
Abstract
The hyperspectral image (HSI) is rich in spectral information and has important applications in the field of ground objects classification. However, HSI data have high dimensions and variable spatial–spectral features, which make it difficult for some models to adequately extract the effective features. [...] Read more.
The hyperspectral image (HSI) is rich in spectral information and has important applications in the field of ground objects classification. However, HSI data have high dimensions and variable spatial–spectral features, which make it difficult for some models to adequately extract the effective features. Recent studies have shown that fusing spatial and spectral features can significantly improve accuracy by exploiting multi-dimensional correlations. Based on this, this article proposes a spectral integration and focused attention network (SIFANet) with a two-branch structure. SIFANet captures the local spatial features and global spectral dependencies through the parallel-designed spatial feature extractor (SFE) and spectral sequence Transformer (SST), respectively. A cross-module attention fusion (CMAF) mechanism dynamically integrates features from both branches before final classification. Experiments on the Salinas dataset and Xiong’an hyperspectral dataset show that the overall accuracy on these two datasets is 99.89% and 99.79%, which is higher than the other models compared. The proposed method also had the lowest standard deviation of category accuracy and optimal computational efficiency metrics, demonstrating robust spatial–spectral feature integration for improved classification. Full article
Show Figures

Figure 1

23 pages, 2066 KB  
Article
Intelligent Attention-Driven Deep Learning for Hip Disease Diagnosis: Fusing Multimodal Imaging and Clinical Text for Enhanced Precision and Early Detection
by Jinming Zhang, He Gong, Pengling Ren, Shuyu Liu, Zhengbin Jia, Lizhen Wang and Yubo Fan
Medicina 2026, 62(2), 250; https://doi.org/10.3390/medicina62020250 (registering DOI) - 24 Jan 2026
Abstract
Background: Hip joint disorders exhibit diverse and overlapping radiological features, complicating early diagnosis and limiting the diagnostic value of single-modality imaging. Isolated imaging or clinical data may therefore inadequately represent disease-specific pathological characteristics. Methods: This retrospective study included 605 hip joints [...] Read more.
Background: Hip joint disorders exhibit diverse and overlapping radiological features, complicating early diagnosis and limiting the diagnostic value of single-modality imaging. Isolated imaging or clinical data may therefore inadequately represent disease-specific pathological characteristics. Methods: This retrospective study included 605 hip joints from Center A (2018–2024), comprising normal hips, osteoarthritis, osteonecrosis of the femoral head (ONFH), and femoroacetabular impingement (FAI). An independent cohort of 24 hips from Center B (2024–2025) was used for external validation. A multimodal deep learning framework was developed to jointly analyze radiographs, CT volumes, and clinical texts. Features were extracted using ResNet50, 3D-ResNet50, and a pretrained BERT model, followed by attention-based fusion for four-class classification. Results: The combined Clinical+X-ray+CT model achieved an AUC of 0.949 on the internal test set, outperforming all single-modality models. Improvements were consistently observed in accuracy, sensitivity, specificity, and decision curve analysis. Grad-CAM visualizations confirmed that the model attended to clinically relevant anatomical regions. Conclusions: Attention-based multimodal feature fusion substantially improves diagnostic performance for hip joint diseases, providing an interpretable and clinically applicable framework for early detection and precise classification in orthopedic imaging. Full article
(This article belongs to the Special Issue Artificial Intelligence in Medicine: Shaping the Future of Healthcare)
Show Figures

Figure 1

24 pages, 7094 KB  
Article
Research on Pilot Workload Identification Based on EEG Time Domain and Frequency Domain
by Weiping Yang, Yixuan Li, Lingbo Liu, Haiqing Si, Haibo Wang, Ting Pan, Yan Zhao and Gen Li
Aerospace 2026, 13(2), 114; https://doi.org/10.3390/aerospace13020114 - 23 Jan 2026
Abstract
Pilot workload is a critical factor influencing flight safety. This study collects both subjective and objective data on pilot workload using the NASA-TLX questionnaire and electroencephalogram acquisition systems during simulated flight tasks. The raw EEG signals are denoised through preprocessing techniques, and relevant [...] Read more.
Pilot workload is a critical factor influencing flight safety. This study collects both subjective and objective data on pilot workload using the NASA-TLX questionnaire and electroencephalogram acquisition systems during simulated flight tasks. The raw EEG signals are denoised through preprocessing techniques, and relevant EEG features are extracted using time-domain and frequency-domain analysis methods. One-way ANOVA is employed to examine the statistical differences in EEG indicators under varying workload levels. A fusion model based on CNN-Bi-LSTM is developed to train and classify the extracted EEG features, enabling accurate identification of pilot workload states. The results demonstrate that the proposed hybrid model achieves a recognition accuracy of 98.2% on the test set, confirming its robustness. Additionally, under increased workload conditions, frequency-domain features outperform time-domain features in discriminative power. The model proposed in this study effectively recognizes pilot workload levels and offers valuable insights for civil aviation safety management and pilot training programs. Full article
(This article belongs to the Special Issue Human Factors and Performance in Aviation Safety)
19 pages, 7177 KB  
Article
MFF-Net: A Study on Soil Moisture Content Inversion in a Summer Maize Field Based on Multi-Feature Fusion of Leaf Images
by Jianqin Ma, Jiaqi Han, Bifeng Cui, Xiuping Hao, Zhengxiong Bai, Yijian Chen, Yan Zhao and Yu Ding
Agriculture 2026, 16(3), 298; https://doi.org/10.3390/agriculture16030298 - 23 Jan 2026
Abstract
Current agricultural irrigation management practices are often extensive, and traditional soil moisture content (SMC) monitoring methods are inefficient, so there is a pressing need for innovative approaches in precision irrigation. This study proposes a Multi-Feature Fusion Network (MFF-Net) for SMC inversion. The model [...] Read more.
Current agricultural irrigation management practices are often extensive, and traditional soil moisture content (SMC) monitoring methods are inefficient, so there is a pressing need for innovative approaches in precision irrigation. This study proposes a Multi-Feature Fusion Network (MFF-Net) for SMC inversion. The model uses a designed Channel-Changeable Residual Block (ResBlockCC) to construct a multi-branch feature extraction and fusion architecture. Integrating the Channel Squeeze and Spatial Excitation (sSE) attention module with U-Net-like skip connections, MFF-Net inverts root-zone SMC from summer maize leaf images. Field experiments were conducted in Zhengzhou, Henan Province, China, from 2024 to 2025, under three irrigation treatments: 60–70% θfc, 70–90% θfc, and 60–90% θfc (θfc denotes field capacity). This study shows that (1) MFF-Net achieved its smallest inversion error under the 60–70% θfc treatment, suggesting the inversion was most effective when SMC variation was small and relatively low; (2) MFF-Net demonstrated superior performance to several benchmark models, achieving an R2 of 0.84; and (3) the ablation study confirmed that each feature branch and the sSE attention module contributed positively to model performance. MFF-Net thus offers a technological reference for real-time precision irrigation and shows promise for field SMC inversion in summer maize. Full article
(This article belongs to the Section Agricultural Soils)
Show Figures

Figure 1

26 pages, 8183 KB  
Article
MEE-DETR: Multi-Scale Edge-Aware Enhanced Transformer for PCB Defect Detection
by Xiaoyu Ma, Xiaolan Xie and Yuhui Song
Electronics 2026, 15(3), 504; https://doi.org/10.3390/electronics15030504 - 23 Jan 2026
Abstract
Defect inspection of Printed Circuit Board (PCB) is essential for maintaining the safety and reliability of electronic products. With the continuous trend toward smaller components and higher integration levels, identifying tiny imperfections on densely packed PCB structures has become increasingly difficult and remains [...] Read more.
Defect inspection of Printed Circuit Board (PCB) is essential for maintaining the safety and reliability of electronic products. With the continuous trend toward smaller components and higher integration levels, identifying tiny imperfections on densely packed PCB structures has become increasingly difficult and remains a major challenge for current inspection systems. To tackle this problem, this study proposes the Multi-scale Edge-Aware Enhanced Detection Transformer (MEE-DETR), a deep learning-based object detection method. Building upon the RT-DETR framework, which is grounded in Transformer-based machine learning, the proposed approach systematically introduces enhancements at three levels: backbone feature extraction, feature interaction, and multi-scale feature fusion. First, the proposed Edge-Strengthened Backbone Network (ESBN) constructs multi-scale edge extraction and semantic fusion pathways, effectively strengthening the structural representation of shallow defect edges. Second, the Entanglement Transformer Block (ETB), synergistically integrates frequency self-attention, spatial self-attention, and a frequency–spatial entangled feed-forward network, enabling deep cross-domain information interaction and consistent feature representation. Finally, the proposed Adaptive Enhancement Feature Pyramid Network (AEFPN), incorporating the Adaptive Cross-scale Fusion Module (ACFM) for cross-scale adaptive weighting and the Enhanced Feature Extraction C3 Module (EFEC3) for local nonlinear enhancement, substantially improves detail preservation and semantic balance during feature fusion. Experiments conducted on the PKU-Market-PCB dataset reveal that MEE-DETR delivers notable performance gains. Specifically, Precision, Recall, and mAP50–95 improve by 2.5%, 9.4%, and 4.2%, respectively. In addition, the model’s parameter size is reduced by 40.7%. These results collectively indicate that MEE-DETR achieves excellent detection performance with a lightweight network architecture. Full article
20 pages, 3656 KB  
Article
Efficient Model for Detecting Steel Surface Defects Utilizing Dual-Branch Feature Enhancement and Downsampling
by Quan Lu, Minsheng Gong and Linfei Yin
Appl. Sci. 2026, 16(3), 1181; https://doi.org/10.3390/app16031181 - 23 Jan 2026
Abstract
Surface defect evaluation in steel production demands both high inference speed and accuracy for efficient production. However, existing methods face two critical challenges: (1) the diverse dimensions and irregular morphologies of surface defects reduce detection accuracy, and (2) computationally intensive feature extraction slows [...] Read more.
Surface defect evaluation in steel production demands both high inference speed and accuracy for efficient production. However, existing methods face two critical challenges: (1) the diverse dimensions and irregular morphologies of surface defects reduce detection accuracy, and (2) computationally intensive feature extraction slows inference. In response to these challenges, this study proposes an innovative network based on dual-branch feature enhancement and downsampling (DFED-Net). First, an atrous convolution and multi-scale dilated attention fusion module (AMFM) is developed, incorporating local–global feature representation. By emphasizing local details and global semantics, the module suppresses noise interference and enhances the capability of the model to separate small-object features from complex backgrounds. Additionally, a dual-branch downsampling module (DBDM) is developed to preserve the fine details related to scale that are typically lost during downsampling. The DBDM efficiently fuses semantic and detailed information, improving consistency across feature maps at different scales. A lightweight dynamic upsampling (DySample) is introduced to supplant traditional fixed methods with a learnable, adaptive approach, which retains critical feature information more flexibly while reducing redundant computation. Experimental evaluation shows a mean average precision (mAP) of 81.5% on the Northeastern University surface defect detection (NEU-DET) dataset, a 5.2% increase compared to the baseline, while maintaining a real-time inference speed of 120 FPS compared to the 118 FPS of the baseline. The proposed DFED-Net provides strong support for the development of automated visual inspection systems for detecting defects on steel surfaces. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

20 pages, 49658 KB  
Article
Dead Chicken Identification Method Based on a Spatial-Temporal Graph Convolution Network
by Jikang Yang, Chuang Ma, Haikun Zheng, Zhenlong Wu, Xiaohuan Chao, Cheng Fang and Boyi Xiao
Animals 2026, 16(3), 368; https://doi.org/10.3390/ani16030368 - 23 Jan 2026
Abstract
In intensive cage rearing systems, accurate dead hen detection remains difficult due to complex environments, severe occlusion, and the high visual similarity between dead hens and live hens in a prone posture. To address these issues, this study proposes a dead hen identification [...] Read more.
In intensive cage rearing systems, accurate dead hen detection remains difficult due to complex environments, severe occlusion, and the high visual similarity between dead hens and live hens in a prone posture. To address these issues, this study proposes a dead hen identification method based on a Spatial-Temporal Graph Convolutional Network (STGCN). Unlike conventional static image-based approaches, the proposed method introduces temporal information to enable dynamic spatial-temporal modeling of hen health states. First, a multimodal fusion algorithm is applied to visible light and thermal infrared images to strengthen multimodal feature representation. Then, an improved YOLOv7-Pose algorithm is used to extract the skeletal keypoints of individual hens, and the ByteTrack algorithm is employed for multi-object tracking. Based on these results, spatial-temporal graph-structured data of hens are constructed by integrating spatial and temporal dimensions. Finally, a spatial-temporal graph convolution model is used to identify dead hens by learning spatial-temporal dependency features from skeleton sequences. Experimental results show that the improved YOLOv7-Pose model achieves an average precision (AP) of 92.8% in keypoint detection. Based on the constructed spatial-temporal graph data, the dead hen identification model reaches an overall classification accuracy of 99.0%, with an accuracy of 98.9% for the dead hen category. These results demonstrate that the proposed method effectively reduces interference caused by feeder occlusion and ambiguous visual features. By using dynamic spatial-temporal information, the method substantially improves robustness and accuracy of dead hen detection in complex cage rearing environments, providing a new technical route for intelligent monitoring of poultry health status. Full article
(This article belongs to the Special Issue Welfare and Behavior of Laying Hens)
Show Figures

Figure 1

27 pages, 2582 KB  
Article
Intent-Aware Collision Avoidance for UAVs in High-Density Non-Cooperative Environments Using Deep Reinforcement Learning
by Xuchuan Liu, Yuan Zheng, Chenglong Li, Bo Jiang and Wenyong Gu
Aerospace 2026, 13(2), 111; https://doi.org/10.3390/aerospace13020111 - 23 Jan 2026
Viewed by 26
Abstract
Collision avoidance between unmanned aerial vehicles (UAVs) and non-cooperative targets (e.g., off-nominal operations or birds) presents significant challenges in urban air mobility (UAM). This difficulty arises due to the highly dynamic and unpredictable flight intentions of these targets. Traditional collision-avoidance methods primarily focus [...] Read more.
Collision avoidance between unmanned aerial vehicles (UAVs) and non-cooperative targets (e.g., off-nominal operations or birds) presents significant challenges in urban air mobility (UAM). This difficulty arises due to the highly dynamic and unpredictable flight intentions of these targets. Traditional collision-avoidance methods primarily focus on cooperative targets or non-cooperative ones with fixed behavior, rendering them ineffective when dealing with highly unpredictable flight patterns. To address this, we introduce a deep reinforcement learning-based collision-avoidance approach leveraging global and local intent prediction. Specifically, we propose a Global and Local Perception Prediction Module (GLPPM) that combines a state-space-based global intent association mechanism with a local feature extraction module, enabling accurate prediction of short- and long-term flight intents. Additionally, we propose a Fusion Sector Flight Control Module (FSFCM) that is trained with a Dueling Double Deep Q-Network (D3QN). The module integrates both predicted future and current intents into the state space and employs a specifically designed reward function, thereby ensuring safe UAV operations. Experimental results demonstrate that the proposed method significantly improves mission success rates in high-density environments, with up to 80 non-cooperative targets per square kilometer. In 1000 flight tests, the mission success rate is 15.2 percentage points higher than that of the baseline D3QN. Furthermore, the approach retains an 88.1% success rate even under extreme target densities of 120 targets per square kilometer. Finally, interpretability analysis via Deep SHAP further verifies the decision-making rationality of the algorithm. Full article
(This article belongs to the Section Aeronautics)
30 pages, 6571 KB  
Article
MRKAN: A Multi-Scale Network for Dual-Polarization Radar Multi-Parameter Extrapolation
by Junfei Wang, Yonghong Zhang, Linglong Zhu, Qi Liu, Haiyang Lin, Huaqing Peng and Lei Wu
Remote Sens. 2026, 18(2), 372; https://doi.org/10.3390/rs18020372 - 22 Jan 2026
Viewed by 13
Abstract
Severe convective weather is marked by abrupt onset, rapid evolution, and substantial destructive potential, posing major threats to economic activities and human safety. To address this challenge, this study proposes MRKAN, a multi-parameter prediction algorithm for dual-polarization radar that integrates Mamba, radial basis [...] Read more.
Severe convective weather is marked by abrupt onset, rapid evolution, and substantial destructive potential, posing major threats to economic activities and human safety. To address this challenge, this study proposes MRKAN, a multi-parameter prediction algorithm for dual-polarization radar that integrates Mamba, radial basis functions (RBFs), and the Kolmogorov–Arnold Network (KAN). The method predicts radar reflectivity, differential reflectivity, and the specific differential phase, enabling a refined depiction of the dynamic structure of severe convective systems. MRKAN incorporates four key innovations. First, a Cross-Scan Mamba module is designed to enhance global spatiotemporal dependencies through point-wise modeling across multiple complementary scans. Second, a Multi-Order KAN module is developed that employs multi-order β-spline functions to overcome the linear limitations of convolution kernels and to achieve high-order representations of nonlinear local features. Third, a Gaussian and Inverse Multiquadratic RBF module is constructed to extract mesoscale features using a combination of Gaussian radial basis functions and Inverse Multiquadratic radial basis functions. Finally, a Multi-Scale Feature Fusion module is designed to integrate global, local, and mesoscale information, thereby enhancing multi-scale adaptive modeling capability. Experimental results show that MRKAN significantly outperforms mainstream methods across multiple key metrics and yields a more accurate depiction of the spatiotemporal evolution of severe convective weather. Full article
22 pages, 2759 KB  
Article
DACL-Net: A Dual-Branch Attention-Based CNN-LSTM Network for DOA Estimation
by Wenjie Xu and Shichao Yi
Sensors 2026, 26(2), 743; https://doi.org/10.3390/s26020743 (registering DOI) - 22 Jan 2026
Viewed by 16
Abstract
While deep learning methods are increasingly applied in the field of DOA estimation, existing approaches generally feed the real and imaginary parts of the covariance matrix directly into neural networks without optimizing the input features, which prevents classical attention mechanisms from improving accuracy. [...] Read more.
While deep learning methods are increasingly applied in the field of DOA estimation, existing approaches generally feed the real and imaginary parts of the covariance matrix directly into neural networks without optimizing the input features, which prevents classical attention mechanisms from improving accuracy. This paper proposes a spatio-temporal fusion model named DACL-Net for DOA estimation. The spatial branch applies a two-dimensional Fourier transform (2D-FT) to the covariance matrix, causing angles to appear as peaks in the magnitude spectrum. This operation transforms the original covariance matrix into a dark image with bright spots, enabling the convolutional neural network (CNN) to focus on the bright-spot components via an attention module. Additionally, a spectrum attention mechanism (SAM) is introduced to enhance the extraction of temporal features in the time branch. The model learns simultaneously from two data branches and finally outputs DOA results through a linear layer. Simulation results demonstrate that DACL-Net outperforms existing algorithms in terms of accuracy, achieving an RMSE of 0.04 at an SNR of 0 dB. Full article
(This article belongs to the Section Communications)
26 pages, 4614 KB  
Article
CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution
by Xia Li, Haicheng Sun and Tie-Qiang Li
Sensors 2026, 26(2), 738; https://doi.org/10.3390/s26020738 (registering DOI) - 22 Jan 2026
Viewed by 7
Abstract
Magnetic resonance imaging (MRI) super-resolution (SR) enables high-resolution reconstruction from low-resolution acquisitions, reducing scan time and easing hardware demands. However, most deep learning-based SR models are large and computationally heavy, limiting deployment in clinical workstations, real-time pipelines, and resource-restricted platforms such as low-field [...] Read more.
Magnetic resonance imaging (MRI) super-resolution (SR) enables high-resolution reconstruction from low-resolution acquisitions, reducing scan time and easing hardware demands. However, most deep learning-based SR models are large and computationally heavy, limiting deployment in clinical workstations, real-time pipelines, and resource-restricted platforms such as low-field and portable MRI. We introduce CHARMS, a lightweight convolutional–Transformer hybrid with attention regularization optimized for MRI SR. CHARMS employs a Reverse Residual Attention Fusion backbone for hierarchical local feature extraction, Pixel–Channel and Enhanced Spatial Attention for fine-grained feature calibration, and a Multi-Depthwise Dilated Transformer Attention block for efficient long-range dependency modeling. Novel attention regularization suppresses redundant activations, stabilizes training, and enhances generalization across contrasts and field strengths. Across IXI, Human Connectome Project Young Adult, and paired 3T/7T datasets, CHARMS (~1.9M parameters; ~30 GFLOPs for 256 × 256) surpasses leading lightweight and hybrid baselines (EDSR, PAN, W2AMSN-S, and FMEN) by 0.1–0.6 dB PSNR and up to 1% SSIM at ×2/×4 upscaling, while reducing inference time ~40%. Cross-field fine-tuning yields 7T-like reconstructions from 3T inputs with ~6 dB PSNR and 0.12 SSIM gains over native 3T. With near-real-time performance (~11 ms/slice, ~1.6–1.9 s per 3D volume on RTX 4090), CHARMS offers a compelling fidelity–efficiency balance for clinical workflows, accelerated protocols, and portable MRI. Full article
(This article belongs to the Special Issue Sensing Technologies in Digital Radiology and Image Analysis)
27 pages, 23394 KB  
Article
YOLO-MSRF: A Multimodal Segmentation and Refinement Framework for Tomato Fruit Detection and Segmentation with Count and Size Estimation Under Complex Illumination
by Ao Li, Chunrui Wang, Aichen Wang, Jianpeng Sun, Fengwei Gu and Tianxue Zhang
Agriculture 2026, 16(2), 277; https://doi.org/10.3390/agriculture16020277 - 22 Jan 2026
Viewed by 23
Abstract
Segmentation of tomato fruits under complex lighting conditions remains technically challenging, especially in low illumination or overexposure, where RGB-only methods often suffer from blurred boundaries and missed small or occluded instances, and simple multimodal fusion cannot fully exploit complementary cues. To address these [...] Read more.
Segmentation of tomato fruits under complex lighting conditions remains technically challenging, especially in low illumination or overexposure, where RGB-only methods often suffer from blurred boundaries and missed small or occluded instances, and simple multimodal fusion cannot fully exploit complementary cues. To address these gaps, we propose YOLO-MSRF, a lightweight RGB–NIR multimodal segmentation and refinement framework for robust tomato perception in facility agriculture. Firstly, we propose a dual-branch multimodal backbone, introduce Cross-Modality Difference Complement Fusion (C-MDCF) for difference-based complementary RGB–NIR fusion, and design C2f-DCB to reduce computation while strengthening feature extraction. Furthermore, we develop a cross-scale attention fusion network and introduce the proposed MS-CPAM to jointly model multi-scale channel and position cues, strengthening fine-grained detail representation and spatial context aggregation for small and occluded tomatoes. Finally, we design the Multi-Scale Fusion and Semantic Refinement Network, MSF-SRNet, which combines the Scale-Concatenate Fusion Module (Scale-Concat) fusion with SDI-based cross-layer detail injection to progressively align and refine multi-scale features, improving representation quality and segmentation accuracy. Extensive experiments show that YOLO-MSRF achieves substantial gains under weak and low-light conditions, where RGB-only models are most prone to boundary degradation and missed instances, and it still delivers consistent improvements on the mixed four-light validation set, increasing mAP0.5 by 2.3 points, mAP0.50.95 by 2.4 points, and mIoU by 3.60 points while maintaining real-time inference at 105.07 FPS. The proposed system further supports counting, size estimation, and maturity analysis of harvestable tomatoes, and can be integrated with depth sensing and yield estimation to enable real-time yield prediction in practical greenhouse operations. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

18 pages, 11508 KB  
Article
Design of MEMS Microphone Array Integrated System for Pipeline Leakage Detection
by Kaixuan Wang, Yong Yang, Daoguang Liu, Di Song and Xiaoli Zhao
Micromachines 2026, 17(1), 140; https://doi.org/10.3390/mi17010140 - 22 Jan 2026
Viewed by 22
Abstract
Pressure pipelines are widely used in the energy and transportation fields for conveying natural gas, water, etc. Under complex and harsh conditions with long-term operation, this easily leads to leakage, threatening the safe and stable operation of transportation systems. Although acoustic sensors support [...] Read more.
Pressure pipelines are widely used in the energy and transportation fields for conveying natural gas, water, etc. Under complex and harsh conditions with long-term operation, this easily leads to leakage, threatening the safe and stable operation of transportation systems. Although acoustic sensors support non-destructive leakage detection, their accuracy is restricted by noise interference and minor leakage uncertainties, and existing systems lack a targeted integration design for pipeline scenarios. To address this, the micro-electromechanical system (MEMS) is specifically designed as an MEMS microphone array integrated system (MEMS-MAIS), which is applied for pipeline leakage detection through data fusion at different levels. First, a dedicated MEMS microphone array system is designed to realize high-sensitivity collection of leakage acoustic data. In addition, the integrated feature extraction and feature-level fusion modules are proposed to retain effective information, and a decision-level fusion module is incorporated to improve the reliability of leakage detection results. To verify the designed system, an experiential platform is established with several microphone data. The results indicate that the proposed MEMS-MAIS exhibits excellent anti-interference performance and leakage detection accuracy of 94.67%. It provides a reliable integrated system solution for pipeline leakage detection and verifying high engineering application value. Full article
Show Figures

Figure 1

Back to TopTop