Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

Search Results (779)

Search Parameters:
Keywords = multi-resolution feature fusion

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
33 pages, 3496 KB  
Article
Modified RefineNet with Attention-Based Fusion for Multi-Class Classification of Corn and Pepper Plant Diseases
by Maramreddy Srinivasulu and Sandipan Maiti
AgriEngineering 2026, 8(4), 122; https://doi.org/10.3390/agriengineering8040122 - 30 Mar 2026
Abstract
Early and precise detection of plant diseases is essential for safeguarding crop yield and ensuring sustainable agricultural practices. In this study, we propose the Modified RefineNet with Attention based Fusion (MoRefNet-AF), a Modified RefineNet architecture enhanced with attention-based fusion for multi-class classification of [...] Read more.
Early and precise detection of plant diseases is essential for safeguarding crop yield and ensuring sustainable agricultural practices. In this study, we propose the Modified RefineNet with Attention based Fusion (MoRefNet-AF), a Modified RefineNet architecture enhanced with attention-based fusion for multi-class classification of corn (maize) and Pepper leaf diseases. Unlike the original RefineNet, which was segmentation-oriented and computationally heavy, MoRefNet-AF is redesigned for lightweight and discriminative classification. The modifications include replacing standard convolutions with depthwise separable convolutions for efficiency, adopting the Mish activation function for smoother gradient flow, redesigning the multi-resolution fusion module with concatenation and shared convolution for richer cross-scale integration, and incorporating Squeeze-and-Excitation (SE) blocks for adaptive channel recalibration. Additionally, Chained Residual Pooling (CRP) with atrous convolutions enhances contextual representation, while global average pooling with dense layers improves classification readiness. When evaluated on a curated six-class dataset combining PlantVillage and Mendeley leaf disease repositories, MoRefNet-AF achieved 99.88% accuracy, 99.74% precision, 99.73% recall, 99.95% F1-score, and 99.73% specificity. These results outperform strong baselines including ResNet152V2, DenseNet201, EfficientNet-B0, and ConvNeXt-Tiny, while maintaining only 0.3 M parameters. With its compact design and TensorFlow Lite (v2.13) compatibility, MoRefNet-AF offers a robust, lightweight, and real-time deployable solution for precision agriculture and smart plant disease monitoring. Full article
27 pages, 7912 KB  
Article
Hierarchical Wetland Mapping in the East China Sea Based on Integrated Multifaceted Source Features
by Jie Wang, Yixuan Zhou, Xin Fang, Shengqi Wang, Haiyang Zhang and Runbin Hu
Remote Sens. 2026, 18(7), 1023; https://doi.org/10.3390/rs18071023 - 29 Mar 2026
Abstract
The East China Sea represents a critical coastal wetland region, characterized by complex geomorphology, heterogeneous land-cover composition, and diverse wetland types. Accurate delineation of coastal wetland extent is essential for ecosystem service assessment and sustainable coastal management, directly contributing to wetland-related Sustainable Development [...] Read more.
The East China Sea represents a critical coastal wetland region, characterized by complex geomorphology, heterogeneous land-cover composition, and diverse wetland types. Accurate delineation of coastal wetland extent is essential for ecosystem service assessment and sustainable coastal management, directly contributing to wetland-related Sustainable Development Goals (SDGs), particularly SDG 15, on ecosystem conservation and biodiversity protection. However, pronounced spectral similarity and structural heterogeneity among wetland classes pose substantial challenges to reliable classification. To address these challenges, this study developed a hierarchical classification framework integrating Random Forest, K-means clustering, and a decision tree classifier based on multi-source Sentinel-1 and Sentinel-2 imagery. Spectral, polarimetric, texture, and morphological features were systematically constructed to enhance class separability. Using this framework, a 10 m resolution coastal wetland map of the East China Sea was generated for 2023. The proposed approach achieved an overall accuracy of 91.32% and improved the discrimination of spectrally similar wetland types. Feature fusion reduced confusion among water-related classes, while object-based clustering improved the extraction of linear riverine wetlands. The resulting 10 m wetland map provides updated spatial information for ecological assessment and coastal management in the East China Sea. Full article
(This article belongs to the Special Issue Big Earth Data in Support of the Sustainable Development Goals)
Show Figures

Figure 1

35 pages, 6116 KB  
Article
Attention-Enhanced GAN for Spatial–Spectral Fusion and Chlorophyll-a Inversion in Chen Lake, China
by Chenxi Zeng, Cheng Shang, Yankun Wang, Shan Jiang, Ningsheng Chen, Chengyu Geng, Yadong Zhou and Yun Du
Sensors 2026, 26(7), 2107; https://doi.org/10.3390/s26072107 - 28 Mar 2026
Viewed by 65
Abstract
The Sentinel-3 Ocean and Land Colour Instrument (OLCI) is designed for water monitoring. Its 21-spectral bands serve as the basis for the precise retrieval of water quality parameters. However, its coarse resolution restricts the depiction of the spatial distribution of water quality parameters [...] Read more.
The Sentinel-3 Ocean and Land Colour Instrument (OLCI) is designed for water monitoring. Its 21-spectral bands serve as the basis for the precise retrieval of water quality parameters. However, its coarse resolution restricts the depiction of the spatial distribution of water quality parameters in small inland water bodies. Spatial–spectral fusion is a common method to address the inherent constraints between the spatial and spectral resolutions of sensors. Central to the popular methods is the deep learning-based method. Nonetheless, deep-learning-based models still face challenges in fusing Sentinel-2 Multi-Spectral Instrument (MSI) and Sentinel-3 OLCI data. Here, we propose a Multi-Scale-Attention-based Unsupervised Generative Adversarial Network (MSA-UGAN), which effectively integrates OLCI’s spectral advantage and MSI’s spatial resolution. Quantitative evaluation was conducted against five benchmark methods, including traditional approaches (GS, SFIM, MTF-GLP) and deep learning models (SRCNN, UCGAN). The results show that MSA-UGAN achieves the best overall performance: QNR (0.9709) and SSIM (0.9087) are the highest, while SAM (1.1331), spatial distortion (DS = 0.0389), and spectral distortion (Dλ = 0.0252) are the lowest. This shows that MSA-UGAN can better preserve the spatial details of S2 MSI and the spectral features of S3 OLCI data. Moreover, ERGAS (2.2734) also performs excellently in the comparative experiments. The experiment of Chlorophyll-a inversion using the fused image in Chen Lake revealed a spatial gradient ranging from 3.25 to 19.33 µg/L, with the highest concentrations in the southwestern nearshore waters, likely associated with aquaculture. These results jointly indicate that MSA-UGAN can generate high-spatial-resolution multispectral images, and the fused images can be effectively utilized for water quality monitoring, thereby providing essential data support for the precision management and scientific decision-making regarding inland lakes. Full article
(This article belongs to the Section Remote Sensors)
16 pages, 10364 KB  
Article
A Method for Filling Blank Stripes in Electrical Imaging Based on the Fusion of Arbitrary Kernel Convolution and Generative Adversarial Networks
by Ruhan A, Die Liu, Ge Cao, Kun Meng, Taiping Zhao, Lili Tian, Bin Zhao, Guilan Lin and Sinan Fang
Appl. Sci. 2026, 16(7), 3267; https://doi.org/10.3390/app16073267 - 27 Mar 2026
Viewed by 178
Abstract
Electrical imaging logging images play a crucial role in petroleum exploration; however, in practical applications, blank strips frequently appear due to instrument malfunctions or data transmission failures, severely compromising geological interpretation and hydrocarbon evaluation. Existing image inpainting methods have limited adaptability to blank [...] Read more.
Electrical imaging logging images play a crucial role in petroleum exploration; however, in practical applications, blank strips frequently appear due to instrument malfunctions or data transmission failures, severely compromising geological interpretation and hydrocarbon evaluation. Existing image inpainting methods have limited adaptability to blank strips at different depth scales and exhibit blurred high-resolution geological textures. To address these issues, this paper proposes a blank strip filling method that integrates Arbitrary Kernel Convolution (AKConv) with the Aggregated Contextual-Transformations Generative Adversarial Network (AOT-GAN). Specifically, the adaptive sampling mechanism of AKConv is incorporated into the generator network of AOT-GAN, enabling the model—to effectively capture long-range contextual information and adaptively handle blank strips of varying scales and shapes through multi-scale feature fusion. Experimental results on real oilfield datasets demonstrate that the proposed method achieves significant improvements in PSNR, SSIM, and MAE, exhibiting superior structural preservation and texture sharpness—especially in restoring deep and large-scale blank strips. Furthermore, visual comparisons confirm the method’s superior performance in recovering key geological features, such as bedding continuity and fracture structures, thus providing an effective approach for electrical imaging logging image restoration. Full article
(This article belongs to the Special Issue Applied Geophysical Imaging and Data Processing, 2nd Edition)
Show Figures

Figure 1

21 pages, 922 KB  
Article
DBCF-Net: A Dual-Branch Cross-Scale Fusion Network for Heterogeneous Satellite–UAV Change Detection
by Yan Ren, Ruiyong Li, Pengbo Zhai and Xinyu Chen
Remote Sens. 2026, 18(7), 1009; https://doi.org/10.3390/rs18071009 - 27 Mar 2026
Viewed by 101
Abstract
Heterogeneous change detection (HCD) using satellite and Unmanned Aerial Vehicle (UAV) imagery is a pivotal task in remote sensing and Earth observation. However, the effective utilization of such multi-source data is significantly hindered by extreme spatial resolution disparities and distinct radiometric characteristics. Existing [...] Read more.
Heterogeneous change detection (HCD) using satellite and Unmanned Aerial Vehicle (UAV) imagery is a pivotal task in remote sensing and Earth observation. However, the effective utilization of such multi-source data is significantly hindered by extreme spatial resolution disparities and distinct radiometric characteristics. Existing deep learning methods, often based on weight-sharing Siamese architectures, struggle to bridge these domain gaps, leading to spectral pseudo-changes and blurred detection boundaries. To address these challenges, we propose a novel Dual-Branch Cross-Scale Fusion Network (DBCF-Net) specifically tailored for heterogeneous satellite–UAV change detection. We introduce a Difference-Aware Attention Module (DAAM) to explicitly align cross-modal feature spaces and suppress domain-related noise through a hybrid local–global attention mechanism. Furthermore, an Adaptive Gated Fusion Module (AGFM) is designed to dynamically weight multi-scale interactions, ensuring the preservation of high-frequency spatial details from UAV imagery while maintaining the semantic consistency of satellite data. Extensive experiments on the Heterogeneous Satellite–UAV Dataset (HSUD) demonstrate that DBCF-Net achieves state-of-the-art performance, reaching an F1-score of 88.75% and an IoU of 80.58%. This study provides a robust technical framework for heterogeneous sensor fusion and high-precision monitoring in complex remote sensing scenarios. Full article
(This article belongs to the Section Remote Sensing Image Processing)
21 pages, 11455 KB  
Article
Cross-Scale Spectral Calibration for Spatiotemporal Fusion of Remote Sensing Images
by Yishuo Tian, Xiaorong Xue, Jingtong Yang, Wen Zhang, Bingyan Lu, Xin Zhao and Wancheng Wang
Sensors 2026, 26(7), 2090; https://doi.org/10.3390/s26072090 - 27 Mar 2026
Viewed by 271
Abstract
Spatiotemporal fusion aims to generate remote sensing images with both high spatial and high temporal resolution by integrating multi-source observations. However, significant spectral inconsistencies often arise when fusing images acquired at different spatial scales, which severely degrade the radiometric fidelity and temporal reliability [...] Read more.
Spatiotemporal fusion aims to generate remote sensing images with both high spatial and high temporal resolution by integrating multi-source observations. However, significant spectral inconsistencies often arise when fusing images acquired at different spatial scales, which severely degrade the radiometric fidelity and temporal reliability of the fused results. Most existing methods focus on enhancing spatial details or temporal consistency, while the cross-scale spectral discrepancy between coarse- and fine-resolution images has not been sufficiently addressed. To tackle this issue, we propose a cross-scale spectral calibration framework for spatiotemporal fusion (XSC-Net), which explicitly models and corrects spectral responses across different spatial scales. The proposed method introduces a spatial feature refinement block to enhance spatially discriminative structures and a hierarchical spectral refinement block to adaptively calibrate channel-wise spectral representations. By jointly exploiting spatial and spectral correlations, the proposed framework effectively suppresses spectral distortion while preserving fine spatial details. Extensive experiments on the public CIA and LGC datasets indicate that XSC-Net compares favorably with state-of-the-art methods, demonstrating superior performance over established baselines. Furthermore, ablation studies verify the efficacy and contribution of the proposed architectural components. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

25 pages, 8205 KB  
Article
Forest Road Extraction via Optimized DeepLabv3+ and Multi-Temporal Remote Sensing for Wildfire Emergency Response
by Zhuoran Gao, Ziyang Li, Weiyuan Yao, Tingtao Zhang, Shi Qiu and Zhaoyan Liu
Appl. Sci. 2026, 16(7), 3228; https://doi.org/10.3390/app16073228 - 26 Mar 2026
Viewed by 251
Abstract
Forest fires occur frequently in China; however, the complex terrain and incomplete road networks severely constrain ground rescue efficiency. Accurate forest road information is essential for the optimization of emergency response and rescue force deployment. Existing road extraction algorithms are primarily designed for [...] Read more.
Forest fires occur frequently in China; however, the complex terrain and incomplete road networks severely constrain ground rescue efficiency. Accurate forest road information is essential for the optimization of emergency response and rescue force deployment. Existing road extraction algorithms are primarily designed for urban environments and exhibit limited efficacy in forest scenarios due to dense canopy, complex background interference and specific forest road features. To address this gap, this study proposes a forest road extraction method based on an enhanced DeepLabv3+ model using multi-temporal, high-resolution satellite imagery. Specifically, a Multi-Scale Channel Attention (MCSA) mechanism is embedded in skip connections to suppress background interference, while strip pooling is integrated into the Atrous Spatial Pyramid Pooling (ASPP) module to better capture slender road features. A composite Focal-Dice loss function is also constructed to mitigate sample imbalance. Finally, by applying the model in multi-temporal remote sensing images, a fusion strategy is introduced to integrate multi-seasonal road masks to enhance overall accuracy and topological integrity. Experimental results show that the proposed method achieves a precision of 54.1%, an F1-Score of 59.3%, and an IoU of 41.8%, effectively enhancing road continuity and providing robust technical support for fire-rescue decision-making. Full article
Show Figures

Figure 1

28 pages, 105542 KB  
Article
Underwater Image Enhancement via HSV-CS Representation and Perception-Driven Adaptive Fusion
by Fengxu Guan, Tong Guo and Yuzhu Zhang
Remote Sens. 2026, 18(7), 986; https://doi.org/10.3390/rs18070986 - 25 Mar 2026
Viewed by 230
Abstract
Underwater images often suffer from color distortion and low contrast, severely limiting the reliability of visual perception systems. Existing methods struggle to balance enhancement quality and computational efficiency. To address this issue, we propose PCF-Net (Perception-driven Color Fusion Network), a lightweight dual-branch network [...] Read more.
Underwater images often suffer from color distortion and low contrast, severely limiting the reliability of visual perception systems. Existing methods struggle to balance enhancement quality and computational efficiency. To address this issue, we propose PCF-Net (Perception-driven Color Fusion Network), a lightweight dual-branch network for underwater image enhancement based on a stable HSV-CS (Hue-Saturation-Value with sine–cosine transformation) color-space representation. Specifically, a sine–cosine transformation is introduced to construct a stable HSV-CS color space, effectively avoiding hue discontinuities at boundary regions in conventional HSV representations. To compensate for underwater degradation, a Color-Bias-Aware module and a Value-Confidence module are designed to adaptively correct color distortion and luminance degradation. Furthermore, a lightweight Channel-Spatial Adaptive Gated Fusion module dynamically aggregates features from the RGB and HSV-CS branches in a perception-driven manner. The overall architecture incorporates multi-branch re-parameterizable convolutions, significantly reducing computational cost while preserving strong representational capacity. Extensive experiments on underwater image enhancement benchmarks, including UIEB and RUIE, demonstrate that PCF-Net achieves state-of-the-art performance in terms of PSNR, SSIM, and UIQM, along with visually superior color correction and contrast enhancement. With only 0.17 M parameters, the proposed model runs at 118.6 FPS on an RTX 3090 and 35.3 FPS on a Jetson Orin Nano at a resolution of 512 × 512, making it well suited for resource-constrained real-time underwater vision applications. Full article
(This article belongs to the Special Issue Deep Learning for Remote Sensing Image Enhancement)
Show Figures

Figure 1

12 pages, 1041 KB  
Communication
Artificial Oxidation: A Major Challenge in Implementing Multi-Attribute Methods for Therapeutic Protein Analysis
by Yaokai Duan, Michael Lanzillotti, Dylan L. Riggs, Albana Nito, Junnichi Mijares, Amanda Helms, Carl Ly, Kevin Millea, Xingwen Li, Hao Zhang and Zhongqi Zhang
Pharmaceuticals 2026, 19(4), 528; https://doi.org/10.3390/ph19040528 - 25 Mar 2026
Viewed by 189
Abstract
Background/Objectives: Mass spectrometry-based multi-attribute methods (MAM) have the potential to transform therapeutic protein analysis by enabling comprehensive monitoring of multiple quality attributes in a single assay. However, the widespread adoption of MAM is hindered by significant challenges, most notably artificial oxidation during [...] Read more.
Background/Objectives: Mass spectrometry-based multi-attribute methods (MAM) have the potential to transform therapeutic protein analysis by enabling comprehensive monitoring of multiple quality attributes in a single assay. However, the widespread adoption of MAM is hindered by significant challenges, most notably artificial oxidation during sample preparation and analysis. This report summarizes long-term operational observations and several case studies that substantiate this concern. Methods: A tryptic digest, high-resolution LC-MS MAM workflow was applied to an Fc-fusion protein and multiple antibody-based therapeutics, with a frozen reference standard analyzed in each run for system suitability and longitudinal trending. Oxidation excursions were investigated by comparing laboratories, consumables, LC-MS configurations, and other method parameters. Results: In a seven-year trending record, apparent total methionine oxidation in the Fc-fusion protein reference standard showed an abrupt, sustained increase (up to ~5-fold); the shift was traced to a specific bag of microcentrifuge-tubes used during buffer exchange and resolved after those tubes were discontinued. In an antibody–drug conjugate, observed methionine oxidation was strongly influenced by the sample preparation procedure. In other antibodies, variability of observed methionine oxidation was attributed to on-column oxidation, which produced a broad and noisy peak that interferes with automated peak integration. EDTA flushing reduced this feature, implicating exposure to metal ions. Conclusions: While advances continue to address many MAM challenges, artificial oxidation remains unpredictable and constitutes a major obstacle to robust implementation in regulated QC environments. Enhanced control strategies and further research are urgently needed to ensure reliable therapeutic protein analysis. Such control strategies include consumable qualification and change control, system suitability/trending using a reference standard, metal management across LC flow path/column lifecycle, reduction of trifluoracetic acid (TFA) exposure, data analysis to safeguard excessive on-column oxidation, etc. Full article
Show Figures

Graphical abstract

28 pages, 14283 KB  
Article
FSD-YOLO: A Fusion Framework for Region Segmentation and Deformable Object Detection in Container Yards
by Linghao Dai, Zhihong Liang, Qi Feng, Shihuan Xie and Hongxu Li
Sensors 2026, 26(7), 2029; https://doi.org/10.3390/s26072029 - 24 Mar 2026
Viewed by 188
Abstract
Safety monitoring in container hoisting operations within rail-road intermodal logistics parks is a critical task in industrial safety management. Such scenarios are characterized by complex environments, large variations in target scales, deformable object shapes, and frequent occlusions, which pose significant challenges to visual [...] Read more.
Safety monitoring in container hoisting operations within rail-road intermodal logistics parks is a critical task in industrial safety management. Such scenarios are characterized by complex environments, large variations in target scales, deformable object shapes, and frequent occlusions, which pose significant challenges to visual perception systems. Conventional single-task models suffer from inherent limitations in handling low recall rates for distant small targets and insufficient adaptability to geometric deformations, making them inadequate for high-precision, real-time safety warning applications. To address these challenges, this study proposes a unified visual analysis framework that integrates semantic segmentation and object detection to enhance the recognition performance of small and deformable targets in complex operational environments, enabling real-time perception and safety warning of key objects and hazardous regions within container yards. Specifically, we introduce FSD-YOLO, a fusion-based architecture composed of the following key components. First, a SegFormer-based semantic segmentation module is employed to achieve pixel-level delineation of different operational regions. Second, an improved object detection network is developed based on the YOLOv8n architecture, incorporating: (1) the integration of C2f modules in the shallow layers of the backbone to enhance high-resolution feature extraction; (2) the embedding of C2fDCN modules within the detection head to improve modeling capability for deformable objects via deformable convolution; (3) the adoption of CARAFE upsampling operators to optimize multi-scale feature fusion; and (4) a dynamic loss-weighting strategy for small objects, where loss weights are adaptively adjusted according to target area to increase training emphasis on small-scale targets. Finally, a decision-level fusion strategy is applied to combine segmentation and detection outputs, enabling real-time safety judgment based on semantic rules. Experimental results on a self-constructed container yard dataset demonstrate that the proposed detection model achieves an mAP50-95 of 0.6433 and an mAP50 of 0.9565, significantly outperforming the baseline YOLOv8n model (mAP50-95: 0.5394, mAP50: 0.8435), thereby validating the effectiveness of the proposed framework. Full article
Show Figures

Figure 1

27 pages, 8177 KB  
Article
DINOv3-PEFT: A Dual-Branch Collaborative Network with Parameter-Efficient Fine-Tuning for Precise Road Segmentation in SAR Imagery
by Debao Chen, Wanlin Yang, Ye Yuan and Juntao Gu
Remote Sens. 2026, 18(7), 973; https://doi.org/10.3390/rs18070973 - 24 Mar 2026
Viewed by 109
Abstract
Extracting road networks from Synthetic Aperture Radar (SAR) data represents a core challenge in remote sensing scene analysis, particularly for applications in traffic monitoring and emergency management. The task is complicated by several inherent limitations: speckle noise degrades image quality, geometric distortions arise [...] Read more.
Extracting road networks from Synthetic Aperture Radar (SAR) data represents a core challenge in remote sensing scene analysis, particularly for applications in traffic monitoring and emergency management. The task is complicated by several inherent limitations: speckle noise degrades image quality, geometric distortions arise from the side-looking acquisition geometry, and roads often exhibit weak radiometric separation from surrounding terrain. Traditional processing pipelines and recent single-branch deep learning frameworks have shown insufficient performance when global contextual reasoning and fine-scale spatial detail must both be addressed. This work presents DINOv3-PEFT, a parameter-efficient dual-encoder network designed specifically for SAR road segmentation. The architecture employs two complementary processing streams tailored to SAR characteristics: one stream utilizes adapter-based fine-tuning applied to pre-trained DINOv3 weights (kept frozen), which captures long-distance spatial relationships crucial for maintaining network connectivity despite speckle corruption. The second stream, based on convolutional operations, focuses on extracting localized geometric features that preserve the narrow, elongated structure and sharp boundaries typical of road infrastructure. Feature fusion occurs through the Topological-Geometric Feature Integration (TGFI) Module, which synthesizes multi-scale representations hierarchically. This mechanism proves effective at bridging fragmented road segments and recovering geometric accuracy in scenarios with heavy shadow casting or signal interference. Performance evaluation on the GF-3 satellite dataset across four spatial resolutions (1 m, 3 m, 5 m, and 10 m) demonstrates the proposed method achieves an 82.61% F1-score, a 76.51% IoU, and a 98.08% overall accuracy, all averaged across the four resolutions. When benchmarked against six state-of-the-art methods, DINOv3-PEFT demonstrates substantial improvements in road class segmentation quality and topological connectivity preservation, supporting its robustness for operational SAR road mapping tasks. Full article
Show Figures

Figure 1

20 pages, 3073 KB  
Article
YOLOv11-WFD: A Multimodal Grape Segmentation Framework with Wavelet Convolution, FasterNeXt, and Dynamic Upsampling for Intelligent Harvesting
by Pengyan Wang, Chengshuai Li and Linjing Wei
Agronomy 2026, 16(7), 679; https://doi.org/10.3390/agronomy16070679 (registering DOI) - 24 Mar 2026
Viewed by 130
Abstract
Grapes are high-value crops, but expanding cultivation has made manual harvesting inefficient and costly due to labor shortages and weather constraints. Automated harvesting requires accurate and lightweight image segmentation to ensure reliable visual perception. Improving segmentation precision, robustness, and model compactness is thus [...] Read more.
Grapes are high-value crops, but expanding cultivation has made manual harvesting inefficient and costly due to labor shortages and weather constraints. Automated harvesting requires accurate and lightweight image segmentation to ensure reliable visual perception. Improving segmentation precision, robustness, and model compactness is thus critical for intelligent grape harvesting. To enhance segmentation robustness in complex orchard environments, this study introduces a multimodal fusion and multi-scale enhancement strategy and develops a lightweight instance segmentation network. Using a multimodal grape dataset containing RGB, near-infrared (NIR), and depth information, a multi-resolution training scheme based on an image-pyramid framework was constructed. Among the three YOLOv11-based fusion strategies, early fusion achieved the best performance. Accordingly, the lightweight model YOLOv11-WFD was designed by integrating FasterNeXt, DySample, and WaveletPool to strengthen feature extraction, adaptive sampling, and small-object perception. The model delivers high segmentation accuracy and strong deployment suitability for intelligent harvesting applications. Experimental results show that YOLOv11-WFD achieves a mAP@50:95 of 79.3% on the validation set with only 2.25 M parameters, demonstrating outstanding performance in both precision and compactness. Compared with YOLOv3-tiny, YOLOv5n, YOLOv8n, YOLOv10n, YOLOv11n, and YOLOv12n, YOLOv11-WFD improves mAP@50:95 by 25.4, 3.0, 2.7, 2.8, 2.0, and 3.1 percentage points, respectively, while reducing parameters by 80.4%, 7.8%, 23.5%, 10.7%, 20.8%, and 18.8%. Overall, YOLOv11-WFD achieves an excellent balance among accuracy, speed, and complexity, verifying the effectiveness of the multimodal fusion and lightweight integration strategy. It shows strong potential for practical applications and large-scale deployment in complex agricultural environments such as intelligent grape harvesting. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

28 pages, 14845 KB  
Article
Spatial Relation Reasoning Based on Keypoints for Railway Intrusion Detection and Risk Assessment
by Shanping Ning, Feng Ding and Bangbang Chen
Appl. Sci. 2026, 16(6), 3026; https://doi.org/10.3390/app16063026 - 20 Mar 2026
Viewed by 157
Abstract
Foreign object intrusion in railway tracks is a major threat to train operation safety, yet current detection methods face challenges in identifying small distant targets and adapting to low-light conditions. Moreover, existing systems often lack the ability to assess intrusion risk levels, limiting [...] Read more.
Foreign object intrusion in railway tracks is a major threat to train operation safety, yet current detection methods face challenges in identifying small distant targets and adapting to low-light conditions. Moreover, existing systems often lack the ability to assess intrusion risk levels, limiting real-time warning and graded response capabilities. To address these gaps, this paper proposes a novel method for intrusion detection and risk assessment based on keypoint spatial discrimination. First, an XS-BiSeNetV2-based track segmentation network is developed, incorporating cross-feature fusion and spatial feature recalibration to improve track extraction accuracy in complex scenes. Second, an enhanced STI-YOLO detection model is introduced, integrating a Shuffle attention mechanism for better feature interaction, a high-resolution Transformer detection head to improve small-target sensitivity, and the Inner-IoU loss function to refine bounding box regression. Detected targets’ bottom keypoints are then analyzed relative to track boundaries to determine intrusion direction. By combining lateral distance and motion state features, a multi-level risk classification system is established for quantitative threat assessment. Experiments on the RailSem19 and GN-rail-Object datasets show that the method achieves a track segmentation mIoU of 88.19% and a detection mAP of 82.6%. The risk assessment module effectively quantifies threats across scenarios and maintains stable performance under low-light and strong-glare conditions. This work offers a quantifiable risk assessment solution for intelligent railway safety systems. Full article
Show Figures

Figure 1

26 pages, 20660 KB  
Article
Sea Ice and Water Segmentation in SAR Imagery Based on Polarization Channel Interaction and Edge Selective Fusion
by Wei Song, Yixun Chen, Bin Liu, Mengying Ge, Yiji Zhou and Huifang Xu
Remote Sens. 2026, 18(6), 945; https://doi.org/10.3390/rs18060945 - 20 Mar 2026
Viewed by 190
Abstract
Sea ice segmentation based on Synthetic Aperture Radar (SAR) images has become an important technical means for polar climate change monitoring and navigation safety guarantee. However, the existing methods have limitations in the utilization of SAR polarization information and the modeling of local [...] Read more.
Sea ice segmentation based on Synthetic Aperture Radar (SAR) images has become an important technical means for polar climate change monitoring and navigation safety guarantee. However, the existing methods have limitations in the utilization of SAR polarization information and the modeling of local diversity details of sea ice, which leads to insufficient segmentation, especially in complex ice-water boundary regions. To address these issues, this paper proposes a novel Polarization-Fused Edge-Enhanced UNet (PFEE-UNet) designed specifically for sea ice segmentation from high-resolution SAR images. Specifically, we design the Cross-Polarization Channel Interaction (CPCI) module, which employs a dual interaction strategy of hierarchical inter-group cascading and symmetric cross-fusion. This approach effectively leverages the complementary features of the HH and HV polarization channels, significantly enhancing the distinction between sea ice and open water. Additionally, we present the Dense–Sparse Diversity Enhancement (DSDE) module, which combines a spatial-channel joint attention mechanism to strengthen the model’s ability to capture spatial relationships within complex ice–water structures, effectively alleviating misclassifications caused by abrupt local texture changes. Finally, we design the Selective Edge Fusion (SEF) module, which dynamically selects and integrates multi-level edge features, improving the continuity of sea ice boundaries and preserving its morphological integrity. The experimental results show that the proposed PFEE-UNet model outperforms mainstream segmentation methods on the AI4Arctic/ASIP sea ice dataset, achieving an average Intersection over Union (IoU) of 84.48%, which surpasses existing methods such as HRNet (82.52%) and DeepLabv3+ (82.40%). Additionally, PFEE-UNet was applied for end-to-end ice–water segmentation on real-world Sentinel-1 SAR scenes, demonstrating its effectiveness and robustness for practical sea ice monitoring. Full article
(This article belongs to the Special Issue Innovative Remote-Sensing Technologies for Sea Ice Observing)
Show Figures

Figure 1

29 pages, 6237 KB  
Article
Development of a Multi-Scale Spectrum Phenotyping Framework for High-Throughput Screening of Salt-Tolerant Rice Varieties
by Xiaorui Li, Jiahao Han, Dongdong Han, Shibo Fang, Zhanhao Zhang, Li Yang, Chunyan Zhou, Chengming Jin and Xuejian Zhang
Agronomy 2026, 16(6), 658; https://doi.org/10.3390/agronomy16060658 - 20 Mar 2026
Viewed by 229
Abstract
Soil salinization severely threatens agricultural sustainability in saline–alkali regions, and high-throughput, efficient screening of salt-tolerant rice varieties is critical to mitigating this threat. Traditional evaluation methods are constrained by low throughput, limited spatiotemporal resolution, and the lack of standardized indicators. To address these [...] Read more.
Soil salinization severely threatens agricultural sustainability in saline–alkali regions, and high-throughput, efficient screening of salt-tolerant rice varieties is critical to mitigating this threat. Traditional evaluation methods are constrained by low throughput, limited spatiotemporal resolution, and the lack of standardized indicators. To address these gaps, this study established a multi-scale spectral phenotyping framework integrating ground-based hyperspectral, UAV-borne multispectral, and Sentinel-2 satellite remote sensing data for high-throughput screening of salt-tolerant rice. Field experiments were conducted with 12 rice lines at five key growth stages in Ningxia, China, with synchronous ground spectral measurements and UAV image acquisition on the same day for each stage. Five feature selection methods were employed to screen salt stress-sensitive hyperspectral bands, with classification accuracy validated via a Support Vector Machine (SVM) model. The results showed that: (1) rice spectral characteristics varied dynamically across growth stages, and first-order differential transformation effectively amplified subtle spectral variations in stress-sensitive regions; (2) the Minimum Redundancy–Maximum Relevance (mRMR) method outperformed other methods, achieving 100% classification accuracy at key growth stages, with sensitive bands dominated by red edge bands (58.33%); (3) the constructed Salt Stress Index (SIR) showed strong correlations with classical vegetation indices and rice yield, and could clearly distinguish salt-tolerant and salt-sensitive rice varieties, with stable performance against field environmental noise; and (4) band matching between UAV and Sentinel-2 data enabled multi-scale data fusion and regional-scale salt stress monitoring. This framework realizes the transformation from qualitative spectral description to quantitative salt tolerance evaluation, providing standardized technical support for salt-tolerant rice breeding and precision management of saline–alkali lands. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

Back to TopTop