Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,570)

Search Parameters:
Keywords = multi sensor fusion

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 11870 KB  
Article
Early Mapping of Farmland and Crop Planting Structures Using Multi-Temporal UAV Remote Sensing
by Lu Wang, Yuan Qi, Juan Zhang, Rui Yang, Hongwei Wang, Jinlong Zhang and Chao Ma
Agriculture 2025, 15(21), 2186; https://doi.org/10.3390/agriculture15212186 (registering DOI) - 22 Oct 2025
Abstract
Fine-grained identification of crop planting structures provides key data for precision agriculture, thereby supporting scientific production and evidence-based policy making. This study selected a representative experimental farmland in Qingyang, Gansu Province, and acquired Unmanned Aerial Vehicle (UAV) multi-temporal data (six epochs) from multiple [...] Read more.
Fine-grained identification of crop planting structures provides key data for precision agriculture, thereby supporting scientific production and evidence-based policy making. This study selected a representative experimental farmland in Qingyang, Gansu Province, and acquired Unmanned Aerial Vehicle (UAV) multi-temporal data (six epochs) from multiple sensors (multispectral [visible–NIR], thermal infrared, and LiDAR). By fusing 59 feature indices, we achieved high-accuracy extraction of cropland and planting structures and identified the key feature combinations that discriminate among crops. The results show that (1) multi-source UAV data from April + June can effectively delineate cropland and enable accurate plot segmentation; (2) July is the optimal time window for fine-scale extraction of all planting-structure types in the area (legumes, millet, maize, buckwheat, wheat, sorghum, maize–legume intercropping, and vegetables), with a cumulative importance of 72.26% for the top ten features, while the April + June combination retains most of the separability (67.36%), enabling earlier but slightly less precise mapping; and (3) under July imagery, the SAM (Segment Anything Model) segmentation + RF (Random Forest) classification approach—using the RF-selected top 10 of the 59 features—achieved an overall accuracy of 92.66% with a Kappa of 0.9163, representing a 7.57% improvement over the contemporaneous SAM + CNN (Convolutional Neural Network) method. This work establishes a basis for UAV-based recognition of typical crops in the Qingyang sector of the Loess Plateau and, by deriving optimal recognition timelines and feature combinations from multi-epoch data, offers useful guidance for satellite-based mapping of planting structures across the Loess Plateau following multi-scale data fusion. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

19 pages, 2186 KB  
Article
A Range Query Method with Data Fusion in Two-Layer Wire-Less Sensor Networks
by Shouxue Chen, Yun Deng and Xiaohui Cheng
Symmetry 2025, 17(11), 1784; https://doi.org/10.3390/sym17111784 (registering DOI) - 22 Oct 2025
Abstract
Wireless sensor networks play a crucial role in IoT applications. Traditional range query methods have limitations in multi-sensor data fusion and energy efficiency. This paper proposes a new range query method for two-layer wireless sensor networks. The method supports data fusion operations directly [...] Read more.
Wireless sensor networks play a crucial role in IoT applications. Traditional range query methods have limitations in multi-sensor data fusion and energy efficiency. This paper proposes a new range query method for two-layer wireless sensor networks. The method supports data fusion operations directly on storage nodes. Communication costs between sink nodes and storage nodes are significantly reduced. Reverse Z-O coding optimizes the encoding process by focusing only on the most valuable data. This approach shortens both encoding time and length. Data security is ensured using the Paillier homomorphic encryption algorithm. A comparison chain for the most valuable data is generated using Reverse Z-O coding and HMAC. Storage nodes can perform multi-sensor data fusion under encryption. Experiments were conducted on Raspberry Pi 2B+ and NVIDIA TX2 platforms. Performance was evaluated in terms of fusion efficiency, query dimensions, and data volume. The results demonstrate secure and efficient multi-sensor data fusion with lower energy consumption. The method outperforms existing approaches in reducing communication and computational costs. Full article
(This article belongs to the Special Issue Symmetry and Asymmetry in Embedded Systems)
Show Figures

Figure 1

25 pages, 2140 KB  
Article
A Bearing Fault Diagnosis Method for Multi-Sensors Using Cloud Model and Dempster–Shafer Evidence Fusion
by Lin Li, Xiafei Zhang, Peng Wang, Chaobo Chen, Tianli Ma and Song Gao
Appl. Sci. 2025, 15(21), 11302; https://doi.org/10.3390/app152111302 - 22 Oct 2025
Abstract
This paper proposes a bearing fault diagnosis method based on the Dempster–Shafer evidence fusion of cloud model memberships from multi-channel data, which provides an explicable calculation process and a final result. Firstly, vibration signals from the drive end and fan end of the [...] Read more.
This paper proposes a bearing fault diagnosis method based on the Dempster–Shafer evidence fusion of cloud model memberships from multi-channel data, which provides an explicable calculation process and a final result. Firstly, vibration signals from the drive end and fan end of the rolling bearing are used as dual-channel data sources to extract multi-dimensional features from time and frequency domains. Then, cloud models are employed to build models for each feature under different conditions, utilizing three digital characteristic parameters to characterize the distribution and uncertainty of features under different operating conditions. Thus, the membership degree vectors of test samples from two channels can be calculated using reference models. Subsequently, D-S evidence theory is applied to fuse membership degree vectors of the two channels, effectively enhancing the robustness and accuracy of the diagnosis. Experiments are conducted on the rolling bearing fault dataset from Case Western Reserve University. Results demonstrate that the proposed method achieves an accuracy of 96.32% using evidence fusion of the drive-end and fan-end data, which is obviously higher than that seen in preliminary single-channel diagnosis. Meanwhile, the final results can give suggestions of the possibilities of anther, which is benefit for technicists seeking to investigate the actual situation. Full article
(This article belongs to the Special Issue Control and Security of Industrial Cyber–Physical Systems)
Show Figures

Figure 1

35 pages, 14047 KB  
Article
Wildfire Susceptibility Mapping Using Deep Learning and Machine Learning Models Based on Multi-Sensor Satellite Data Fusion: A Case Study of Serbia
by Uroš Durlević, Velibor Ilić and Aleksandar Valjarević
Fire 2025, 8(10), 407; https://doi.org/10.3390/fire8100407 - 20 Oct 2025
Abstract
To prevent or mitigate the negative impact of fires, spatial prediction maps of wildfires are created to identify susceptible locations and key factors that influence the occurrence of fires. This study uses artificial intelligence models, specifically machine learning (XGBoost) and deep learning (Kolmogorov-Arnold [...] Read more.
To prevent or mitigate the negative impact of fires, spatial prediction maps of wildfires are created to identify susceptible locations and key factors that influence the occurrence of fires. This study uses artificial intelligence models, specifically machine learning (XGBoost) and deep learning (Kolmogorov-Arnold networks—KANs, and deep neural network—DNN), with data obtained from multi-sensor satellite imagery (MODIS, VIIRS, Sentinel-2, Landsat 8/9) for spatial modeling wildfires in Serbia (88,361 km2). Based on geographic information systems (GIS) and 199,598 wildfire samples, 16 quantitative variables (geomorphological, climatological, hydrological, vegetational, and anthropogenic) are presented, together with 3 synthesis maps and an integrated susceptibility map of the 3 applied models. The results show a varying percentage of Serbia’s very high vulnerability to wildfires (XGBoost = 11.5%; KAN = 14.8%; DNN = 15.2%; Ensemble = 12.7%). Among the applied models, the DNN achieved the highest predictive performance (Accuracy = 83.4%, ROC-AUC = 92.3%), followed by XGBoost and KANs, both of which also demonstrated strong predictive accuracy (ROC-AUC > 90%). These results confirm the robustness of deep and machine learning approaches for wildfire susceptibility mapping in Serbia. SHAP analysis determined that the most influential factors are elevation, air temperature, and humidity regime (precipitation, aridity, and series of consecutive dry/wet days). Full article
Show Figures

Graphical abstract

18 pages, 5353 KB  
Communication
A Reconfigurable Memristor-Based Computing-in-Memory Circuit for Content-Addressable Memory in Sensor Systems
by Hao Hu, Yian Liu, Shuang Liu, Junjie Wang, Siyu Xiao, Shiqin Yan, Ruicheng Pan, Yang Wang, Xingyu Liao, Tianhao Mao, Yutong Chen, Xiangzhan Wang and Yang Liu
Sensors 2025, 25(20), 6464; https://doi.org/10.3390/s25206464 - 19 Oct 2025
Viewed by 223
Abstract
To meet the demand for energy-efficient and high-performance computing in resource-limited sensor edge applications, this paper presents a reconfigurable memristor-based computing-in-memory circuit for Content-Addressable Memory (CAM). The scheme exploits the analog multi-level resistance characteristics of memristors to enable parallel multi-bit processing, overcoming the [...] Read more.
To meet the demand for energy-efficient and high-performance computing in resource-limited sensor edge applications, this paper presents a reconfigurable memristor-based computing-in-memory circuit for Content-Addressable Memory (CAM). The scheme exploits the analog multi-level resistance characteristics of memristors to enable parallel multi-bit processing, overcoming the constraints of traditional binary computing and significantly improving storage density and computational efficiency. Furthermore, by employing dynamic adjustment of the mapping between input signals and reference voltages, the circuit supports dynamic switching between exact and approximate CAM modes, substantially enhancing functional flexibility. Experimental results demonstrate that the 32 × 36 memristor array based on a TiN/TiOx/HfO2/TiN structure exhibits eight stable and distinguishable resistance states with excellent retention characteristics. In large-scale array simulations, the minimum voltage separation between state-representing waveforms exceeds 6.5 mV, ensuring reliable discrimination by the readout circuit. This work provides an efficient and scalable hardware solution for intelligent edge computing in next-generation sensor networks, particularly suitable for real-time biometric recognition, distributed sensor data fusion, and lightweight artificial intelligence inference, effectively reducing system dependence on cloud communication and overall power consumption. Full article
Show Figures

Figure 1

19 pages, 1603 KB  
Article
BiLSTM-LN-SA: A Novel Integrated Model with Self-Attention for Multi-Sensor Fire Detection
by Zhaofeng He, Yu Si, Liyuan Yang, Nuo Xu, Xinglong Zhang, Mingming Wang and Xiaoyun Sun
Sensors 2025, 25(20), 6451; https://doi.org/10.3390/s25206451 - 18 Oct 2025
Viewed by 233
Abstract
Multi-sensor fire detection technology has been widely adopted in practical applications; however, existing methods still suffer from high false alarm rates and inadequate adaptability in complex environments due to their limited capacity to capture deep time-series dependencies in sensor data. To enhance robustness [...] Read more.
Multi-sensor fire detection technology has been widely adopted in practical applications; however, existing methods still suffer from high false alarm rates and inadequate adaptability in complex environments due to their limited capacity to capture deep time-series dependencies in sensor data. To enhance robustness and accuracy, this paper proposes a novel model named BiLSTM-LN-SA, which integrates a Bidirectional Long Short-Term Memory (BiLSTM) network with Layer Normalization (LN) and a Self-Attention (SA) mechanism. The BiLSTM module extracts intricate time-series features and long-term dependencies. The incorporation of Layer Normalization mitigates feature distribution shifts across different environments, thereby improving the model’s adaptability to cross-scenario data and its generalization capability. Simultaneously, the Self-Attention mechanism dynamically recalibrates the importance of features at different time steps, adaptively enhancing fire-critical information and enabling deeper, process-aware feature fusion. Extensive evaluation on a real-world dataset demonstrates the superiority of the BiLSTM-LN-SA model, which achieves a test accuracy of 98.38%, an F1-score of 0.98, and an AUC of 0.99, significantly outperforming existing methods including EIF-LSTM, rTPNN, and MLP. Notably, the model also maintains low false positive and false negative rates of 1.50% and 1.85%, respectively. Ablation studies further elucidate the complementary roles of each component: the self-attention mechanism is pivotal for dynamic feature weighting, while layer normalization is key to stabilizing the learning process. This validated design confirms the model’s strong generalization capability and practical reliability across varied environmental scenarios. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

20 pages, 11855 KB  
Article
High-Precision Extrinsic Calibration for Multi-LiDAR Systems with Narrow FoV via Synergistic Planar and Circular Features
by Xinbao Sun, Zhi Zhang, Shuo Xu and Jinyue Liu
Sensors 2025, 25(20), 6432; https://doi.org/10.3390/s25206432 - 17 Oct 2025
Viewed by 300
Abstract
Precise extrinsic calibration is a fundamental prerequisite for data fusion in multi-LiDAR systems. However, conventional methods are often encumbered by dependencies on initial estimates, auxiliary sensors, or manual feature selection, which renders them complex, time-consuming, and limited in adaptability across diverse environments. To [...] Read more.
Precise extrinsic calibration is a fundamental prerequisite for data fusion in multi-LiDAR systems. However, conventional methods are often encumbered by dependencies on initial estimates, auxiliary sensors, or manual feature selection, which renders them complex, time-consuming, and limited in adaptability across diverse environments. To address these limitations, this paper proposes a novel, high-precision extrinsic calibration method for multi-LiDAR systems with a narrow Field of View (FoV), achieved through the synergistic use of circular and planar features. Our approach commences with the automatic segmentation of the calibration target’s point cloud using an improved VoxelNet. Subsequently, a denoising step, combining RANSAC and a Gaussian Mean Intensity Filter (GMIF), is applied to ensure high-quality feature extraction. From the refined point cloud, planar and circular features are robustly extracted via Principal Component Analysis (PCA) and least-squares fitting, respectively. Finally, the extrinsic parameters are optimized by minimizing a nonlinear objective function formulated with joint constraints from both geometric features. Simulation results validate the high precision of our method, with rotational and translational errors contained within 0.08° and 0.8 cm. Furthermore, real-world experiments confirm its effectiveness and superiority, outperforming conventional point-cloud registration techniques. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

22 pages, 1678 KB  
Article
Image Completion Network Considering Global and Local Information
by Yubo Liu, Ke Chen and Alan Penn
Buildings 2025, 15(20), 3746; https://doi.org/10.3390/buildings15203746 - 17 Oct 2025
Viewed by 174
Abstract
Accurate depth image inpainting in complex urban environments remains a critical challenge due to occlusions, reflections, and sensor limitations, which often result in significant data loss. We propose a hybrid deep learning framework that explicitly combines local and global modelling through Convolutional Neural [...] Read more.
Accurate depth image inpainting in complex urban environments remains a critical challenge due to occlusions, reflections, and sensor limitations, which often result in significant data loss. We propose a hybrid deep learning framework that explicitly combines local and global modelling through Convolutional Neural Networks (CNNs) and Transformer modules. The model employs a multi-branch parallel architecture, where the CNN branch captures fine-grained local textures and edges, while the Transformer branch models global semantic structures and long-range dependencies. We introduce an optimized attention mechanism, Agent Attention, which differs from existing efficient/linear attention methods by using learnable proxy tokens tailored for urban scene categories (e.g., façades, sky, ground). A content-guided dynamic fusion module adaptively combines multi-scale features to enhance structural alignment and texture recovery. The frame-work is trained with a composite loss function incorporating pixel accuracy, perceptual similarity, adversarial realism, and structural consistency. Extensive experiments on the Paris StreetView dataset demonstrate that the proposed method achieves state-of-the-art performance, outperforming existing approaches in PSNR, SSIM, and LPIPS metrics. The study highlights the potential of multi-scale modeling for urban depth inpainting and discusses challenges in real-world deployment, ethical considerations, and future directions for multimodal integration. Full article
Show Figures

Figure 1

37 pages, 1690 KB  
Review
Advances in Crop Row Detection for Agricultural Robots: Methods, Performance Indicators, and Scene Adaptability
by Zhen Ma, Xinzhong Wang, Xuegeng Chen, Bin Hu and Jingbin Li
Agriculture 2025, 15(20), 2151; https://doi.org/10.3390/agriculture15202151 - 16 Oct 2025
Viewed by 414
Abstract
Crop row detection technology, as one of the key technologies for agricultural robots to achieve autonomous navigation and precise operations, is related to the precision and stability of agricultural machinery operations. Its research and development will also significantly determine the development process of [...] Read more.
Crop row detection technology, as one of the key technologies for agricultural robots to achieve autonomous navigation and precise operations, is related to the precision and stability of agricultural machinery operations. Its research and development will also significantly determine the development process of intelligent agriculture. The paper first summarizes the mainstream technical methods, performance evaluation systems, and adaptability analysis of typical agricultural scenes for crop row detection. The paper also summarizes and explains the technical principles and characteristics of traditional methods based on visual sensors, point cloud preprocessing based on LiDAR, line structure extraction and 3D feature calculation methods, and multi-sensor fusion methods. Secondly, a review was conducted on performance evaluation criteria such as accuracy, efficiency, robustness, and practicality, analyzing and comparing the applicability of different methods in typical scenarios such as open fields, facility agriculture, orchards, and special terrains. Based on the multidimensional analysis above, it is concluded that a single technology has specific environmental adaptability limitations. Multi-sensor fusion can help improve robustness in complex scenarios, and the fusion advantage will gradually increase with the increase in the number of sensors. Suggestions on the development of agricultural robot navigation technology are made based on the current status of technological applications in the past five years and the needs for future development. This review systematically summarizes crop row detection technology, providing a clear technical framework and scenario adaptation reference for research in this field, and striving to promote the development of precision and efficiency in agricultural production. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

28 pages, 1236 KB  
Article
Transfer Entropy-Based Causal Inference for Industrial Alarm Overload Mitigation
by Yaofang Zhang, Haikuo Qu, Yang Liu, Hongri Liu and Bailing Wang
Electronics 2025, 14(20), 4066; https://doi.org/10.3390/electronics14204066 - 16 Oct 2025
Viewed by 180
Abstract
In tightly coupled Industrial Control Systems (ICS), abnormal disturbances often propagate throughout the process, triggering a large number of time-correlated alarms that exceed the handling capacity of the operator. Consequently, a key challenge is how to leverage the directional and temporal characteristics of [...] Read more.
In tightly coupled Industrial Control Systems (ICS), abnormal disturbances often propagate throughout the process, triggering a large number of time-correlated alarms that exceed the handling capacity of the operator. Consequently, a key challenge is how to leverage the directional and temporal characteristics of disturbance propagation to alleviate alarm overload. This paper proposes a delay-sensitive causal inference approach for industrial alarm analysis to address this problem. On the one hand, time delay estimation is introduced to precisely align the responses of two sensor sequences to disturbances, thereby improving the accuracy of causal relationship identification in the temporal domain. On the other hand, a multi-scale subgraph fusion strategy is designed to address the inconsistency in causal strength caused by disturbances of varying intensities. By integrating significant causal subgraphs from multiple scenarios into a unified graph, the method reveals the overall causal structure among alarm variables and provides guidance for alarm mitigation. To validate the proposed method, a case study is conducted on the Tennessee Eastman Process. The results demonstrate that the approach identifies causal relationships more accurately and reasonably and can effectively reduce the number of alarms by up to 51.6%. Full article
Show Figures

Figure 1

19 pages, 2733 KB  
Article
Style Transfer from Sentinel-1 to Sentinel-2 for Fluvial Scenes with Multi-Modal and Multi-Temporal Image Fusion
by Patrice E. Carbonneau
Remote Sens. 2025, 17(20), 3445; https://doi.org/10.3390/rs17203445 - 15 Oct 2025
Viewed by 199
Abstract
Recently, there has been significant progress in the area of semantic classification of water bodies at global scales with deep learning. For the key purposes of water inventory and change detection, advanced deep learning classifiers such as UNets and Vision Transformers have been [...] Read more.
Recently, there has been significant progress in the area of semantic classification of water bodies at global scales with deep learning. For the key purposes of water inventory and change detection, advanced deep learning classifiers such as UNets and Vision Transformers have been shown to be both accurate and flexible when applied to large-scale, or even global, satellite image datasets from optical (e.g., Sentinel-2) and radar sensors (e.g., Sentinel-1). Most of this work is conducted with optical sensors, which usually have better image quality, but their obvious limitation is cloud cover, which is why radar imagery is an important complementary dataset. However, radar imagery is generally more sensitive to soil moisture than optical data. Furthermore, topography and wind-ripple effects can alter the reflected intensity of radar waves, which can induce errors in water classification models that fundamentally rely on the fact that water is darker than the surrounding landscape. In this paper, we develop a solution to the use of Sentinel-1 radar images for the semantic classification of water bodies that uses style transfer with multi-modal and multi-temporal image fusion. Instead of developing new semantic classification models that work directly on Sentinel-1 images, we develop a global style transfer model that produces synthetic Sentinel-2 images from Sentinel-1 input. The resulting synthetic Sentinel-2 imagery can then be classified with existing models. This has the advantage of obviating the need for large volumes of manually labeled Sentinel-1 water masks. Next, we show that fusing an 8-year cloud-free composite of the near-infrared band 8 of Sentinel-2 to the input Sentinel-1 image improves the classification performance. Style transfer models were trained and validated with global scale data covering the years 2017 to 2024, and include every month of the year. When tested against a global independent benchmark, S1S2-Water, the semantic classifications produced from our synthetic imagery show a marked improvement with the use of image fusion. When we use only Sentinel-1 data, we find an overall IoU (Intersection over Union) score of 0.70, but when we add image fusion, the overall IoU score rises to 0.93. Full article
(This article belongs to the Special Issue Multimodal Remote Sensing Data Fusion, Analysis and Application)
Show Figures

Figure 1

29 pages, 7541 KB  
Article
An Underwater Salvage Robot for Retrieving Foreign Objects in Nuclear Reactor Pools
by Ming Zhong, Zihan Gao, Zhengxiong Mao, Ruifei Lyu and Yaxin Liu
Drones 2025, 9(10), 714; https://doi.org/10.3390/drones9100714 - 15 Oct 2025
Viewed by 252
Abstract
In this paper, an underwater salvage robot is developed to retrieve foreign objects scattered in nuclear reactor pools. The robot mainly consists of an ROV platform and a 3-DOF Delta robotic arm. Utilizing fused IMU and LED beacon visual data for localization, it [...] Read more.
In this paper, an underwater salvage robot is developed to retrieve foreign objects scattered in nuclear reactor pools. The robot mainly consists of an ROV platform and a 3-DOF Delta robotic arm. Utilizing fused IMU and LED beacon visual data for localization, it achieves pool traversal via six dynamically controlled thrusters. An improved YOLOv8s algorithm is employed to identify foreign objects in underwater environments. During traversal, the robot identifies and retrieves foreign objects along the way. The prototype of the robot was subjected to a series of experiments in an indoor pool. Results show that the improved YOLOv8 algorithm achieves 92.2% mAP, surpassing the original YOLOv8s and Faster-RCNN by 3.7 and 3.3 percentage points, respectively. The robot achieved a foreign-object identification rate of 95.42% and a retrieval success rate of 90.64% under dynamic traversal conditions, indicating that it meets the operational requirements and has significant engineering application value. Full article
Show Figures

Figure 1

17 pages, 3498 KB  
Article
Self-Supervised Learning and Multi-Sensor Fusion for Alpine Wetland Vegetation Mapping: Bayinbuluke, China
by Muhammad Murtaza Zaka, Alim Samat, Jilili Abuduwaili, Enzhao Zhu, Arslan Akhtar and Wenbo Li
Plants 2025, 14(20), 3153; https://doi.org/10.3390/plants14203153 - 13 Oct 2025
Viewed by 641
Abstract
Accurate mapping of wetland vegetation is essential for ecological monitoring and conservation, yet it remains challenging due to the spatial heterogeneity of wetlands, the scarcity of ground-truth data, and the spread of invasive species. Invasive plants alter native vegetation patterns, making their early [...] Read more.
Accurate mapping of wetland vegetation is essential for ecological monitoring and conservation, yet it remains challenging due to the spatial heterogeneity of wetlands, the scarcity of ground-truth data, and the spread of invasive species. Invasive plants alter native vegetation patterns, making their early detection critical for preserving ecosystem integrity. This study proposes a novel framework that integrates self-supervised learning (SSL), supervised segmentation, and multi-sensor data fusion to enhance vegetation classification in the Bayinbuluke Alpine Wetland, China. High-resolution satellite imagery from PlanetScope-3 and Jilin-1 was fused, and SSL methods—including BYOL, DINO, and MoCo v3—were employed to learn transferable feature representations without extensive labeled data. The results show that SSL methods exhibit consistent variations in classification performance, while multi-sensor fusion significantly improves the detection of rare and fragmented vegetation patches and enables the early identification of invasive species. Overall, the proposed SSL–fusion strategy reduces reliance on labor-intensive field data collection and provides a scalable, high-precision solution for wetland monitoring and invasive species management. Full article
(This article belongs to the Special Issue Computer Vision Techniques for Plant Phenomics Applications)
Show Figures

Figure 1

30 pages, 23104 KB  
Article
MSAFNet: Multi-Modal Marine Aquaculture Segmentation via Spatial–Frequency Adaptive Fusion
by Guolong Wu and Yimin Lu
Remote Sens. 2025, 17(20), 3425; https://doi.org/10.3390/rs17203425 - 13 Oct 2025
Viewed by 378
Abstract
Accurate mapping of marine aquaculture areas is critical for environmental management and sustainable development for marine ecosystem protection and sustainable resource utilization. However, remote sensing imagery based on single-sensor modalities has inherent limitations when extracting aquaculture zones in complex marine environments. To address [...] Read more.
Accurate mapping of marine aquaculture areas is critical for environmental management and sustainable development for marine ecosystem protection and sustainable resource utilization. However, remote sensing imagery based on single-sensor modalities has inherent limitations when extracting aquaculture zones in complex marine environments. To address this challenge, we constructed a multi-modal dataset from five Chinese coastal regions using cloud detection methods and developed Multi-modal Spatial–Frequency Adaptive Fusion Network (MSAFNet) for optical-radar data fusion. MSAFNet employs a dual-path architecture utilizing a Multi-scale Dual-path Feature Module (MDFM) that combines CNN and Transformer capabilities to extract multi-scale features. Additionally, it implements a Dynamic Frequency Domain Adaptive Fusion Module (DFAFM) to achieve deep integration of multi-modal features in both spatial and frequency domains, effectively leveraging the complementary advantages of different sensor data. Results demonstrate that MSAFNet achieves 76.93% mean intersection over union (mIoU), 86.96% mean F1 score (mF1), and 93.26% mean Kappa coefficient (mKappa) in extracting floating raft aquaculture (FRA) and cage aquaculture (CA), significantly outperforming existing methods. Applied to China’s coastal waters, the model generated 2020 nearshore aquaculture distribution maps, demonstrating its generalization capability and practical value in complex marine environments. This approach provides reliable technical support for marine resource management and ecological monitoring. Full article
Show Figures

Figure 1

21 pages, 3794 KB  
Article
DEIM-SFA: A Multi-Module Enhanced Model for Accurate Detection of Weld Surface Defects
by Yan Sun, Yingjie Xie, Ran Peng, Yifan Zhang, Wei Chen and Yan Guo
Sensors 2025, 25(20), 6314; https://doi.org/10.3390/s25206314 - 13 Oct 2025
Viewed by 667
Abstract
High-precision automated detection of metal welding defects is critical to ensuring structural safety and reliability in modern manufacturing. However, existing methods often struggle with insufficient fine-grained feature retention, low efficiency in multi-scale information fusion, and vulnerability to complex background interference, resulting in low [...] Read more.
High-precision automated detection of metal welding defects is critical to ensuring structural safety and reliability in modern manufacturing. However, existing methods often struggle with insufficient fine-grained feature retention, low efficiency in multi-scale information fusion, and vulnerability to complex background interference, resulting in low detection accuracy. This work addresses the limitations by introducing the DEIM-SFA, a novel detection framework designed for automated visual inspection in industrial machine vision sensors. The model introduces a novel structure-aware dynamic convolution (SPD-Conv), effectively focusing on the fine-grained structure of defects while suppressing background noise interference; an innovative multi-scale dynamic fusion pyramid (FTPN) architecture is designed to achieve efficient and adaptive aggregation of feature information from different receptive fields, ensuring consistent detection of multi-scale targets; combined with a lightweight and efficient multi-scale attention module (EMA), this further enhances the model’s ability to locate salient regions in complex scenarios. The network is evaluated on a self-built welding defect dataset. Experimental results show that DEIM-SFA achieves a 3.9% improvement in mAP50 compared to the baseline model, mAP75 by 4.3%, mAP50–95 by 3.7%, and Recall by 1.4%. The model exhibits consistently significant superiority in detection accuracy across targets of various sizes, while maintaining well-balanced model complexity and inference efficiency, comprehensively surpassing existing state-of-the-art (SOTA) methods. Full article
(This article belongs to the Section Industrial Sensors)
Show Figures

Figure 1

Back to TopTop