Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (274)

Search Parameters:
Keywords = refinement mask

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 16950 KB  
Article
Using High-Resolution Satellite Imagery and Deep Learning to Map Artisanal Mining Spatial Extent in the Democratic Republic of the Congo
by Francesco Pasanisi, Robert N. Masolele and Johannes Reiche
Remote Sens. 2025, 17(24), 4057; https://doi.org/10.3390/rs17244057 - 18 Dec 2025
Abstract
Artisanal and Small-scale Mining (ASM) significantly impacts the Democratic Republic of Congo’s (DRC) socio-economic landscape and environmental integrity, yet its dynamic and informal nature makes monitoring challenging. This study addresses this challenge by implementing a novel deep learning approach to map ASM sites [...] Read more.
Artisanal and Small-scale Mining (ASM) significantly impacts the Democratic Republic of Congo’s (DRC) socio-economic landscape and environmental integrity, yet its dynamic and informal nature makes monitoring challenging. This study addresses this challenge by implementing a novel deep learning approach to map ASM sites across the DRC using satellite imagery. We tackled key obstacles including ground truth data scarcity, insufficient spatial resolution of conventional satellite sensors, and persistent cloud cover in the region. We developed a methodology to generate a pseudo-ground truth dataset by converting point-based ASM locations to segmented areas through a multi-stage process involving clustering, auxiliary dataset masking, and manual refinement. Four model configurations were evaluated: Planet-NICFI standalone, Sentinel-1 standalone, Early Fusion, and Late Fusion approaches. The Late Fusion model, which integrated high-resolution Planet-NICFI optical imagery (4.77 m resolution) with Sentinel-1 SAR data, achieved the highest performance with an average precision of 71%, recall of 75%, and F1-score of 73% for ASM detection. This superior performance demonstrated how SAR data’s textural features complemented optical data’s spectral information, particularly improving discrimination between ASM sites and water bodies—a common source of misclassification in optical-only approaches. We deployed the optimized model to map ASM extent in the Mwenga territory, achieving an overall accuracy of 88.4% when validated against high-resolution reference imagery. Despite these achievements, challenges persist in distinguishing ASM sites from built-up areas, suggesting avenues for future research through multi-class approaches. This study advances the domain of ASM mapping by offering methodologies that enhance remote sensing capabilities in ASM-impacted regions, providing valuable tools for monitoring, regulation, and environmental management. Full article
Show Figures

Figure 1

20 pages, 14411 KB  
Article
An Integrated Framework with SAM and OCR for Pavement Crack Quantification and Geospatial Mapping
by Nut Sovanneth, Asnake Adraro Angelo, Felix Obonguta and Kiyoyuki Kaito
Infrastructures 2025, 10(12), 348; https://doi.org/10.3390/infrastructures10120348 - 15 Dec 2025
Viewed by 157
Abstract
Pavement condition assessment using computer vision has emerged as an efficient alternative to traditional manual surveys, which are often labor-intensive and time-consuming. Leveraging deep learning, pavement distress such as cracks can be automatically detected, segmented, and quantified from high-resolution images captured by survey [...] Read more.
Pavement condition assessment using computer vision has emerged as an efficient alternative to traditional manual surveys, which are often labor-intensive and time-consuming. Leveraging deep learning, pavement distress such as cracks can be automatically detected, segmented, and quantified from high-resolution images captured by survey vehicles. Although numerous segmentation models have been proposed to generate crack masks, they typically require extensive pixel-level annotations, leading to high labeling costs. To overcome this limitation, this study integrates the Segmentation Anything Model (SAM), which produces accurate segmentation masks from simple bounding box prompts while leveraging its zero-shot capability to generalize to unseen images with minimal retraining. However, since SAM alone is not an end-to-end solution, we incorporate YOLOv8 for automated crack detection, eliminating the need for manual box annotation. Furthermore, the framework applies local refinement techniques to enhance mask precision and employs Optical Character Recognition (OCR) to automatically extract embedded GPS coordinates for geospatial mapping. The proposed framework is empirically validated using open-source pavement images from Yamanashi, demonstrating effective automated detection, classification, quantification, and geospatial mapping of pavement cracks. The results support automated pavement distress mapping onto real-world road networks, facilitating efficient maintenance planning for road agencies. Full article
Show Figures

Figure 1

22 pages, 1479 KB  
Article
VMPANet: Vision Mamba Skin Lesion Image Segmentation Model Based on Prompt and Attention Mechanism Fusion
by Zinuo Peng, Shuxian Liu and Chenhao Li
J. Imaging 2025, 11(12), 443; https://doi.org/10.3390/jimaging11120443 - 11 Dec 2025
Viewed by 156
Abstract
In the realm of medical image processing, the segmentation of dermatological lesions is a pivotal technique for the early detection of skin cancer. However, existing methods for segmenting images of skin lesions often encounter limitations when dealing with intricate boundaries and diverse lesion [...] Read more.
In the realm of medical image processing, the segmentation of dermatological lesions is a pivotal technique for the early detection of skin cancer. However, existing methods for segmenting images of skin lesions often encounter limitations when dealing with intricate boundaries and diverse lesion shapes. To address these challenges, we propose VMPANet, designed to accurately localize critical targets and capture edge structures. VMPANet employs an inverted pyramid convolution to extract multi-scale features while utilizing the visual Mamba module to capture long-range dependencies among image features. Additionally, we leverage previously extracted masks as cues to facilitate efficient feature propagation. Furthermore, VMPANet integrates parallel depthwise separable convolutions to enhance feature extraction and introduces innovative mechanisms for edge enhancement, spatial attention, and channel attention to adaptively extract edge information and complex spatial relationships. Notably, VMPANet refines a novel cross-attention mechanism, which effectively facilitates the interaction between deep semantic cues and shallow texture details, thereby generating comprehensive feature representations while reducing computational load and redundancy. We conducted comparative and ablation experiments on two public skin lesion datasets (ISIC2017 and ISIC2018). The results demonstrate that VMPANet outperforms existing mainstream methods. On the ISIC2017 dataset, its mIoU and DSC metrics are 1.38% and 0.83% higher than those of VM-Unet respectively; on the ISIC2018 dataset, these metrics are 1.10% and 0.67% higher than those of EMCAD, respectively. Moreover, VMPANet boasts a parameter count of only 0.383 M and a computational load of 1.159 GFLOPs. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

28 pages, 15780 KB  
Article
Towards Near-Real-Time Estimation of Live Fuel Moisture Content from Sentinel-2 for Fire Management in Northern Thailand
by Chakrit Chotamonsak, Duangnapha Lapyai and Punnathorn Thanadolmethaphorn
Fire 2025, 8(12), 475; https://doi.org/10.3390/fire8120475 - 11 Dec 2025
Viewed by 205
Abstract
Wildfires are a recurring dry-season hazard in northern Thailand, contributing to severe air pollution and trans-boundary haze. However, the region lacks the ground-based measurements necessary for monitoring Live Fuel Moisture Content (LFMC), a key variable influencing vegetation flammability. This study presents a preliminary [...] Read more.
Wildfires are a recurring dry-season hazard in northern Thailand, contributing to severe air pollution and trans-boundary haze. However, the region lacks the ground-based measurements necessary for monitoring Live Fuel Moisture Content (LFMC), a key variable influencing vegetation flammability. This study presents a preliminary framework for near-real-time (NRT) LFMC estimation using Sentinel-2 multispectral imagery. The system integrates normalized vegetation and moisture-related indices, including the Normalized Difference Vegetation Index (NDVI), the Normalized Difference Infrared Index (NDII), and the Moisture Stress Index (MSI) with an NDVI-derived evapotranspiration fraction (ETf) within a heuristic modeling approach. The workflow includes cloud and shadow masking, weekly to biweekly compositing, and pixel-wise normalization to address the persistent cloud cover and heterogeneous land surfaces. Although currently unvalidated, the LFMC estimates capture the relative spatial and temporal variations in vegetation moisture across northern Thailand during the 2024 dry season (January–April). Evergreen forests maintained higher moisture levels, whereas deciduous forests and agricultural landscapes exhibited pronounced drying from January to March. Short-lag responses to rainfall suggest modest moisture recovery following precipitation, although the relationship is influenced by additional climatic and ecological factors not represented in the heuristic model. LFMC-derived moisture classes reflect broad seasonal dryness patterns but should not be interpreted as direct fire danger indicators. This study demonstrates the feasibility of generating regional LFMC indicators in a data-scarce tropical environment and outlines a clear pathway for future calibration and validation, including field sampling, statistical optimization, and benchmarking against global LFMC products. Until validated, the proposed NRT LFMC estimation product should be used to assess relative vegetation dryness and to support the refinement and development of future operational fire management tools, including early warnings, burn-permit regulation, and resource allocation. Full article
(This article belongs to the Section Fire Science Models, Remote Sensing, and Data)
Show Figures

Figure 1

28 pages, 29492 KB  
Article
RSAM: Vision-Language Two-Way Guidance for Referring Remote Sensing Image Segmentation
by Zilong Zhao, Xin Xu, Bingxin Huang, Hongjia Chen and Fangling Pu
Remote Sens. 2025, 17(24), 3960; https://doi.org/10.3390/rs17243960 - 8 Dec 2025
Viewed by 261
Abstract
Referring remote sensing image segmentation (RRSIS) aims to accurately segment target objects in remote sensing images based on natural language instructions. Despite its growing relevance, progress in this field is constrained by limited datasets and weak cross-modal alignment. To support RRSIS research, we [...] Read more.
Referring remote sensing image segmentation (RRSIS) aims to accurately segment target objects in remote sensing images based on natural language instructions. Despite its growing relevance, progress in this field is constrained by limited datasets and weak cross-modal alignment. To support RRSIS research, we construct referring image segmentation in optical remote sensing (RISORS), a large-scale benchmark containing 36,697 instruction–mask pairs. RISORS provides diverse and high-quality samples that enable comprehensive experiment in remote sensing contexts. Building on this foundation, we propose Referring-SAM (RSAM), a novel framework that extends Segment Anything Model 2 to support text-prompted segmentation. RSAM integrates a Two-Way Guidance Module (TWGM) and a Multimodal Mask Decoder (MMMD). TWGM facilitates a two-way guidance mechanism that mutually refines image and text features, with positional encodings incorporated across all attention layers to significantly enhance relational reasoning. MMMD effectively separates textual prompts from spatial prompts, improving segmentation accuracy in complex multimodal settings. Extensive experiments on RISORS, as well as RefSegRS and RRSIS-D datasets, demonstrate that RSAM achieves state-of-the-art performance, particularly in segmenting small and diverse targets. Ablation studies further validate the individual contributions of TWGM and MMMD. This work provides a solid foundation for further developments in integrated vision-language analysis within remote sensing applications. Full article
Show Figures

Figure 1

24 pages, 3036 KB  
Article
MPG-SwinUMamba: High-Precision Segmentation and Automated Measurement of Eye Muscle Area in Live Sheep Based on Deep Learning
by Zhou Zhang, Yaojing Yue, Fuzhong Li, Leifeng Guo and Svitlana Pavlova
Animals 2025, 15(24), 3509; https://doi.org/10.3390/ani15243509 - 5 Dec 2025
Viewed by 220
Abstract
Accurate EMA assessment in live sheep is crucial for genetic breeding and production management within the meat sheep industry. However, the segmentation accuracy and reliability of existing automated methods are limited by challenges inherent to B-mode ultrasound images, such as low contrast and [...] Read more.
Accurate EMA assessment in live sheep is crucial for genetic breeding and production management within the meat sheep industry. However, the segmentation accuracy and reliability of existing automated methods are limited by challenges inherent to B-mode ultrasound images, such as low contrast and noise interference. To address these challenges, we present MPG-SwinUMamba, a novel deep learning-based segmentation network. This model uniquely combines the state-space model with a U-Net architecture. It also integrates an edge-enhancement multi-scale attention module (MSEE) and a pyramid attention refinement module (PARM) to improve the detection of indistinct boundaries and better capture global context. The global context aggregation decoder (GCAD) is employed to precisely reconstruct the segmentation mask, enabling automated measurement of the EMA. Compared to 12 other leading segmentation models, MPG-SwinUMamba achieved superior performance, with an intersection-over-union of 91.62% and a Dice similarity coefficient of 95.54%. Additionally, automated measurements show excellent agreement with expert manual assessments (correlation coefficient r = 0.9637), with a mean absolute percentage error of only 4.05%. This method offers non-invasive and efficient and objective evaluation of carcass performance in live sheep, with the potential to reduce measurement costs and enhance breeding efficiency. Full article
(This article belongs to the Section Animal System and Management)
Show Figures

Figure 1

19 pages, 2788 KB  
Article
Universal Image Segmentation with Arbitrary Granularity for Efficient Pest Monitoring
by L. Minh Dang, Sufyan Danish, Muhammad Fayaz, Asma Khan, Gul E. Arzu, Lilia Tightiz, Hyoung-Kyu Song and Hyeonjoon Moon
Horticulturae 2025, 11(12), 1462; https://doi.org/10.3390/horticulturae11121462 - 3 Dec 2025
Viewed by 286
Abstract
Accurate and timely pest monitoring is essential for sustainable agriculture and effective crop protection. While recent deep learning-based pest recognition systems have significantly improved accuracy, they are typically trained for fixed label sets and narrowly defined tasks. In this paper, we present RefPestSeg, [...] Read more.
Accurate and timely pest monitoring is essential for sustainable agriculture and effective crop protection. While recent deep learning-based pest recognition systems have significantly improved accuracy, they are typically trained for fixed label sets and narrowly defined tasks. In this paper, we present RefPestSeg, a universal, language-promptable segmentation model specifically designed for pest monitoring. RefPestSeg can segment targets at any semantic level, such as species, genus, life stage, or damage type, conditioned on flexible natural language instructions. The model adopts a symmetric architecture with self-attention and cross-attention mechanisms to tightly align visual features with language embeddings in a unified feature space. To further enhance performance in challenging field conditions, we integrate an optimized super-resolution module to improve image quality and employ diverse data augmentation strategies to enrich the training distribution. A lightweight postprocessing step refines segmentation masks by suppressing highly overlapping regions and removing noise blobs introduced by cluttered backgrounds. Extensive experiments on a challenging pest dataset show that RefPestSeg achieves an Intersection over Union (IoU) of 69.08 while maintaining robustness in real-world scenarios. By enabling language-guided pest segmentation, RefPestSeg advances toward more intelligent, adaptable monitoring systems that can respond to real-time agricultural demands without costly model retraining. Full article
Show Figures

Figure 1

24 pages, 11779 KB  
Article
Effective SAR Image Despeckling Using Noise-Guided Transformer and Multi-Scale Feature Fusion
by Linna Zhang, Le Zheng, Yuxin Wen, Fugui Zhang, Fuyu Bo and Yigang Cen
Remote Sens. 2025, 17(23), 3863; https://doi.org/10.3390/rs17233863 - 28 Nov 2025
Viewed by 263
Abstract
Speckle noise is a significant challenge in synthetic aperture radar (SAR) images, severely degrading the visual quality and compromising subsequent image interpretation tasks. While existing despeckling methods can reduce noise, they often fail to strike a appropriate balance between noise suppression and the [...] Read more.
Speckle noise is a significant challenge in synthetic aperture radar (SAR) images, severely degrading the visual quality and compromising subsequent image interpretation tasks. While existing despeckling methods can reduce noise, they often fail to strike a appropriate balance between noise suppression and the preservation of fine image details. To address this issue, in this paper, we propose a novel SAR image despeckling method that leverages both structural image priors and noise distribution characteristics in an end-to-end framework. Our approach consists of two key components: a dual-branch subnet for coarse despeckling and noise estimation, and a noise-guided Transformer-based subnet for final image refinement. The dual-branch subnet decouples the tasks of noise estimation and despeckling, improving both noise suppression accuracy and structural detail preservation. Furthermore, a combination of grouped pooling attention (GPA) and context-aware fusion (CAF) modules enables effective multi-scale feature fusion by jointly capturing local details and global contextual information. The noise estimation branch generates adaptive priors that guide the Transformer refinement, which incorporates deformable convolutions and a masked self-attention mechanism to selectively focus on relevant image regions. Extensive experiments conducted on both synthetic and real SAR datasets demonstrate that the proposed method consistently outperforms current state-of-the-art methods, achieving superior speckle suppression while preserving fine details more effectively. Full article
(This article belongs to the Special Issue SAR Images Processing and Analysis (3rd Edition))
Show Figures

Figure 1

38 pages, 7210 KB  
Article
Vision–Geometry Fusion for Measuring Pupillary Height and Interpupillary Distance via RC-BlendMask and Ensemble Regression Trees
by Shishuo Han, Zihan Yang and Huiyu Xiang
Appl. Syst. Innov. 2025, 8(6), 181; https://doi.org/10.3390/asi8060181 - 27 Nov 2025
Viewed by 493
Abstract
This study proposes an automated, visual–geometric fusion method for measuring pupillary height (PH) and interpupillary distance (PD), aiming to replace manual measurements while balancing accuracy, efficiency, and cost accessibility. To this end, a two-layer Ensemble of Regression Tree (ERT) is used to coarsely [...] Read more.
This study proposes an automated, visual–geometric fusion method for measuring pupillary height (PH) and interpupillary distance (PD), aiming to replace manual measurements while balancing accuracy, efficiency, and cost accessibility. To this end, a two-layer Ensemble of Regression Tree (ERT) is used to coarsely localize facial landmarks and the pupil center, which is then refined via direction-aware ray casting and edge-side-stratified RANSAC followed by least-squares fitting; in parallel, an RC-BlendMask instance-segmentation module extracts the lowest rim point of the spectacle lens. Head pose and lens-plane depth are estimated with the Perspective-n-Point (PnP) algorithm to enable pixel-to-millimeter calibration and pose gating, thereby achieving 3D quantification of PH/PD under a single-camera setup. In a comparative study with 30 participants against the Zeiss i.Terminal2, the proposed method achieved mean absolute errors of 1.13 mm (PD), 0.73 mm (PH-L), and 0.89 mm (PH-R); Pearson correlation coefficients were r = 0.944 (PD), 0.964 (PH-L), and 0.916 (PH-R), and Bland–Altman 95% limits of agreement were −2.00 to 2.70 mm (PD), −0.84 to 1.76 mm (PH-L), and −1.85 to 1.79 mm (PH-R). Lens segmentation performance reached a Precision of 97.5% and a Recall of 93.8%, supporting robust PH extraction. Overall, the proposed approach delivers measurement agreement comparable to high-end commercial devices on low-cost hardware, satisfies ANSI Z80.1/ISO 21987 clinical tolerances for decentration and prism error, and is suitable for both in-store dispensing and tele-dispensing scenarios. Full article
Show Figures

Figure 1

22 pages, 4907 KB  
Article
Detection of Fusarium Head Blight in Individual Wheat Spikes Using Monocular Depth Estimation with Depth Anything V2
by Jiacheng Wang, Jianliang Wang, Yuanyuan Zhao, Fei Wu, Wei Wu, Zhen Li, Chengming Sun, Tao Li and Tao Liu
Agronomy 2025, 15(11), 2651; https://doi.org/10.3390/agronomy15112651 - 19 Nov 2025
Viewed by 344
Abstract
Fusarium head blight (FHB) poses a significant threat to global wheat yields and food security, underscoring the importance of timely detection and severity assessment. Although existing approaches based on semantic segmentation and stereo vision have shown promise, their scalability is constrained by limited [...] Read more.
Fusarium head blight (FHB) poses a significant threat to global wheat yields and food security, underscoring the importance of timely detection and severity assessment. Although existing approaches based on semantic segmentation and stereo vision have shown promise, their scalability is constrained by limited training datasets and the high maintenance cost and complexity of visual sensor systems. In this study, AR glasses were employed for image acquisition, and wheat spike segmentation was performed using Depth Anything V2, a monocular depth estimation model. Through geometric localization methods—such as identifying abrupt changes in stem width—redundant elements (e.g., awns and stems) were effectively excluded, yielding high-precision spike masks (Precision: 0.945; IoU: 0.878) that outperformed leading semantic segmentation models including Mask R-CNN and DeepLabv3+. The study further conducted a comprehensive analysis of differences between diseased and healthy spikelets across RGB, HSV, and Lab color spaces, as well as three color indices: Excess Green–Excess Red (ExGR), Normalized Difference Index (NDI), and Visible Atmospherically Resistant Index (VARI). A dynamic fusion weighting strategy was developed by combining the Lab-a* component with the ExGR index, thereby enhancing visual contrast between symptomatic and asymptomatic regions. This fused index enabled quantitative assessment of FHB severity, achieving an R2 of 0.815 and an RMSE of 8.91%, indicating strong predictive accuracy. The proposed framework offers an intelligent, cost-effective solution for FHB detection, and its core methodologies—depth-guided segmentation, geometric refinement, and multi-feature fusion—present a transferable model for similar tasks in other crop segmentation applications. Full article
Show Figures

Figure 1

15 pages, 3327 KB  
Article
Investigation of the Carbonation Behavior of Cement Mortar Containing Interior Stone Sludge and Recycled Mask Fibers
by Junhyeok Choi, Seongjin Cho, Dongkyu Lee, Gwang Mok Kim, Beomjoo Yang and Daeik Jang
Materials 2025, 18(22), 5218; https://doi.org/10.3390/ma18225218 - 18 Nov 2025
Viewed by 298
Abstract
This study examines the carbonation and mechanical behavior of cement mortar incorporating artificial interior stone (AIS) sludge and recycled mask fibers (RMFs). Sludge, derived from AIS waste, replaced 30 wt.% of fine aggregate, while RMF from polypropylene masks was added at 0–1 wt.% [...] Read more.
This study examines the carbonation and mechanical behavior of cement mortar incorporating artificial interior stone (AIS) sludge and recycled mask fibers (RMFs). Sludge, derived from AIS waste, replaced 30 wt.% of fine aggregate, while RMF from polypropylene masks was added at 0–1 wt.% of cement. Specimens were cured under normal and carbonation conditions (10% CO2, 25 °C, 60% RH) for 7 and 28 days. Carbonation curing improved compressive and flexural strengths by up to 28% and 88%, respectively, and enhanced microstructural densification. Although the incorporation of AIS sludge reduced compressive strength due to its inert and irregular particle characteristics, it effectively refined the pore structure and decreased overall porosity. The inclusion of RMF at moderate contents (0.25–0.5 wt.%) improved crack resistance and lowered thermal conductivity, demonstrating a favorable balance between strength and thermal performance. TGA/DTG results confirmed increased CaCO3 formation and greater CO2 uptake. After exposure to 500 °C, carbonation-cured mortars retained higher residual strength, indicating superior thermal stability. Full article
(This article belongs to the Special Issue Advanced Concrete Formulations: Nanotechnology and Hybrid Materials)
Show Figures

Figure 1

26 pages, 61479 KB  
Article
Graph-Based Multi-Resolution Cosegmentation for Coarse-to-Fine Object-Level SAR Image Change Detection
by Jingxing Zhu, Miao Yu, Feng Wang, Guangyao Zhou, Niangang Jiao, Yuming Xiang and Hongjian You
Remote Sens. 2025, 17(22), 3736; https://doi.org/10.3390/rs17223736 - 17 Nov 2025
Viewed by 310
Abstract
The ongoing launch of high-resolution satellites has led to a significant increase in the volume of synthetic aperture radar data, resulting in a high-resolution and high-revisit Earth observation that efficiently supports subsequent high-resolution SAR change detection. To address the issues of speckle noise [...] Read more.
The ongoing launch of high-resolution satellites has led to a significant increase in the volume of synthetic aperture radar data, resulting in a high-resolution and high-revisit Earth observation that efficiently supports subsequent high-resolution SAR change detection. To address the issues of speckle noise interference, insufficient integrity of change targets and blurred boundary location of high-resolution SAR change detection, we propose a coarse-to-fine framework based on the multi-scale segmentation and hybrid structure graph (HSG), which consists of three modules: multi-scale segmentation, difference measurement, and change refinement. First, we propose a graph-based multi-resolution co-segmentation (GMRCS) in the multi-scale segmentation module to generate hierarchically nested superpixel masks. And, a two-stage ranking (TSR) strategy is designed to help GMRCS better approximate the target edges and preserve the spatio-temporal structure of changed regions. Then, we introduce a graph model and measuring difference level based on the HSG. The multi-scale difference image (DI) is generated by constructing the HSG for bi-temporal SAR images and comparing the consistency of the HSGs to reduce the effect of speckle noise. Finally, the coarse-scale change information is gradually mapped to the fine-scale based on the multi-scale fusion refinement (FR) strategy, and we can get the binary change map (BCM). Experimental results on three high-resolution SAR change detection datasets demonstrates the superiority of our proposed algorithm in preserving the integrity and structural precision of change targets compared with several state-of-the-art methods. Full article
(This article belongs to the Special Issue SAR Image Change Detection: From Hand-Crafted to Deep Learning)
Show Figures

Figure 1

23 pages, 3917 KB  
Article
Multi-Fluid Pipeline Leak Detection and Classification Using Savitzky–Golay Scalograms and Lightweight Vision Transformer Featuring Streamlined Self-Attention
by Niamat Ullah, Zahoor Ahmad and Jong-Myon Kim
Sensors 2025, 25(22), 7001; https://doi.org/10.3390/s25227001 - 16 Nov 2025
Viewed by 585
Abstract
This paper presents a novel pipeline leak diagnosis framework that combines Savitzky–Golay scalograms with a lightweight advanced deep learning architecture. Pipelines are critical for transporting fluids and gases, but leaks can lead to operational disruptions, environmental hazards, and financial losses. Leak events generate [...] Read more.
This paper presents a novel pipeline leak diagnosis framework that combines Savitzky–Golay scalograms with a lightweight advanced deep learning architecture. Pipelines are critical for transporting fluids and gases, but leaks can lead to operational disruptions, environmental hazards, and financial losses. Leak events generate acoustic emissions (AE), captured as transient signals by AE sensors; however, these signals are often masked by noise and affected by the transported medium. To overcome this challenge, a fluid-independent detection approach is proposed that begins with acquiring AE data under various operational conditions, including multiple intensities of pinhole leaks and normal states. The transient signals are transformed into detailed scalograms using the Continuous Wavelet Transform (CWT), revealing subtle time–frequency patterns associated with leak events. To enhance these leak-specific features, a targeted Savitzky–Golay (SG) filter is applied, producing refined Savitzky–Golay scalograms (SG scalograms). These SG scalograms are then used to train a Convolutional Neural Network (CNN) and a newly developed lightweight Vision Transformer with streamlined self-attention (LViT-S), which autonomously learn both local and global features. The LViT-S achieves reduced embedding dimensions and fewer Transformer layers, significantly lowering computational cost while maintaining high performance. Extracted local and global features are merged into a unified feature vector, representing diverse visual characteristics learned by each network through their respective loss functions. This comprehensive feature representation is then passed to an Artificial Neural Network (ANN) for final classification, accurately identifying the presence, severity, and absence of leaks. The effectiveness of the proposed method is evaluated under two different pressure conditions, two fluid types (gas and water), and three distinct leak sizes, achieving a high classification accuracy of 98.6%. Additionally, a comparative evaluation against four state-of-the-art methods demonstrates that the proposed framework consistently delivers superior accuracy across diverse operational scenarios. Full article
(This article belongs to the Special Issue Advanced Sensing Technology in Structural Health Monitoring)
Show Figures

Figure 1

26 pages, 3233 KB  
Article
Analysis of Regional Surface CO2 Fluxes Using the MEGA Satellite Data Assimilation System
by Liting Hu, Xiaoyi Hu, Fei Jiang, Wei He, Zhu Deng, Shuangxi Fang and Xuekun Fang
Remote Sens. 2025, 17(22), 3720; https://doi.org/10.3390/rs17223720 - 14 Nov 2025
Viewed by 511
Abstract
Understanding the dynamics of terrestrial carbon sources and sinks is crucial for addressing climate change, yet significant uncertainties remain at regional scales. We developed the Monitoring and Evaluation of Greenhouse gAs Flux (MEGA) inversion system with satellite data assimilation and applied it to [...] Read more.
Understanding the dynamics of terrestrial carbon sources and sinks is crucial for addressing climate change, yet significant uncertainties remain at regional scales. We developed the Monitoring and Evaluation of Greenhouse gAs Flux (MEGA) inversion system with satellite data assimilation and applied it to China using OCO-2 V11.1r XCO2 retrievals. Our results show that China’s terrestrial ecosystems acted as a carbon sink of 0.28 ± 0.15 PgC yr−1 during 2018–2023, consistent with other inversion estimates. Validation against surface CO2 flask measurements demonstrated significant improvement, with RMSE and MAE reduced by 30%–46% and 24–44%, respectively. Six sets of prior sensitivity experiments conclusively demonstrated the robustness of MEGA. In addition, this study is the first to systematically compare model-derived and observation-based background fields in satellite data assimilation. Ten sets of background sensitivity experiments revealed that model-based background fields exhibit superior capability in resolving seasonal flux dynamics, though their performance remains contingent on three key factors: (1) initial fields, (2) flux fields, and (3) flux masks (used to control regional flux switches). These findings highlight the potential for further refinement of the atmospheric inversion system. Full article
(This article belongs to the Section Atmospheric Remote Sensing)
Show Figures

Figure 1

12 pages, 1662 KB  
Article
A High-Resolution Machine Vision System Using Computational Imaging Based on Multiple Image Capture During Object Transport
by Giseok Oh, Jeonghong Ha and Hyun Choi
Photonics 2025, 12(11), 1104; https://doi.org/10.3390/photonics12111104 - 9 Nov 2025
Viewed by 514
Abstract
This study adapts Fourier ptychography (FP) for high-resolution imaging in machine vision settings. We replace multi-angle illumination hardware with a single fixed light source and controlled object translation to enable a sequence of slightly shifted low-resolution frames to produce the requisite frequency-domain diversity [...] Read more.
This study adapts Fourier ptychography (FP) for high-resolution imaging in machine vision settings. We replace multi-angle illumination hardware with a single fixed light source and controlled object translation to enable a sequence of slightly shifted low-resolution frames to produce the requisite frequency-domain diversity for FP. The concept is validated in simulation using an embedded pupil function recovery algorithm to reconstruct a high-resolution complex field, recovering both amplitude and phase. For conveyor-belt transport, we introduce a lightweight preprocessing pipeline—background estimation, difference-based foreground detection, and morphological refinement—that yields robust masks and cropped inputs suitable for FP updates. The reconstructed images exhibit sharper fine structures and enhanced contrast relative to native lens imagery, indicating effective pupil synthesis without multi-LED arrays. The approach preserves compatibility with standard industrial optics and conveyor-style acquisition while reducing hardware complexity. We also discuss practical operating considerations, including blur-free capture and synchronization strategies. Full article
Show Figures

Figure 1

Back to TopTop