Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (540)

Search Parameters:
Keywords = masked loss

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
13 pages, 7106 KiB  
Article
Multi-Scale Universal Style-Transfer Network Based on Diffusion Model
by Na Su, Jingtao Wang and Yun Pan
Algorithms 2025, 18(8), 481; https://doi.org/10.3390/a18080481 - 4 Aug 2025
Viewed by 142
Abstract
Artistic style transfer aims to transfer the style of an artwork to a photograph while maintaining its original overall content. Although current style-transfer methods have achieved promising results when processing photorealistic images, they often struggle with brushstroke preservation in artworks, especially in styles [...] Read more.
Artistic style transfer aims to transfer the style of an artwork to a photograph while maintaining its original overall content. Although current style-transfer methods have achieved promising results when processing photorealistic images, they often struggle with brushstroke preservation in artworks, especially in styles such as oil painting and pointillism. In such cases, the extracted style and content features tend to include redundant information, leading to issues such as blurred edges and a loss of fine details in the transferred images. To address this problem, this paper proposes a multi-scale general style-transfer network based on diffusion models. The proposed network consists of a coarse style-transfer module and a refined style-transfer module. First, the coarse style-transfer module is designed to perform mainstream style-transfer tasks more efficiently by operating on downsampled images, enabling faster processing with satisfactory results. Next, to further enhance edge fidelity, a refined style-transfer module is introduced. This module utilizes a segmentation component to generate a mask of the main subject in the image and performs edge-aware refinement. This enhances the fusion between the subject’s edges and the target style while preserving more detailed features. To improve overall image quality and better integrate the style along the content boundaries, the output from the coarse module is upsampled by a factor of two and combined with the subject mask. With the assistance of ControlNet and Stable Diffusion, the model performs content-aware edge redrawing to enhance the overall visual quality of the stylized image. Compared with state-of-the-art style-transfer methods, the proposed model preserves more edge details and achieves more natural fusion between style and content. Full article
(This article belongs to the Section Evolutionary Algorithms and Machine Learning)
Show Figures

Figure 1

10 pages, 1129 KiB  
Article
Optimal Sound Presentation Level for Sound Localization Testing in Unilateral Conductive Hearing Loss
by Miki Takahara, Takanori Nishiyama, Yu Fumiiri, Tsubasa Kitama, Makoto Hosoya, Marie N. Shimanuki, Masafumi Ueno, Takeshi Wakabayashi, Hiroyuki Ozawa and Naoki Oishi
Audiol. Res. 2025, 15(4), 95; https://doi.org/10.3390/audiolres15040095 (registering DOI) - 2 Aug 2025
Viewed by 91
Abstract
Background/Objectives: This study aimed to investigate the optimal sound presentation level for sound localization testing to assess the effect of hearing interventions in individuals with unilateral conductive hearing loss (UCHL). Methods: Nine participants with normal hearing were tested, and simulated two-stage [...] Read more.
Background/Objectives: This study aimed to investigate the optimal sound presentation level for sound localization testing to assess the effect of hearing interventions in individuals with unilateral conductive hearing loss (UCHL). Methods: Nine participants with normal hearing were tested, and simulated two-stage UCHL was created using earmuffs and earplugs. We created two types of masking conditions: (1) only an earplug inserted, and (2) an earplug inserted with an earmuff worn. A sound localization test was performed for each condition. The sound presentation levels were 40, 45, 50, 55, 60, 65, and 70 dB SPL, and the results were evaluated using root mean square and d-values. Results: Both values showed little difference in masking Condition 2, regardless of the sound presentation level, whereas in masking Condition 1, the values were at their minimum at 55 dB SPL. In addition, comparing the differences between masking Conditions 1 and 2 for each sound presentation level, the greatest difference was observed at 55 dB SPL for both values. Conclusions: The optimal sound presentation level for sound localization testing to assess hearing intervention effects in UCHL was 55 dB. This result may be attributed to the effect of input from the non-masked ear, accounting for interaural attenuation; the effect was considered minimal at 55 dB SPL. Full article
(This article belongs to the Section Hearing)
Show Figures

Figure 1

18 pages, 5956 KiB  
Article
Improving the Universal Performance of Land Cover Semantic Segmentation Through Training Data Refinement and Multi-Dataset Fusion via Redundant Models
by Jae Young Chang, Kwan-Young Oh and Kwang-Jae Lee
Remote Sens. 2025, 17(15), 2669; https://doi.org/10.3390/rs17152669 - 1 Aug 2025
Viewed by 133
Abstract
Artificial intelligence (AI) has become the mainstream of analysis tools in remote sensing. Various semantic segmentation models have been introduced to segment land cover from aerial or satellite images, and remarkable results have been achieved. However, they often lack universal performance on unseen [...] Read more.
Artificial intelligence (AI) has become the mainstream of analysis tools in remote sensing. Various semantic segmentation models have been introduced to segment land cover from aerial or satellite images, and remarkable results have been achieved. However, they often lack universal performance on unseen images, making them challenging to provide as a service. One of the primary reasons for the lack of robustness is overfitting, resulting from errors and inconsistencies in the ground truth (GT). In this study, we propose a method to mitigate these inconsistencies by utilizing redundant models and verify the improvement using a public dataset based on Google Earth images. Redundant models share the same network architecture and hyperparameters but are trained with different combinations of training and validation data on the same dataset. Because of the variations in sample exposure during training, these models yield slightly different inference results. This variability allows for the estimation of pixel-level confidence levels for the GT. The confidence level is incorporated into the GT to influence the loss calculation during the training of the enhanced model. Furthermore, we implemented a consensus model that employs modified masks, where classes with low confidence are substituted by the dominant classes identified through a majority vote from the redundant models. To further improve robustness, we extended the same approach to fuse the dataset with different class compositions based on imagery from the Korea Multipurpose Satellite 3A (KOMPSAT-3A). Performance evaluations were conducted on three network architectures: a simple network, U-Net, and DeepLabV3. In the single-dataset case, the performance of the enhanced and consensus models improved by an average of 2.49% and 2.59% across the network architectures. In the multi-dataset scenario, the enhanced models and consensus models showed an average performance improvement of 3.37% and 3.02% across the network architectures, respectively, compared to an average increase of 1.55% without the proposed method. Full article
Show Figures

Graphical abstract

21 pages, 97817 KiB  
Article
Compression of 3D Optical Encryption Using Singular Value Decomposition
by Kyungtae Park, Min-Chul Lee and Myungjin Cho
Sensors 2025, 25(15), 4742; https://doi.org/10.3390/s25154742 - 1 Aug 2025
Viewed by 231
Abstract
In this paper, we propose a compressionmethod for optical encryption using singular value decomposition (SVD). Double random phase encryption (DRPE), which employs two distinct random phase masks, is adopted as the optical encryption technique. Since the encrypted data in DRPE have the same [...] Read more.
In this paper, we propose a compressionmethod for optical encryption using singular value decomposition (SVD). Double random phase encryption (DRPE), which employs two distinct random phase masks, is adopted as the optical encryption technique. Since the encrypted data in DRPE have the same size as the input data and consists of complex values, a compression technique is required to improve data efficiency. To address this issue, we introduce SVD as a compression method. SVD decomposes any matrix into simpler components, such as a unitary matrix, a rectangular diagonal matrix, and a complex unitary matrix. By leveraging this property, the encrypted data generated by DRPE can be effectively compressed. However, this compression may lead to some loss of information in the decrypted data. To mitigate this loss, we employ volumetric computational reconstruction based on integral imaging. As a result, the proposed method enhances the visual quality, compression ratio, and security of DRPE simultaneously. To validate the effectiveness of the proposed method, we conduct both computer simulations and optical experiments. The performance is evaluated quantitatively using peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and peak sidelobe ratio (PSR) as evaluation metrics. Full article
Show Figures

Figure 1

19 pages, 3328 KiB  
Article
Enhancing Trauma Care: Machine Learning-Based Photoplethysmography Analysis for Estimating Blood Volume During Hemorrhage and Resuscitation
by Jose M. Gonzalez, Lawrence Holland, Sofia I. Hernandez Torres, John G. Arrington, Tina M. Rodgers and Eric J. Snider
Bioengineering 2025, 12(8), 833; https://doi.org/10.3390/bioengineering12080833 - 31 Jul 2025
Viewed by 189
Abstract
Hemorrhage is the leading cause of preventable death in trauma care, requiring rapid and accurate detection to guide effective interventions. Hemorrhagic shock can be masked by underlying compensatory mechanisms, which may lead to delayed decision-making that can compromise casualty care. In this proof-of-concept [...] Read more.
Hemorrhage is the leading cause of preventable death in trauma care, requiring rapid and accurate detection to guide effective interventions. Hemorrhagic shock can be masked by underlying compensatory mechanisms, which may lead to delayed decision-making that can compromise casualty care. In this proof-of-concept study, we aimed to develop and evaluate machine learning models to predict Percent Estimated Blood Loss from a photoplethysmography waveform, offering non-invasive, field deployable solutions. Different model types were tuned and optimized using data captured during a hemorrhage and resuscitation swine study. Through this optimization process, we evaluated different time-lengths of prediction windows, machine learning model architectures, and data normalization approaches. Models were successful at predicting Percent Estimated Blood Loss in blind swine subjects with coefficient of determination values exceeding 0.8. This provides evidence that Percent Estimated Blood Loss can be accurately derived from non-invasive signals, improving its utility for trauma care and casualty triage in the pre-hospital and emergency medicine environment. Full article
Show Figures

Figure 1

19 pages, 3130 KiB  
Article
Deep Learning-Based Instance Segmentation of Galloping High-Speed Railway Overhead Contact System Conductors in Video Images
by Xiaotong Yao, Huayu Yuan, Shanpeng Zhao, Wei Tian, Dongzhao Han, Xiaoping Li, Feng Wang and Sihua Wang
Sensors 2025, 25(15), 4714; https://doi.org/10.3390/s25154714 - 30 Jul 2025
Viewed by 234
Abstract
The conductors of high-speed railway OCSs (Overhead Contact Systems) are susceptible to conductor galloping due to the impact of natural elements such as strong winds, rain, and snow, resulting in conductor fatigue damage and significantly compromising train operational safety. Consequently, monitoring the galloping [...] Read more.
The conductors of high-speed railway OCSs (Overhead Contact Systems) are susceptible to conductor galloping due to the impact of natural elements such as strong winds, rain, and snow, resulting in conductor fatigue damage and significantly compromising train operational safety. Consequently, monitoring the galloping status of conductors is crucial, and instance segmentation techniques, by delineating the pixel-level contours of each conductor, can significantly aid in the identification and study of galloping phenomena. This work expands upon the YOLO11-seg model and introduces an instance segmentation approach for galloping video and image sensor data of OCS conductors. The algorithm, designed for the stripe-like distribution of OCS conductors in the data, employs four-direction Sobel filters to extract edge features in horizontal, vertical, and diagonal orientations. These features are subsequently integrated with the original convolutional branch to form the FDSE (Four Direction Sobel Enhancement) module. It integrates the ECA (Efficient Channel Attention) mechanism for the adaptive augmentation of conductor characteristics and utilizes the FL (Focal Loss) function to mitigate the class-imbalance issue between positive and negative samples, hence enhancing the model’s sensitivity to conductors. Consequently, segmentation outcomes from neighboring frames are utilized, and mask-difference analysis is performed to autonomously detect conductor galloping locations, emphasizing their contours for the clear depiction of galloping characteristics. Experimental results demonstrate that the enhanced YOLO11-seg model achieves 85.38% precision, 77.30% recall, 84.25% AP@0.5, 81.14% F1-score, and a real-time processing speed of 44.78 FPS. When combined with the galloping visualization module, it can issue real-time alerts of conductor galloping anomalies, providing robust technical support for railway OCS safety monitoring. Full article
(This article belongs to the Section Industrial Sensors)
Show Figures

Figure 1

26 pages, 6348 KiB  
Article
Building Envelope Thermal Anomaly Detection Using an Integrated Vision-Based Technique and Semantic Segmentation
by Shayan Mirzabeigi, Ryan Razkenari and Paul Crovella
Buildings 2025, 15(15), 2672; https://doi.org/10.3390/buildings15152672 - 29 Jul 2025
Viewed by 329
Abstract
Infrared thermography is a common approach used in building inspection for identifying building envelope thermal anomalies that cause energy loss and occupant thermal discomfort. Detecting these anomalies is essential to improve the thermal performance of energy-inefficient buildings through energy retrofit design and correspondingly [...] Read more.
Infrared thermography is a common approach used in building inspection for identifying building envelope thermal anomalies that cause energy loss and occupant thermal discomfort. Detecting these anomalies is essential to improve the thermal performance of energy-inefficient buildings through energy retrofit design and correspondingly reduce operational energy costs and environmental impacts. A thermal bridge is an unwanted conductive heat transfer. On the other hand, an infiltration/exfiltration anomaly is an uncontrollable convective heat transfer, typically happening around windows and doors, but it can also be due to a defect that comprises a building envelope’s integrity. While the existing literature underscores the significance of automatic thermal anomaly identification and offers insights into automated methodologies, there is a notable gap in addressing an automated workflow that leverages building envelope component segmentation for enhanced detection accuracy. Consequently, an automatic thermal anomaly identification workflow from visible and thermal images was developed to test it, utilizing segmented building envelope information compared to a workflow without any semantic segmentation. Therefore, building envelope images (e.g., walls and windows) were segmented based on a U-Net architecture compared to a more conventional semantic segmentation approach. The results were discussed to better understand the importance of the availability of training data and for scaling the workflow. Then, thermal anomaly thresholds for different target domains were detected using probability distributions. Finally, thermal anomaly masks of those domains were computed. This study conducted a comprehensive examination of a campus building in Syracuse, New York, utilizing a drone-based data collection approach. The case study successfully detected diverse thermal anomalies associated with various envelope components. The proposed approach offers the potential for immediate and accurate in situ thermal anomaly detection in building inspections. Full article
Show Figures

Figure 1

22 pages, 825 KiB  
Article
Conformal Segmentation in Industrial Surface Defect Detection with Statistical Guarantees
by Cheng Shen and Yuewei Liu
Mathematics 2025, 13(15), 2430; https://doi.org/10.3390/math13152430 - 28 Jul 2025
Viewed by 265
Abstract
Detection of surface defects can significantly elongate mechanical service time and mitigate potential risks during safety management. Traditional defect detection methods predominantly rely on manual inspection, which suffers from low efficiency and high costs. Some machine learning algorithms and artificial intelligence models for [...] Read more.
Detection of surface defects can significantly elongate mechanical service time and mitigate potential risks during safety management. Traditional defect detection methods predominantly rely on manual inspection, which suffers from low efficiency and high costs. Some machine learning algorithms and artificial intelligence models for defect detection, such as Convolutional Neural Networks (CNNs), present outstanding performance, but they are often data-dependent and cannot provide guarantees for new test samples. To this end, we construct a detection model by combining Mask R-CNN, selected for its strong baseline performance in pixel-level segmentation, with Conformal Risk Control. The former evaluates the distribution that discriminates defects from all samples based on probability. The detection model is improved by retraining with calibration data that is assumed to be independent and identically distributed (i.i.d) with the test data. The latter constructs a prediction set on which a given guarantee for detection will be obtained. First, we define a loss function for each calibration sample to quantify detection error rates. Subsequently, we derive a statistically rigorous threshold by optimization of error rates and a given guarantee significance as the risk level. With the threshold, defective pixels with high probability in test images are extracted to construct prediction sets. This methodology ensures that the expected error rate on the test set remains strictly bounded by the predefined risk level. Furthermore, our model shows robust and efficient control over the expected test set error rate when calibration-to-test partitioning ratios vary. Full article
Show Figures

Figure 1

30 pages, 92065 KiB  
Article
A Picking Point Localization Method for Table Grapes Based on PGSS-YOLOv11s and Morphological Strategies
by Jin Lu, Zhongji Cao, Jin Wang, Zhao Wang, Jia Zhao and Minjie Zhang
Agriculture 2025, 15(15), 1622; https://doi.org/10.3390/agriculture15151622 - 26 Jul 2025
Viewed by 294
Abstract
During the automated picking of table grapes, the automatic recognition and segmentation of grape pedicels, along with the positioning of picking points, are vital components for all the following operations of the harvesting robot. In the actual scene of a grape plantation, however, [...] Read more.
During the automated picking of table grapes, the automatic recognition and segmentation of grape pedicels, along with the positioning of picking points, are vital components for all the following operations of the harvesting robot. In the actual scene of a grape plantation, however, it is extremely difficult to accurately and efficiently identify and segment grape pedicels and then reliably locate the picking points. This is attributable to the low distinguishability between grape pedicels and the surrounding environment such as branches, as well as the impacts of other conditions like weather, lighting, and occlusion, which are coupled with the requirements for model deployment on edge devices with limited computing resources. To address these issues, this study proposes a novel picking point localization method for table grapes based on an instance segmentation network called Progressive Global-Local Structure-Sensitive Segmentation (PGSS-YOLOv11s) and a simple combination strategy of morphological operators. More specifically, the network PGSS-YOLOv11s is composed of an original backbone of the YOLOv11s-seg, a spatial feature aggregation module (SFAM), an adaptive feature fusion module (AFFM), and a detail-enhanced convolutional shared detection head (DE-SCSH). And the PGSS-YOLOv11s have been trained with a new grape segmentation dataset called Grape-⊥, which includes 4455 grape pixel-level instances with the annotation of ⊥-shaped regions. After the PGSS-YOLOv11s segments the ⊥-shaped regions of grapes, some morphological operations such as erosion, dilation, and skeletonization are combined to effectively extract grape pedicels and locate picking points. Finally, several experiments have been conducted to confirm the validity, effectiveness, and superiority of the proposed method. Compared with the other state-of-the-art models, the main metrics F1 score and mask mAP@0.5 of the PGSS-YOLOv11s reached 94.6% and 95.2% on the Grape-⊥ dataset, as well as 85.4% and 90.0% on the Winegrape dataset. Multi-scenario tests indicated that the success rate of positioning the picking points reached up to 89.44%. In orchards, real-time tests on the edge device demonstrated the practical performance of our method. Nevertheless, for grapes with short pedicels or occluded pedicels, the designed morphological algorithm exhibited the loss of picking point calculations. In future work, we will enrich the grape dataset by collecting images under different lighting conditions, from various shooting angles, and including more grape varieties to improve the method’s generalization performance. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

25 pages, 27219 KiB  
Article
KCUNET: Multi-Focus Image Fusion via the Parallel Integration of KAN and Convolutional Layers
by Jing Fang, Ruxian Wang, Xinglin Ning, Ruiqing Wang, Shuyun Teng, Xuran Liu, Zhipeng Zhang, Wenfeng Lu, Shaohai Hu and Jingjing Wang
Entropy 2025, 27(8), 785; https://doi.org/10.3390/e27080785 - 24 Jul 2025
Viewed by 179
Abstract
Multi-focus image fusion (MFIF) is an image-processing method that aims to generate fully focused images by integrating source images from different focal planes. However, the defocus spread effect (DSE) often leads to blurred or jagged focus/defocus boundaries in fused images, which affects the [...] Read more.
Multi-focus image fusion (MFIF) is an image-processing method that aims to generate fully focused images by integrating source images from different focal planes. However, the defocus spread effect (DSE) often leads to blurred or jagged focus/defocus boundaries in fused images, which affects the quality of the image. To address this issue, this paper proposes a novel model that embeds the Kolmogorov–Arnold network with convolutional layers in parallel within the U-Net architecture (KCUNet). This model keeps the spatial dimensions of the feature map constant to maintain high-resolution details while progressively increasing the number of channels to capture multi-level features at the encoding stage. In addition, KCUNet incorporates a content-guided attention mechanism to enhance edge information processing, which is crucial for DSE reduction and edge preservation. The model’s performance is optimized through a hybrid loss function that evaluates in several aspects, including edge alignment, mask prediction, and image quality. Finally, comparative evaluations against 15 state-of-the-art methods demonstrate KCUNet’s superior performance in both qualitative and quantitative analyses. Full article
(This article belongs to the Section Signal and Data Analysis)
Show Figures

Figure 1

20 pages, 4920 KiB  
Article
Martian Skylight Identification Based on the Deep Learning Model
by Lihong Li, Lingli Mu, Wei Zhang, Weihua Dong and Yuqing He
Remote Sens. 2025, 17(15), 2571; https://doi.org/10.3390/rs17152571 - 24 Jul 2025
Viewed by 294
Abstract
As a type of distinctive pit on Mars, skylights are entrances to subsurface lava caves. They are very important for studying volcanic activity and potential preserved water ice, and are also considered as potential sites for human extraterrestrial bases in the future. Most [...] Read more.
As a type of distinctive pit on Mars, skylights are entrances to subsurface lava caves. They are very important for studying volcanic activity and potential preserved water ice, and are also considered as potential sites for human extraterrestrial bases in the future. Most skylights are manually identified, which has low efficiency and is highly subjective. Although deep learning methods have recently been used to identify skylights, they face challenges of few effective samples and low identification accuracy. In this article, 151 positive samples and 920 negative samples based on the MRO-HiRISE image data was used to create an initial skylight dataset, which contained few positive samples. To augment the initial dataset, StyleGAN2-ADA was selected to synthesize some positive samples and generated an augmented dataset with 896 samples. On the basis of the augmented skylight dataset, we proposed YOLOv9-Skylight for skylight identification by incorporating Inner-EIoU loss and DySample to enhance localization accuracy and feature extracting ability. Compared with YOLOv9, the P, R, and the F1 of YOLOv9-Skylight were improved by about 9.1%, 2.8%, and 5.6%, respectively. Compared with other mainstream models such as YOLOv5, YOLOv10, Faster R-CNN, Mask R-CNN, and DETR, YOLOv9-Skylight achieved the highest accuracy (F1 = 92.5%), which shows a strong performance in skylight identification. Full article
(This article belongs to the Special Issue Remote Sensing and Photogrammetry Applied to Deep Space Exploration)
Show Figures

Figure 1

38 pages, 6851 KiB  
Article
FGFNet: Fourier Gated Feature-Fusion Network with Fractal Dimension Estimation for Robust Palm-Vein Spoof Detection
by Seung Gu Kim, Jung Soo Kim and Kang Ryoung Park
Fractal Fract. 2025, 9(8), 478; https://doi.org/10.3390/fractalfract9080478 - 22 Jul 2025
Viewed by 264
Abstract
The palm-vein recognition system has garnered attention as a biometric technology due to its resilience to external environmental factors, protection of personal privacy, and low risk of external exposure. However, with recent advancements in deep learning-based generative models for image synthesis, the quality [...] Read more.
The palm-vein recognition system has garnered attention as a biometric technology due to its resilience to external environmental factors, protection of personal privacy, and low risk of external exposure. However, with recent advancements in deep learning-based generative models for image synthesis, the quality and sophistication of fake images have improved, leading to an increased security threat from counterfeit images. In particular, palm-vein images acquired through near-infrared illumination exhibit low resolution and blurred characteristics, making it even more challenging to detect fake images. Furthermore, spoof detection specifically targeting palm-vein images has not been studied in detail. To address these challenges, this study proposes the Fourier-gated feature-fusion network (FGFNet) as a novel spoof detector for palm-vein recognition systems. The proposed network integrates masked fast Fourier transform, a map-based gated feature fusion block, and a fast Fourier convolution (FFC) attention block with global contrastive loss to effectively detect distortion patterns caused by generative models. These components enable the efficient extraction of critical information required to determine the authenticity of palm-vein images. In addition, fractal dimension estimation (FDE) was employed for two purposes in this study. In the spoof attack procedure, FDE was used to evaluate how closely the generated fake images approximate the structural complexity of real palm-vein images, confirming that the generative model produced highly realistic spoof samples. In the spoof detection procedure, the FDE results further demonstrated that the proposed FGFNet effectively distinguishes between real and fake images, validating its capability to capture subtle structural differences induced by generative manipulation. To evaluate the spoof detection performance of FGFNet, experiments were conducted using real palm-vein images from two publicly available palm-vein datasets—VERA Spoofing PalmVein (VERA dataset) and PLUSVein-contactless (PLUS dataset)—as well as fake palm-vein images generated based on these datasets using a cycle-consistent generative adversarial network. The results showed that, based on the average classification error rate, FGFNet achieved 0.3% and 0.3% on the VERA and PLUS datasets, respectively, demonstrating superior performance compared to existing state-of-the-art spoof detection methods. Full article
Show Figures

Figure 1

15 pages, 4146 KiB  
Article
Monitoring Forest Cover Trends in Nepal: Insights from 2000–2020
by Aditya Eaturu
Sustainability 2025, 17(14), 6511; https://doi.org/10.3390/su17146511 - 16 Jul 2025
Viewed by 544
Abstract
This study investigates the spatial relationship between population distribution and tree cover loss in Nepal from 2000 to 2020, using satellite-based forest cover and population data along with statistical and geospatial analysis. Two statistical methods—linear regression (LR) and Geographically Weighted Regression (GWR)—were used [...] Read more.
This study investigates the spatial relationship between population distribution and tree cover loss in Nepal from 2000 to 2020, using satellite-based forest cover and population data along with statistical and geospatial analysis. Two statistical methods—linear regression (LR) and Geographically Weighted Regression (GWR)—were used to assess the influence of population on forest cover change. The correlation between total population and forest loss at the national level suggested little to no direct impact of population growth on forest loss. However, sub-national analysis revealed localized forest degradation, highlighting the importance of spatial and regional assessments to uncover land cover changes masked by national trends. While LR showed a weak national-level correlation, GWR revealed substantial spatial variation, with the coefficient of determination values increasing from 0.21 in 2000 to 0.59 in 2020. In some regions, local R2 exceeded 0.75 during 2015 and 2020, highlighting emerging hotspot clusters where population pressure is strongly linked to deforestation, especially along major infrastructure corridors. Using very high-resolution spatial data enabled pixel-level analysis, capturing fine-scale deforestation patterns, and confirming hotspot accuracy. Overall, the findings emphasize the value of spatially explicit models like GWR for understanding human–environment interactions guiding targeted land use planning to balance development with environmental sustainability in Nepal. Full article
Show Figures

Figure 1

24 pages, 5976 KiB  
Article
Spatial Downscaling of Sea Level Anomaly Using a Deep Separable Distillation Network
by Senmin Shi, Yineng Li, Yuhang Zhu, Tao Song and Shiqiu Peng
Remote Sens. 2025, 17(14), 2428; https://doi.org/10.3390/rs17142428 - 13 Jul 2025
Viewed by 428
Abstract
The use of high-resolution sea level anomaly (SLA) data in climate change research and ocean forecasting has become increasingly important. However, existing datasets often lack the fine spatial resolution required for capturing mesoscale ocean processes accurately. This has led to the development of [...] Read more.
The use of high-resolution sea level anomaly (SLA) data in climate change research and ocean forecasting has become increasingly important. However, existing datasets often lack the fine spatial resolution required for capturing mesoscale ocean processes accurately. This has led to the development of conventional deep learning models for SLA spatial downscaling, but these models often overlook spatial disparities between land and ocean regions and do not adequately address the spatial structures of SLA data. As a result, their accuracy and structural consistency are suboptimal. To address these issues, we propose a Deep Separable Distillation Network (DSDN) that integrates Depthwise Separable Distillation Blocks (DSDB) and a Landmask Contextual Attention Mechanism (M_CAMB) to achieve efficient and accurate spatial downscaling. The M_CAMB employs geographically-informed land masks to enhance the attention mechanism, prioritizing ocean regions. Additionally, we introduce a novel Pixel-Structure Loss (PSLoss) to enforce spatial structure constraints, significantly improving the structural fidelity of the SLA downscaling results. Experimental results demonstrate that DSDN achieves a root mean square error (RMSE) of 0.062 cm, a peak signal-to-noise ratio (PSNR) of 42.22 dB, and a structural similarity index (SSIM) of 0.976 in SLA downscaling. These results surpass those of baseline models and highlight the superior precision and structural consistency of DSDN. Full article
Show Figures

Figure 1

14 pages, 6691 KiB  
Article
Remote Sensing Extraction of Damaged Buildings in the Shigatse Earthquake, 2025: A Hybrid YOLO-E and SAM2 Approach
by Zhimin Wu, Chenyao Qu, Wei Wang, Zelang Miao and Huihui Feng
Sensors 2025, 25(14), 4375; https://doi.org/10.3390/s25144375 - 12 Jul 2025
Viewed by 374
Abstract
In January 2025, a magnitude 6.8 earthquake struck Dingri County, Shigatse, Tibet, causing severe damage. Rapid and precise extraction of damaged buildings is essential for emergency relief and rebuilding efforts. This study proposes an approach integrating YOLO-E (Real-Time Seeing Anything) and the Segment [...] Read more.
In January 2025, a magnitude 6.8 earthquake struck Dingri County, Shigatse, Tibet, causing severe damage. Rapid and precise extraction of damaged buildings is essential for emergency relief and rebuilding efforts. This study proposes an approach integrating YOLO-E (Real-Time Seeing Anything) and the Segment Anything Model 2 (SAM2) to extract damaged buildings with multi-source remote sensing images, including post-earthquake Gaofen-7 imagery (0.80 m), Beijing-3 imagery (0.30 m), and pre-earthquake Google satellite imagery (0.15 m), over the affected region. In this hybrid approach, YOLO-E functions as the preliminary segmentation module for initial segmentation. It leverages its real-time detection and segmentation capability to locate potential damaged building regions and generate coarse segmentation masks rapidly. Subsequently, SAM2 follows as a refinement step, incorporating shapefile information from pre-disaster sources to apply precise, pixel-level segmentation. The dataset used for training contained labeled examples of damaged buildings, and the model optimization was carried out using stochastic gradient descent (SGD), with cross-entropy and mean squared error as the selected loss functions. Upon evaluation, the model reached a precision of 0.840, a recall of 0.855, an F1-score of 0.847, and an IoU of 0.735. It successfully extracted 492 suspected damaged building patches within a radius of 20 km from the earthquake epicenter, clearly showing the distribution characteristics of damaged buildings concentrated in the earthquake fault zone. In summary, this hybrid YOLO-E and SAM2 approach, leveraging multi-source remote sensing imagery, delivers precise and rapid extraction of damaged buildings with a precision of 0.840, recall of 0.855, and IoU of 0.735, effectively supporting targeted earthquake rescue and post-disaster reconstruction efforts in the Dingri County fault zone. Full article
Show Figures

Figure 1

Back to TopTop