Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (652)

Search Parameters:
Keywords = super-pixel

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 1476 KiB  
Article
Ambiguities, Built-In Biases, and Flaws in Big Data Insight Extraction
by Serge Galam
Information 2025, 16(8), 661; https://doi.org/10.3390/info16080661 (registering DOI) - 2 Aug 2025
Abstract
I address the challenge of extracting reliable insights from large datasets using a simplified model that illustrates how hierarchical classification can distort outcomes. The model consists of discrete pixels labeled red, blue, or white. Red and blue indicate distinct properties, while white represents [...] Read more.
I address the challenge of extracting reliable insights from large datasets using a simplified model that illustrates how hierarchical classification can distort outcomes. The model consists of discrete pixels labeled red, blue, or white. Red and blue indicate distinct properties, while white represents unclassified or ambiguous data. A macro-color is assigned only if one color holds a strict majority among the pixels. Otherwise, the aggregate is labeled white, reflecting uncertainty. This setup mimics a percolation threshold at fifty percent. Assuming that directly accessing the various proportions from the data of colors is infeasible, I implement a hierarchical coarse-graining procedure. Elements (first pixels, then aggregates) are recursively grouped and reclassified via local majority rules, ultimately producing a single super-aggregate for which the color represents the inferred macro-property of the collection of pixels as a whole. Analytical results supported by simulations show that the process introduces additional white aggregates beyond white pixels, which could be present initially; these arise from groups lacking a clear majority, requiring arbitrary symmetry-breaking decisions to attribute a color to them. While each local resolution may appear minor and inconsequential, their repetitions introduce a growing systematic bias. Even with complete data, unavoidable asymmetries in local rules are shown to skew outcomes. This study highlights a critical limitation of recursive data reduction. Insight extraction is shaped not only by data quality but also by how local ambiguity is handled, resulting in built-in biases. Thus, the related flaws are not due to the data but to structural choices made during local aggregations. Although based on a simple model, these findings expose a high likelihood of inherent flaws in widely used hierarchical classification techniques. Full article
(This article belongs to the Section Artificial Intelligence)
24 pages, 17531 KiB  
Article
Efficient Unsupervised Clustering of Hyperspectral Images via Flexible Multi-Anchor Graphs
by Yihong Li, Ting Wang, Zhe Cao, Haonan Xin and Rong Wang
Remote Sens. 2025, 17(15), 2647; https://doi.org/10.3390/rs17152647 - 30 Jul 2025
Viewed by 99
Abstract
Unsupervised hyperspectral image (HSI) clustering is a fundamental yet challenging task due to high dimensionality and complex spectral–spatial characteristics. In this paper, we propose a novel and efficient clustering framework centered on adaptive and diverse anchor graph modeling. First, we introduce a parameter-free [...] Read more.
Unsupervised hyperspectral image (HSI) clustering is a fundamental yet challenging task due to high dimensionality and complex spectral–spatial characteristics. In this paper, we propose a novel and efficient clustering framework centered on adaptive and diverse anchor graph modeling. First, we introduce a parameter-free construction strategy that employs Entropy Rate Superpixel (ERS) segmentation to generate multiple anchor graphs of varying sizes from a single HSI, overcoming the limitation of fixed anchor quantities and enhancing structural expressiveness. Second, we propose an anchor-to-pixel label propagation mechanism to transfer anchor-level cluster labels back to the pixel level, reinforcing spatial coherence and spectral discriminability. Third, we perform clustering directly at the anchor level, which substantially reduces computational cost while retaining structure-aware accuracy. Extensive experiments on three benchmark datasets (Trento, Salinas, and Pavia Center) demonstrate the effectiveness and efficiency of our approach. Full article
Show Figures

Figure 1

27 pages, 8755 KiB  
Article
Mapping Wetlands with High-Resolution Planet SuperDove Satellite Imagery: An Assessment of Machine Learning Models Across the Diverse Waterscapes of New Zealand
by Md. Saiful Islam Khan, Maria C. Vega-Corredor and Matthew D. Wilson
Remote Sens. 2025, 17(15), 2626; https://doi.org/10.3390/rs17152626 - 29 Jul 2025
Viewed by 288
Abstract
(1) Background: Wetlands are ecologically significant ecosystems that support biodiversity and contribute to essential environmental functions such as water purification, carbon storage and flood regulation. However, these ecosystems face increasing pressures from land-use change and degradation, prompting the need for scalable and accurate [...] Read more.
(1) Background: Wetlands are ecologically significant ecosystems that support biodiversity and contribute to essential environmental functions such as water purification, carbon storage and flood regulation. However, these ecosystems face increasing pressures from land-use change and degradation, prompting the need for scalable and accurate classification methods to support conservation and policy efforts. In this research, our motivation was to test whether high-spatial-resolution PlanetScope imagery can be used with pixel-based machine learning to support the mapping and monitoring of wetlands at a national scale. (2) Methods: This study compared four machine learning classification models—Random Forest (RF), XGBoost (XGB), Histogram-Based Gradient Boosting (HGB) and a Multi-Layer Perceptron Classifier (MLPC)—to detect and map wetland areas across New Zealand. All models were trained using eight-band SuperDove satellite imagery from PlanetScope, with a spatial resolution of ~3 m, and ancillary geospatial datasets representing topography and soil drainage characteristics, each of which is available globally. (3) Results: All four machine learning models performed well in detecting wetlands from SuperDove imagery and environmental covariates, with varying strengths. The highest accuracy was achieved using all eight image bands alongside features created from supporting geospatial data. For binary wetland classification, the highest F1 scores were recorded by XGB (0.73) and RF/HGB (both 0.72) when including all covariates. MLPC also showed competitive performance (wetland F1 score of 0.71), despite its relatively lower spatial consistency. However, each model over-predicts total wetland area at a national level, an issue which was able to be reduced by increasing the classification probability threshold and spatial filtering. (4) Conclusions: The comparative analysis highlights the strengths and trade-offs of RF, XGB, HGB and MLPC models for wetland classification. While all four methods are viable, RF offers some key advantages, including ease of deployment and transferability, positioning it as a promising candidate for scalable, high-resolution wetland monitoring across diverse ecological settings. Further work is required for verification of small-scale wetlands (<~0.5 ha) and the addition of fine-spatial-scale covariates. Full article
Show Figures

Figure 1

27 pages, 8957 KiB  
Article
DFAN: Single Image Super-Resolution Using Stationary Wavelet-Based Dual Frequency Adaptation Network
by Gyu-Il Kim and Jaesung Lee
Symmetry 2025, 17(8), 1175; https://doi.org/10.3390/sym17081175 - 23 Jul 2025
Viewed by 277
Abstract
Single image super-resolution is the inverse problem of reconstructing a high-resolution image from its low-resolution counterpart. Although recent Transformer-based architectures leverage global context integration to improve reconstruction quality, they often overlook frequency-specific characteristics, resulting in the loss of high-frequency information. To address this [...] Read more.
Single image super-resolution is the inverse problem of reconstructing a high-resolution image from its low-resolution counterpart. Although recent Transformer-based architectures leverage global context integration to improve reconstruction quality, they often overlook frequency-specific characteristics, resulting in the loss of high-frequency information. To address this limitation, we propose the Dual Frequency Adaptive Network (DFAN). DFAN first decomposes the input into low- and high-frequency components via Stationary Wavelet Transform. In the low-frequency branch, Swin Transformer layers restore global structures and color consistency. In contrast, the high-frequency branch features a dedicated module that combines Directional Convolution with Residual Dense Blocks, precisely reinforcing edges and textures. A frequency fusion module then adaptively merges these complementary features using depthwise and pointwise convolutions, achieving a balanced reconstruction. During training, we introduce a frequency-aware multi-term loss alongside the standard pixel-wise loss to explicitly encourage high-frequency preservation. Extensive experiments on the Set5, Set14, BSD100, Urban100, and Manga109 benchmarks show that DFAN achieves up to +0.64 dBpeak signal-to-noise ratio, +0.01 structural similarity index measure, and −0.01learned perceptual image patch similarity over the strongest frequency-domain baselines, while also delivering visibly sharper textures and cleaner edges. By unifying spatial and frequency-domain advantages, DFAN effectively mitigates high-frequency degradation and enhances SISR performance. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

24 pages, 7933 KiB  
Article
Multi-Temporal Dual Polarimetric SAR Crop Classification Based on Spatial Information Comprehensive Utilization
by Qiang Yin, Yuming Du, Fangfang Li, Yongsheng Zhou and Fan Zhang
Remote Sens. 2025, 17(13), 2304; https://doi.org/10.3390/rs17132304 - 4 Jul 2025
Viewed by 184
Abstract
Dual polarimetric SAR is capable of reflecting the biophysical and geometrical information of terrain with open access data availability. When it is combined with time-series observations, it can effectively capture the dynamic evolution of scattering characteristics of crops in different growth cycles. However, [...] Read more.
Dual polarimetric SAR is capable of reflecting the biophysical and geometrical information of terrain with open access data availability. When it is combined with time-series observations, it can effectively capture the dynamic evolution of scattering characteristics of crops in different growth cycles. However, the actual planting of crops often shows spatial dispersion, and the same crop may be dispersed in different plots, which fails to adequately consider the correlation information between dispersed plots of the same crop in spatial distribution. This study proposed a crop classification method based on multi-temporal dual polarimetric data, which considered the utilization of information between near and far spatial plots, by employing superpixel segmentation and a HyperGraph neural network, respectively. Firstly, the method utilized the dual polarimetric covariance matrix of multi-temporal data to perform superpixel segmentation on neighboring pixels, so that the segmented superpixel blocks were highly compatible with the actual plot shapes from a long-term period perspective. Then, a HyperGraph adjacency matrix was constructed, and a HyperGraph neural network (HGNN) was utilized to better learn the features of plots of the same crop that are distributed far from each other. The method fully utilizes the three dimensions of time, polarization and space information, which complement each other so as to effectively realize high-precision crop classification. The Sentinel-1 experimental results show that, under the optimal parameter settings, the classified accuracy of combined temporal superpixel scattering features using the HGNN was obviously improved, considering the near and far distance spatial correlations of crop types. Full article
(This article belongs to the Special Issue Cutting-Edge PolSAR Imaging Applications and Techniques)
Show Figures

Graphical abstract

32 pages, 58845 KiB  
Article
Using New York City’s Geographic Data in an Innovative Application of Generative Adversarial Networks (GANs) to Produce Cooling Comparisons of Urban Design
by Yuanyuan Li, Lina Zhao, Hao Zheng and Xiaozhou Yang
Land 2025, 14(7), 1393; https://doi.org/10.3390/land14071393 - 2 Jul 2025
Cited by 1 | Viewed by 498
Abstract
Urban blue–green space (UBGS) plays a critical role in mitigating the urban heat island (UHI) effect and reducing land surface temperatures (LSTs). However, existing research has not sufficiently explored the optimization of UBGS spatial configurations or their interactions with urban morphology. This study [...] Read more.
Urban blue–green space (UBGS) plays a critical role in mitigating the urban heat island (UHI) effect and reducing land surface temperatures (LSTs). However, existing research has not sufficiently explored the optimization of UBGS spatial configurations or their interactions with urban morphology. This study takes New York City as a case and systematically investigates small-scale urban cooling strategies by integrating multiple factors, including adjustments to the blue–green ratio, spatial layouts, vegetation composition, building density, building height, and layout typologies. We utilize multi-source geographic data, including LiDAR derived land cover, OpenStreetMap data, and building footprint data, together with LST data retrieved from Landsat imagery, to develop a prediction model based on generative adversarial networks (GANs). This model can rapidly generate visual LST predictions under various configuration scenarios. This study employs a combination of qualitative and quantitative metrics to evaluate the performance of different model stages, selecting the most accurate model as the final experimental framework. Furthermore, the experimental design strictly controls the study area and pixel allocation, combining manual and automated methods to ensure the comparability of different ratio configurations. The main findings indicate that a blue–green ratio of 3:7 maximizes cooling efficiency; a shrub-to-tree coverage ratio of 2:8 performs best, with tree-dominated configurations outperforming shrub-dominated ones; concentrated linear layouts achieve up to a 10.01% cooling effect; and taller buildings exhibit significantly stronger UBGS cooling performance, with super-tall areas achieving cooling effects approximately 31 percentage points higher than low-rise areas. Courtyard layouts enhance airflow and synergistic cooling effects, whereas compact designs limit the cooling potential of UBGS. This study proposes an innovative application of GANs to address a key research gap in the quantitative optimization of UBGS configurations and provides a methodological reference for sustainable microclimate planning at the neighborhood scale. Full article
Show Figures

Figure 1

20 pages, 3406 KiB  
Article
Single-Image Super-Resolution via Cascaded Non-Local Mean Network and Dual-Path Multi-Branch Fusion
by Yu Xu and Yi Wang
Sensors 2025, 25(13), 4044; https://doi.org/10.3390/s25134044 - 28 Jun 2025
Viewed by 549
Abstract
Image super-resolution (SR) aims to reconstruct high-resolution (HR) images from low-resolution (LR) inputs. It plays a crucial role in applications such as medical imaging, surveillance, and remote sensing. However, due to the ill-posed nature of the task and the inherent limitations of imaging [...] Read more.
Image super-resolution (SR) aims to reconstruct high-resolution (HR) images from low-resolution (LR) inputs. It plays a crucial role in applications such as medical imaging, surveillance, and remote sensing. However, due to the ill-posed nature of the task and the inherent limitations of imaging sensors, obtaining accurate HR images remains challenging. While numerous methods have been proposed, the traditional approaches suffer from oversmoothing and limited generalization; CNN-based models lack the ability to capture long-range dependencies; and Transformer-based solutions, although effective in modeling global context, are computationally intensive and prone to texture loss. To address these issues, we propose a hybrid CNN–Transformer architecture that cascades a pixel-wise self-attention non-local means module (PSNLM) and an adaptive dual-path multi-scale fusion block (ADMFB). The PSNLM is inspired by the non-local means (NLM) algorithm. We use weighted patches to estimate the similarity between pixels centered at each patch while limiting the search region and constructing a communication mechanism across ranges. The ADMFB enhances texture reconstruction by adaptively aggregating multi-scale features through dual attention paths. The experimental results demonstrate that our method achieves superior performance on multiple benchmarks. For instance, in challenging ×4 super-resolution, our method outperforms the second-best method by 0.0201 regarding the Structural Similarity Index (SSIM) on the BSD100 dataset. On the texture-rich Urban100 dataset, our method achieves a 26.56 dB Peak Signal-to-Noise Ratio (PSNR) and 0.8133 SSIM. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

27 pages, 4947 KiB  
Article
From Coarse to Crisp: Enhancing Tree Species Maps with Deep Learning and Satellite Imagery
by Taebin Choe, Seungpyo Jeon, Byeongcheol Kim and Seonyoung Park
Remote Sens. 2025, 17(13), 2222; https://doi.org/10.3390/rs17132222 - 28 Jun 2025
Viewed by 423
Abstract
Accurate, detailed, and up-to-date tree species distribution information is essential for effective forest management and environmental research. However, existing tree species maps face limitations in resolution and update cycle, making it difficult to meet modern demands. To overcome these limitations, this study proposes [...] Read more.
Accurate, detailed, and up-to-date tree species distribution information is essential for effective forest management and environmental research. However, existing tree species maps face limitations in resolution and update cycle, making it difficult to meet modern demands. To overcome these limitations, this study proposes a novel framework that utilizes existing medium-resolution national tree species maps as ‘weak labels’ and fuses multi-temporal Sentinel-2 and PlanetScope satellite imagery data. Specifically, a super-resolution (SR) technique, using PlanetScope imagery as a reference, was first applied to Sentinel-2 data to enhance its resolution to 2.5 m. Then, these enhanced Sentinel-2 bands were combined with PlanetScope bands to construct the final multi-spectral, multi-temporal input data. Deep learning (DL) model training data was constructed by strategically sampling information-rich pixels from the national tree species map. Applying the proposed methodology to Sobaeksan and Jirisan National Parks in South Korea, the performance of various machine learning (ML) and deep learning (DL) models was compared, including traditional ML (linear regression, random forest) and DL architectures (multilayer perceptron (MLP), spectral encoder block (SEB)—linear, and SEB-transformer). The MLP model demonstrated optimal performance, achieving over 85% overall accuracy (OA) and more than 81% accuracy in classifying spectrally similar and difficult-to-distinguish species, specifically Quercus mongolica (QM) and Quercus variabilis (QV). Furthermore, while spectral and temporal information were confirmed to contribute significantly to tree species classification, the contribution of spatial (texture) information was experimentally found to be limited at the 2.5 m resolution level. This study presents a practical method for creating high-resolution tree species maps scalable to the national level by fusing existing tree species maps with Sentinel-2 and PlanetScope imagery without requiring costly separate field surveys. Its significance lies in establishing a foundation that can contribute to various fields such as forest resource management, biodiversity conservation, and climate change research. Full article
(This article belongs to the Special Issue Digital Modeling for Sustainable Forest Management)
Show Figures

Figure 1

40 pages, 4919 KiB  
Article
NGSTGAN: N-Gram Swin Transformer and Multi-Attention U-Net Discriminator for Efficient Multi-Spectral Remote Sensing Image Super-Resolution
by Chao Zhan, Chunyang Wang, Bibo Lu, Wei Yang, Xian Zhang and Gaige Wang
Remote Sens. 2025, 17(12), 2079; https://doi.org/10.3390/rs17122079 - 17 Jun 2025
Viewed by 537
Abstract
The reconstruction of high-resolution (HR) remote sensing images (RSIs) from low-resolution (LR) counterparts is a critical task in remote sensing image super-resolution (RSISR). Recent advancements in convolutional neural networks (CNNs) and Transformers have significantly improved RSISR performance due to their capabilities in local [...] Read more.
The reconstruction of high-resolution (HR) remote sensing images (RSIs) from low-resolution (LR) counterparts is a critical task in remote sensing image super-resolution (RSISR). Recent advancements in convolutional neural networks (CNNs) and Transformers have significantly improved RSISR performance due to their capabilities in local feature extraction and global modeling. However, several limitations remain, including the underutilization of multi-scale features in RSIs, the limited receptive field of Swin Transformer’s window self-attention (WSA), and the computational complexity of existing methods. To address these issues, this paper introduces the NGSTGAN model, which employs an N-Gram Swin Transformer as the generator and a multi-attention U-Net as the discriminator. The discriminator enhances attention to multi-scale key features through the addition of channel, spatial, and pixel attention (CSPA) modules, while the generator utilizes an improved shallow feature extraction (ISFE) module to extract multi-scale and multi-directional features, enhancing the capture of complex textures and details. The N-Gram concept is introduced to expand the receptive field of Swin Transformer, and sliding window self-attention (S-WSA) is employed to facilitate interaction between neighboring windows. Additionally, channel-reducing group convolution (CRGC) is used to reduce the number of parameters and computational complexity. A cross-sensor multispectral dataset combining Landsat-8 (L8) and Sentinel-2 (S2) is constructed for the resolution enhancement of L8’s blue (B), green (G), red (R), and near-infrared (NIR) bands from 30 m to 10 m. Experiments show that NGSTGAN outperforms the state-of-the-art (SOTA) method, achieving improvements of 0.5180 dB in the peak signal-to-noise ratio (PSNR) and 0.0153 in the structural similarity index measure (SSIM) over the second best method, offering a more effective solution to the task. Full article
Show Figures

Figure 1

19 pages, 1230 KiB  
Article
A Graph Convolutional Network Framework for Area Attention and Tracking Compensation of In-Orbit Satellite
by Shuai Wang, Ruoke Wu, Yizhi Jiang, Xiaoqiang Di, Yining Mu, Guanyu Wen, Makram Ibrahim and Jinqing Li
Appl. Sci. 2025, 15(12), 6742; https://doi.org/10.3390/app15126742 - 16 Jun 2025
Viewed by 264
Abstract
In order to solve the problems of low tracking accuracy of in-orbit satellites by ground stations and slow processing speed of satellite target tracking images, this paper proposes an orbital satellite regional tracking and prediction model based on graph convolutional networks (GCNs). By [...] Read more.
In order to solve the problems of low tracking accuracy of in-orbit satellites by ground stations and slow processing speed of satellite target tracking images, this paper proposes an orbital satellite regional tracking and prediction model based on graph convolutional networks (GCNs). By performing superpixel segmentation on the satellite tracking image information, we constructed an intra-frame superpixel seed graph node network, enabling the conversion of spatial optical image information into artificial-intelligence-based graph feature data. On this basis, we propose and build an in-orbit satellite region of interest prediction model, which effectively enhances the perception of in-orbit satellite feature information and can be used for in-orbit target prediction. This model, for the first time, combines intra-frame and inter-frame graph structures to improve the sensitivity of GCNs to the spatial feature information of in-orbit satellites. Finally, the model is trained and validated using real satellite target tracking image datasets, demonstrating the effectiveness of the proposed model. Full article
Show Figures

Figure 1

9 pages, 949 KiB  
Article
A Superpixel-Based Algorithm for Detecting Optical Density Changes in Choroidal Optical Coherence Tomography Images of Diabetic Patients
by Sofia Otin, Victor Mallen-Gracia, Luis Perez-Maña, Francisco J. Ávila and Elena Garcia-Martin
Sensors 2025, 25(12), 3619; https://doi.org/10.3390/s25123619 - 9 Jun 2025
Viewed by 474
Abstract
Background: This study explored the diagnostic potential of image-processing analysis in optical coherence tomography (OCT) images to detect systemic vascular changes in individuals with systemic diseases. Methods: Ocular OCT images from two cohorts diabetic patients and healthy control subjects were analyzed. A novel [...] Read more.
Background: This study explored the diagnostic potential of image-processing analysis in optical coherence tomography (OCT) images to detect systemic vascular changes in individuals with systemic diseases. Methods: Ocular OCT images from two cohorts diabetic patients and healthy control subjects were analyzed. A novel Superpixel Segmentation (SpS) algorithm was used to process these images and extract optical image density information from ocular vascular tissue. The algorithm was applied to isolate the choroid layer for analysis of its optical properties. The procedure was performed by separate examiners, and both inter- and intra-observer repeatability were assessed. Choroidal area (CA) and choroidal optical image density (COID) metrics were used to assess structural changes in the vascular tissue and predict alterations in the choroidal parameters. Results: A total of 110 diabetic patient eye images and 92 healthy control images were processed. The results showed significant differences in CA and COID between diabetic and healthy eyes, indicating that these parameters could serve as valuable biomarkers for early vascular damage. Conclusions: The use of the SpS algorithm on OCT B-scan images allows for the identification of new parameters linked to ocular vascular damage. These findings suggest that digital image-processing techniques can reveal differences in vascular tissue, offering potential new indicators of pathology. Full article
Show Figures

Figure 1

23 pages, 5811 KiB  
Article
Multi-Attitude Hybrid Network for Remote Sensing Hyperspectral Images Super-Resolution
by Chi Chen, Yunhan Sun, Xueyan Hu, Ning Zhang, Hao Feng, Zheng Li and Yongcheng Wang
Remote Sens. 2025, 17(11), 1947; https://doi.org/10.3390/rs17111947 - 4 Jun 2025
Cited by 1 | Viewed by 579
Abstract
Benefiting from the development of deep learning, the super-resolution technology for remote sensing hyperspectral images (HSIs) has achieved impressive progress. However, due to the high coupling of complex components in remote sensing HSIs, it is challenging to achieve a complete characterization of the [...] Read more.
Benefiting from the development of deep learning, the super-resolution technology for remote sensing hyperspectral images (HSIs) has achieved impressive progress. However, due to the high coupling of complex components in remote sensing HSIs, it is challenging to achieve a complete characterization of the internal information, which in turn limits the precise reconstruction of detailed texture and spectral features. Therefore, we propose the multi-attitude hybrid network (MAHN) for extracting and characterizing information from multiple feature spaces. On the one hand, we construct the spectral hypergraph cross-attention module (SHCAM) and the spatial hypergraph self-attention module (SHSAM) based on the high and low-frequency features in the spectral and the spatial domains, respectively, which are used to capture the main structure and detail changes within the image. On the other hand, high-level semantic information in mixed pixels is parsed by spectral mixture analysis, and semantic hypergraph 3D module (SH3M) are constructed based on the abundance of each category to enhance the propagation and reconstruction of semantic information. Furthermore, to mitigate the domain discrepancies among features, we introduce a sensitive bands attention mechanism (SBAM) to enhance the cross-guidance and fusion of multi-domain features. Extensive experiments demonstrate that our method achieves optimal reconstruction results compared to other state-of-the-art algorithms while effectively reducing the computational complexity. Full article
Show Figures

Figure 1

21 pages, 10091 KiB  
Article
Scalable Hyperspectral Enhancement via Patch-Wise Sparse Residual Learning: Insights from Super-Resolved EnMAP Data
by Parth Naik, Rupsa Chakraborty, Sam Thiele and Richard Gloaguen
Remote Sens. 2025, 17(11), 1878; https://doi.org/10.3390/rs17111878 - 28 May 2025
Viewed by 720
Abstract
A majority of hyperspectral super-resolution methods aim to enhance the spatial resolution of hyperspectral imaging data (HSI) by integrating high-resolution multispectral imaging data (MSI), leveraging rich spectral information for various geospatial applications. Key challenges include spectral distortions from high-frequency spatial data, high computational [...] Read more.
A majority of hyperspectral super-resolution methods aim to enhance the spatial resolution of hyperspectral imaging data (HSI) by integrating high-resolution multispectral imaging data (MSI), leveraging rich spectral information for various geospatial applications. Key challenges include spectral distortions from high-frequency spatial data, high computational complexity, and limited training data, particularly for new-generation sensors with unique noise patterns. In this contribution, we propose a novel parallel patch-wise sparse residual learning (P2SR) algorithm for resolution enhancement based on fusion of HSI and MSI. The proposed method uses multi-decomposition techniques (i.e., Independent component analysis, Non-negative matrix factorization, and 3D wavelet transforms) to extract spatial and spectral features to form a sparse dictionary. The spectral and spatial characteristics of the scene encoded in the dictionary enable reconstruction through a first-order optimization algorithm to ensure an efficient sparse representation. The final spatially enhanced HSI is reconstructed by combining the learned features from low-resolution HSI and applying an MSI-regulated guided filter to enhance spatial fidelity while minimizing artifacts. P2SR is deployable on a high-performance computing (HPC) system with parallel processing, ensuring scalability and computational efficiency for large HSI datasets. Extensive evaluations on three diverse study sites demonstrate that P2SR consistently outperforms traditional and state-of-the-art (SOA) methods in both quantitative metrics and qualitative spatial assessments. Specifically, P2SR achieved the best average PSNR (25.2100) and SAM (12.4542) scores, indicating superior spatio-spectral reconstruction contributing to sharper spatial features, reduced mixed pixels, and enhanced geological features. P2SR also achieved the best average ERGAS (8.9295) and Q2n (0.5156), which suggests better overall fidelity across all bands and perceptual accuracy with the least spectral distortions. Importantly, we show that P2SR preserves critical spectral signatures, such as Fe2+ absorption, and improves the detection of fine-scale environmental and geological structures. P2SR’s ability to maintain spectral fidelity while enhancing spatial detail makes it a powerful tool for high-precision remote sensing applications, including mineral mapping, land-use analysis, and environmental monitoring. Full article
Show Figures

Graphical abstract

15 pages, 2498 KiB  
Article
Research on Image Stitching Based on an Improved LightGlue Algorithm
by Yuening Feng, Fei Zhang, Xiaozhan Li, Xiong Xiao, Lijun Wang and Xiaofei Xiang
Processes 2025, 13(6), 1687; https://doi.org/10.3390/pr13061687 - 28 May 2025
Viewed by 686
Abstract
In traditional centralized steel plant production monitoring systems, there are two major problems. On the one hand, the limited shooting angles of cameras make it impossible to capture comprehensive information. On the other hand, using multiple cameras to display monitoring screens separately on [...] Read more.
In traditional centralized steel plant production monitoring systems, there are two major problems. On the one hand, the limited shooting angles of cameras make it impossible to capture comprehensive information. On the other hand, using multiple cameras to display monitoring screens separately on a large screen leads to clutter and easy omission of key information. To address the above-mentioned issues, this paper proposes an image stitching technique based on an improved LightGlue algorithm. First of all, this paper employs the SuperPoint (Self-Supervised Interest Point Detection and Description) algorithm as the feature extraction algorithm. The experimental results show that this algorithm outperforms traditional algorithms both in terms of feature extraction speed and extraction accuracy. Then, the LightGlue (Local Feature Matching at Light Speed) algorithm is selected as the feature matching algorithm, and it is optimized and improved by combining it with the Agglomerative Clustering (AGG) algorithm. The experimental results indicate that this improvement effectively increases the speed of feature matching. Compared with the original LightGlue algorithm, the matching efficiency is improved by 26.2%. Finally, aiming at the problems of parallax and ghosting existing in the image fusion process, this paper proposes a pixel dynamic adaptive fusion strategy. A local homography matrix strategy is proposed in the geometric alignment stage, and a pixel difference fusion strategy is proposed in the pixel fusion stage. The experimental results show that this improvement successfully solves the problems of parallax and ghosting and achieves a good image stitching effect. Full article
Show Figures

Figure 1

42 pages, 4633 KiB  
Article
Resolution-Aware Deep Learning with Feature Space Optimization for Reliable Identity Verification in Electronic Know Your Customer Processes
by Mahasak Ketcham, Pongsarun Boonyopakorn and Thittaporn Ganokratanaa
Mathematics 2025, 13(11), 1726; https://doi.org/10.3390/math13111726 - 23 May 2025
Viewed by 667
Abstract
In modern digital transactions involving government agencies, financial institutions, and commercial enterprises, reliable identity verification is essential to ensure security and trust. Traditional methods, such as submitting photocopies of ID cards, are increasingly susceptible to identity theft and fraud. To address these challenges, [...] Read more.
In modern digital transactions involving government agencies, financial institutions, and commercial enterprises, reliable identity verification is essential to ensure security and trust. Traditional methods, such as submitting photocopies of ID cards, are increasingly susceptible to identity theft and fraud. To address these challenges, this study proposes a novel and robust identity verification framework that integrates super-resolution preprocessing, a convolutional neural network (CNN), and Monte Carlo dropout-based Bayesian uncertainty estimation for enhanced facial recognition in electronic know your customer (e-KYC) processes. The key contribution of this research lies in its ability to handle low-resolution and degraded facial images simulating real-world conditions where image quality is inconsistent while providing confidence-aware predictions to support transparent and risk-aware decision making. The proposed model is trained on facial images resized to 24 × 24 pixels, with a super-resolution module enhancing feature clarity prior to classification. By incorporating Monte Carlo dropout, the system estimates predictive uncertainty, addressing critical limitations of conventional black-box deep learning models. Experimental evaluations confirmed the effectiveness of the framework, achieving a classification accuracy of 99.7%, precision of 99.2%, recall of 99.3%, and an AUC score of 99.5% under standard testing conditions. The model also demonstrated strong robustness against noise and image blur, maintaining reliable performance even under challenging input conditions. In addition, the proposed system is designed to comply with international digital identity standards, including the Identity Assurance Level (IAL) and Authenticator Assurance Level (AAL), ensuring practical applicability in regulated environments. Overall, this research contributes a scalable, secure, and interpretable solution that advances the application of deep learning and uncertainty modeling in real-world e-KYC systems. Full article
(This article belongs to the Special Issue Advanced Studies in Mathematical Optimization and Machine Learning)
Show Figures

Figure 1

Back to TopTop