Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (62)

Search Parameters:
Keywords = face image super-resolution

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 6539 KB  
Article
A High-Precision Ionospheric Channel Estimation Method Based on Oblique Projection and Double-Space Decomposition
by Zhengkai Wei, Baiyang Guo, Zhihui Li and Qingsong Zhou
Sensors 2025, 25(18), 5727; https://doi.org/10.3390/s25185727 - 14 Sep 2025
Viewed by 789
Abstract
Accurate ionospheric channel estimation is of great significance for acquisition of ionospheric structure, error correction of remote sensing data, high-precision Synthetic Aperture Radar (SAR) imaging, over-the-horizon (OTH) detection, and the establishment of stable communication links. Traditional super-resolution channel estimation algorithms face challenges in [...] Read more.
Accurate ionospheric channel estimation is of great significance for acquisition of ionospheric structure, error correction of remote sensing data, high-precision Synthetic Aperture Radar (SAR) imaging, over-the-horizon (OTH) detection, and the establishment of stable communication links. Traditional super-resolution channel estimation algorithms face challenges in terms of multipath correlation and noise interference when estimating ionospheric channel information. Meanwhile, some super-resolution algorithms struggle to meet the requirements of real-time measurement due to their high computational complexity. In this paper, we propose the Cross-correlation Oblique Projection Pursuit (CC-OPMP) algorithm, which constructs an atom selection strategy for anti-interference correlation metric and a dual-space multipath separation mechanism based on a greedy framework to effectively suppress noise and separate neighboring multipath components. Simulations demonstrate that the CC-OPMP algorithm outperforms other algorithms in both channel estimation accuracy and computational efficiency. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

28 pages, 19672 KB  
Article
A Multi-Fidelity Data Fusion Approach Based on Semi-Supervised Learning for Image Super-Resolution in Data-Scarce Scenarios
by Hongzheng Zhu, Yingjuan Zhao, Ximing Qiao, Jinshuo Zhang, Jingnan Ma and Sheng Tong
Sensors 2025, 25(17), 5373; https://doi.org/10.3390/s25175373 - 31 Aug 2025
Viewed by 732
Abstract
Image super-resolution (SR) techniques can significantly enhance visual quality and information density. However, existing methods often rely on large amounts of paired low- and high-resolution (LR-HR) data, which limits their generalization and robustness when faced with data scarcity, distribution inconsistencies, and missing high-frequency [...] Read more.
Image super-resolution (SR) techniques can significantly enhance visual quality and information density. However, existing methods often rely on large amounts of paired low- and high-resolution (LR-HR) data, which limits their generalization and robustness when faced with data scarcity, distribution inconsistencies, and missing high-frequency details. To tackle the challenges of image reconstruction in data-scarce scenarios, this paper proposes a semi-supervised learning-driven multi-fidelity fusion (SSLMF) method, which integrates multi-fidelity data fusion (MFDF) and semi-supervised learning (SSL) to reduce reliance on high-fidelity data. More specifically, (1) an MFDF strategy is employed to leverage low-fidelity data for global structural constraints, enhancing information compensation; (2) an SSL mechanism is introduced to reduce data dependence by using only a small amount of labeled HR samples along with a large quantity of unlabeled multi-fidelity data. This framework significantly improves data efficiency and reconstruction quality. We first validate the reconstruction accuracy of SSLMF on benchmark functions and then apply it to image reconstruction tasks. The results demonstrate that SSLMF can effectively model both linear and nonlinear relationships among multi-fidelity data, maintaining high performance even with limited high-fidelity samples. Finally, its cross-disciplinary potential is illustrated through an audio restoration case study, offering a novel solution for efficient image reconstruction, especially in data-scarce scenarios where high-fidelity samples are limited. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

26 pages, 5964 KB  
Article
Super-Resolution Reconstruction of Part Images Using Adaptive Multi-Scale Object Tracking
by Yaohe Li, Long Jin, Yindi Bai, Zhiwen Song and Dongyuan Ge
Processes 2025, 13(8), 2563; https://doi.org/10.3390/pr13082563 - 14 Aug 2025
Viewed by 470
Abstract
Computer vision-based part surface inspection is widely used for quality evaluation. However, challenges such as low image quality, caused by factors like inadequate acquisition equipment, camera vibrations, and environmental conditions, often lead to reduced detection accuracy. Although super-resolution reconstruction can enhance image quality, [...] Read more.
Computer vision-based part surface inspection is widely used for quality evaluation. However, challenges such as low image quality, caused by factors like inadequate acquisition equipment, camera vibrations, and environmental conditions, often lead to reduced detection accuracy. Although super-resolution reconstruction can enhance image quality, existing methods face issues such as limited accuracy, information distortion, and high computational cost. To overcome these challenges, we propose a novel super-resolution reconstruction method for part images that incorporates adaptive multi-scale object tracking. Our approach first adaptively segments the input sequence of part images into blocks of varying scales, improving both reconstruction accuracy and computational efficiency. Optical flow is then applied to estimate the motion parameters between sequence images, followed by the construction of a feature tracking and sampling model to extract detailed features from all images, addressing information distortion caused by pixel misalignment. Finally, a non-linear reconstruction algorithm is employed to generate the high-resolution target image. Experimental results demonstrate that our method achieves superior performance in terms of both quantitative metrics and visual quality, outperforming existing methods. This contributes to a significant improvement in subsequent part detection accuracy and production efficiency. Full article
(This article belongs to the Section Manufacturing Processes and Systems)
Show Figures

Figure 1

27 pages, 8755 KB  
Article
Mapping Wetlands with High-Resolution Planet SuperDove Satellite Imagery: An Assessment of Machine Learning Models Across the Diverse Waterscapes of New Zealand
by Md. Saiful Islam Khan, Maria C. Vega-Corredor and Matthew D. Wilson
Remote Sens. 2025, 17(15), 2626; https://doi.org/10.3390/rs17152626 - 29 Jul 2025
Viewed by 1289
Abstract
(1) Background: Wetlands are ecologically significant ecosystems that support biodiversity and contribute to essential environmental functions such as water purification, carbon storage and flood regulation. However, these ecosystems face increasing pressures from land-use change and degradation, prompting the need for scalable and accurate [...] Read more.
(1) Background: Wetlands are ecologically significant ecosystems that support biodiversity and contribute to essential environmental functions such as water purification, carbon storage and flood regulation. However, these ecosystems face increasing pressures from land-use change and degradation, prompting the need for scalable and accurate classification methods to support conservation and policy efforts. In this research, our motivation was to test whether high-spatial-resolution PlanetScope imagery can be used with pixel-based machine learning to support the mapping and monitoring of wetlands at a national scale. (2) Methods: This study compared four machine learning classification models—Random Forest (RF), XGBoost (XGB), Histogram-Based Gradient Boosting (HGB) and a Multi-Layer Perceptron Classifier (MLPC)—to detect and map wetland areas across New Zealand. All models were trained using eight-band SuperDove satellite imagery from PlanetScope, with a spatial resolution of ~3 m, and ancillary geospatial datasets representing topography and soil drainage characteristics, each of which is available globally. (3) Results: All four machine learning models performed well in detecting wetlands from SuperDove imagery and environmental covariates, with varying strengths. The highest accuracy was achieved using all eight image bands alongside features created from supporting geospatial data. For binary wetland classification, the highest F1 scores were recorded by XGB (0.73) and RF/HGB (both 0.72) when including all covariates. MLPC also showed competitive performance (wetland F1 score of 0.71), despite its relatively lower spatial consistency. However, each model over-predicts total wetland area at a national level, an issue which was able to be reduced by increasing the classification probability threshold and spatial filtering. (4) Conclusions: The comparative analysis highlights the strengths and trade-offs of RF, XGB, HGB and MLPC models for wetland classification. While all four methods are viable, RF offers some key advantages, including ease of deployment and transferability, positioning it as a promising candidate for scalable, high-resolution wetland monitoring across diverse ecological settings. Further work is required for verification of small-scale wetlands (<~0.5 ha) and the addition of fine-spatial-scale covariates. Full article
Show Figures

Figure 1

28 pages, 3794 KB  
Article
A Robust System for Super-Resolution Imaging in Remote Sensing via Attention-Based Residual Learning
by Rogelio Reyes-Reyes, Yeredith G. Mora-Martinez, Beatriz P. Garcia-Salgado, Volodymyr Ponomaryov, Jose A. Almaraz-Damian, Clara Cruz-Ramos and Sergiy Sadovnychiy
Mathematics 2025, 13(15), 2400; https://doi.org/10.3390/math13152400 - 25 Jul 2025
Viewed by 905
Abstract
Deep learning-based super-resolution (SR) frameworks are widely used in remote sensing applications. However, existing SR models still face limitations, particularly in recovering contours, fine features, and textures, as well as in effectively integrating channel information. To address these challenges, this study introduces a [...] Read more.
Deep learning-based super-resolution (SR) frameworks are widely used in remote sensing applications. However, existing SR models still face limitations, particularly in recovering contours, fine features, and textures, as well as in effectively integrating channel information. To address these challenges, this study introduces a novel residual model named OARN (Optimized Attention Residual Network) specifically designed to enhance the visual quality of low-resolution images. The network operates on the Y channel of the YCbCr color space and integrates LKA (Large Kernel Attention) and OCM (Optimized Convolutional Module) blocks. These components can restore large-scale spatial relationships and refine textures and contours, improving feature reconstruction without significantly increasing computational complexity. The performance of OARN was evaluated using satellite images from WorldView-2, GaoFen-2, and Microsoft Virtual Earth. Evaluation was conducted using objective quality metrics, such as Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index Measure (SSIM), Edge Preservation Index (EPI), and Perceptual Image Patch Similarity (LPIPS), demonstrating superior results compared to state-of-the-art methods in both objective measurements and subjective visual perception. Moreover, OARN achieves this performance while maintaining computational efficiency, offering a balanced trade-off between processing time and reconstruction quality. Full article
Show Figures

Figure 1

21 pages, 4388 KB  
Article
An Omni-Dimensional Dynamic Convolutional Network for Single-Image Super-Resolution Tasks
by Xi Chen, Ziang Wu, Weiping Zhang, Tingting Bi and Chunwei Tian
Mathematics 2025, 13(15), 2388; https://doi.org/10.3390/math13152388 - 25 Jul 2025
Viewed by 820
Abstract
The goal of single-image super-resolution (SISR) tasks is to generate high-definition images from low-quality inputs, with practical uses spanning healthcare diagnostics, aerial imaging, and surveillance systems. Although cnns have considerably improved image reconstruction quality, existing methods still face limitations, including inadequate restoration of [...] Read more.
The goal of single-image super-resolution (SISR) tasks is to generate high-definition images from low-quality inputs, with practical uses spanning healthcare diagnostics, aerial imaging, and surveillance systems. Although cnns have considerably improved image reconstruction quality, existing methods still face limitations, including inadequate restoration of high-frequency details, high computational complexity, and insufficient adaptability to complex scenes. To address these challenges, we propose an Omni-dimensional Dynamic Convolutional Network (ODConvNet) tailored for SISR tasks. Specifically, ODConvNet comprises four key components: a Feature Extraction Block (FEB) that captures low-level spatial features; an Omni-dimensional Dynamic Convolution Block (DCB), which utilizes a multidimensional attention mechanism to dynamically reweight convolution kernels across spatial, channel, and kernel dimensions, thereby enhancing feature expressiveness and context modeling; a Deep Feature Extraction Block (DFEB) that stacks multiple convolutional layers with residual connections to progressively extract and fuse high-level features; and a Reconstruction Block (RB) that employs subpixel convolution to upscale features and refine the final HR output. This mechanism significantly enhances feature extraction and effectively captures rich contextual information. Additionally, we employ an improved residual network structure combined with a refined Charbonnier loss function to alleviate gradient vanishing and exploding to enhance the robustness of model training. Extensive experiments conducted on widely used benchmark datasets, including DIV2K, Set5, Set14, B100, and Urban100, demonstrate that, compared with existing deep learning-based SR methods, our ODConvNet method improves Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM), and the visual quality of SR images is also improved. Ablation studies further validate the effectiveness and contribution of each component in our network. The proposed ODConvNet offers an effective, flexible, and efficient solution for the SISR task and provides promising directions for future research. Full article
Show Figures

Figure 1

23 pages, 4237 KB  
Article
Debris-Flow Erosion Volume Estimation Using a Single High-Resolution Optical Satellite Image
by Peng Zhang, Shang Wang, Guangyao Zhou, Yueze Zheng, Kexin Li and Luyan Ji
Remote Sens. 2025, 17(14), 2413; https://doi.org/10.3390/rs17142413 - 12 Jul 2025
Viewed by 693
Abstract
Debris flows pose significant risks to mountainous regions, and quick, accurate volume estimation is crucial for hazard assessment and post-disaster response. Traditional volume estimation methods, such as ground surveys and aerial photogrammetry, are often limited by cost, accessibility, and timeliness. While remote sensing [...] Read more.
Debris flows pose significant risks to mountainous regions, and quick, accurate volume estimation is crucial for hazard assessment and post-disaster response. Traditional volume estimation methods, such as ground surveys and aerial photogrammetry, are often limited by cost, accessibility, and timeliness. While remote sensing offers wide coverage, existing optical and Synthetic Aperture Radar (SAR)-based techniques face challenges in direct volume estimation due to resolution constraints and rapid terrain changes. This study proposes a Super-Resolution Shape from Shading (SRSFS) approach enhanced by a Non-local Piecewise-smooth albedo Constraint (NPC), hereafter referred to as NPC SRSFS, to estimate debris-flow erosion volume using single high-resolution optical satellite imagery. By integrating publicly available global Digital Elevation Model (DEM) data as prior terrain reference, the method enables accurate post-disaster topography reconstruction from a single optical image, thereby reducing reliance on stereo imagery. The NPC constraint improves the robustness of albedo estimation under heterogeneous surface conditions, enhancing depth recovery accuracy. The methodology is evaluated using Gaofen-6 satellite imagery, with quantitative comparisons to aerial Light Detection and Ranging (LiDAR) data. Results show that the proposed method achieves reliable terrain reconstruction and erosion volume estimates, with accuracy comparable to airborne LiDAR. This study demonstrates the potential of NPC SRSFS as a rapid, cost-effective alternative for post-disaster debris-flow assessment. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Figure 1

17 pages, 7786 KB  
Article
Video Coding Based on Ladder Subband Recovery and ResGroup Module
by Libo Wei, Aolin Zhang, Lei Liu, Jun Wang and Shuai Wang
Entropy 2025, 27(7), 734; https://doi.org/10.3390/e27070734 - 8 Jul 2025
Viewed by 512
Abstract
With the rapid development of video encoding technology in the field of computer vision, the demand for tasks such as video frame reconstruction, denoising, and super-resolution has been continuously increasing. However, traditional video encoding methods typically focus on extracting spatial or temporal domain [...] Read more.
With the rapid development of video encoding technology in the field of computer vision, the demand for tasks such as video frame reconstruction, denoising, and super-resolution has been continuously increasing. However, traditional video encoding methods typically focus on extracting spatial or temporal domain information, often facing challenges of insufficient accuracy and information loss when reconstructing high-frequency details, edges, and textures of images. To address this issue, this paper proposes an innovative LadderConv framework, which combines discrete wavelet transform (DWT) with spatial and channel attention mechanisms. By progressively recovering wavelet subbands, it effectively enhances the video frame encoding quality. Specifically, the LadderConv framework adopts a stepwise recovery approach for wavelet subbands, first processing high-frequency detail subbands with relatively less information, then enhancing the interaction between these subbands, and ultimately synthesizing a high-quality reconstructed image through inverse wavelet transform. Moreover, the framework introduces spatial and channel attention mechanisms, which further strengthen the focus on key regions and channel features, leading to notable improvements in detail restoration and image reconstruction accuracy. To optimize the performance of the LadderConv framework, particularly in detail recovery and high-frequency information extraction tasks, this paper designs an innovative ResGroup module. By using multi-layer convolution operations along with feature map compression and recovery, the ResGroup module enhances the network’s expressive capability and effectively reduces computational complexity. The ResGroup module captures multi-level features from low level to high level and retains rich feature information through residual connections, thus improving the overall reconstruction performance of the model. In experiments, the combination of the LadderConv framework and the ResGroup module demonstrates superior performance in video frame reconstruction tasks, particularly in recovering high-frequency information, image clarity, and detail representation. Full article
(This article belongs to the Special Issue Rethinking Representation Learning in the Age of Large Models)
Show Figures

Figure 1

23 pages, 14051 KB  
Article
A Novel Method for Water Surface Debris Detection Based on YOLOV8 with Polarization Interference Suppression
by Yi Chen, Honghui Lin, Lin Xiao, Maolin Zhang and Pingjun Zhang
Photonics 2025, 12(6), 620; https://doi.org/10.3390/photonics12060620 - 18 Jun 2025
Viewed by 837
Abstract
Aquatic floating debris detection is a key technological foundation for ecological monitoring and integrated water environment management. It holds substantial scientific and practical value in applications such as pollution source tracing, floating debris control, and maritime navigation safety. However, this field faces ongoing [...] Read more.
Aquatic floating debris detection is a key technological foundation for ecological monitoring and integrated water environment management. It holds substantial scientific and practical value in applications such as pollution source tracing, floating debris control, and maritime navigation safety. However, this field faces ongoing challenges due to water surface polarization. Reflections of polarized light produce intense glare, resulting in localized overexposure, detail loss, and geometric distortion in captured images. These optical artifacts severely impair the performance of conventional detection algorithms, increasing both false positives and missed detections. To overcome these imaging challenges in complex aquatic environments, we propose a novel YOLOv8-based detection framework with integrated polarized light suppression mechanisms. The framework consists of four key components: a fisheye distortion correction module, a polarization feature processing layer, a customized residual network with Squeeze-and-Excitation (SE) attention, and a cascaded pipeline for super-resolution reconstruction and deblurring. Additionally, we developed the PSF-IMG dataset (Polarized Surface Floats), which includes common floating debris types such as plastic bottles, bags, and foam boards. Extensive experiments demonstrate the network’s robustness in suppressing polarization artifacts and enhancing feature stability under dynamic optical conditions. Full article
(This article belongs to the Special Issue Advancements in Optical Measurement Techniques and Applications)
Show Figures

Figure 1

27 pages, 11612 KB  
Article
FACDIM: A Face Image Super-Resolution Method That Integrates Conditional Diffusion Models with Prior Attributes
by Jianhua Ren, Yuze Guo and Qiangkui Leng
Electronics 2025, 14(10), 2070; https://doi.org/10.3390/electronics14102070 - 20 May 2025
Viewed by 1124
Abstract
Facial image super-resolution seeks to reconstruct high-quality details from low-resolution inputs, yet traditional methods, such as interpolation, convolutional neural networks (CNNs), and generative adversarial networks (GANs), often fall short, suffering from insufficient realism, loss of high-frequency details, and training instability. Furthermore, many existing [...] Read more.
Facial image super-resolution seeks to reconstruct high-quality details from low-resolution inputs, yet traditional methods, such as interpolation, convolutional neural networks (CNNs), and generative adversarial networks (GANs), often fall short, suffering from insufficient realism, loss of high-frequency details, and training instability. Furthermore, many existing models inadequately incorporate facial structural attributes and semantic information, leading to semantically inconsistent generated images. To overcome these limitations, this study introduces an attribute-prior conditional diffusion implicit model that enhances the controllability of super-resolution generation and improves detail restoration capabilities. Methodologically, the framework consists of four components: a pre-super-resolution module, a facial attribute extraction module, a global feature encoder, and an enhanced conditional diffusion implicit model. Specifically, low-resolution images are subjected to preliminary super-resolution and attribute extraction, followed by adaptive group normalization to integrate feature vectors. Additionally, residual convolutional blocks are incorporated into the diffusion model to utilize attribute priors, complemented by self-attention mechanisms and skip connections to optimize feature transmission. Experiments conducted on the CelebA and FFHQ datasets demonstrate that the proposed model achieves an increase of 2.16 dB in PSNR and 0.08 in SSIM under an 8× magnification factor compared to SR3, with the generated images displaying more realistic textures. Moreover, manual adjustment of attribute vectors allows for directional control over generation outcomes (e.g., modifying facial features or lighting conditions), ensuring alignment with anthropometric characteristics. This research provides a flexible and robust solution for high-fidelity face super-resolution, offering significant advantages in detail preservation and user controllability. Full article
(This article belongs to the Special Issue AI-Driven Image Processing: Theory, Methods, and Applications)
Show Figures

Figure 1

14 pages, 8597 KB  
Article
AI-Based Enhancing of xBn MWIR Thermal Camera Performance at 180 Kelvin
by Michael Zadok, Zeev Zalevsky and Benjamin Milgrom
Sensors 2025, 25(10), 3200; https://doi.org/10.3390/s25103200 - 19 May 2025
Viewed by 741
Abstract
Thermal imaging technology has revolutionized various fields, but current high operating temperature (HOT) mid-wave infrared (MWIR) cameras, particularly those based on xBn detectors, face limitations in size and cost due to the need for cooling to 150 Kelvin. This study explores the potential [...] Read more.
Thermal imaging technology has revolutionized various fields, but current high operating temperature (HOT) mid-wave infrared (MWIR) cameras, particularly those based on xBn detectors, face limitations in size and cost due to the need for cooling to 150 Kelvin. This study explores the potential of extending the operating temperature of these cameras to 180 Kelvin, leveraging advanced AI algorithms to mitigate the increased thermal noise expected at higher temperatures. This research investigates the feasibility and effectiveness of this approach for remote sensing applications, combining experimental data with cutting-edge image enhancement techniques like Enhanced Super-Resolution Generative Adversarial Networks (ESRGAN). The findings demonstrate the potential of 180 Kelvin operation for xBn MWIR cameras, particularly in daylight conditions, paving the way for a new generation of more affordable and compact thermal imaging systems. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

16 pages, 3751 KB  
Article
Improved Face Image Super-Resolution Model Based on Generative Adversarial Network
by Qingyu Liu, Yeguo Sun, Lei Chen and Lei Liu
J. Imaging 2025, 11(5), 163; https://doi.org/10.3390/jimaging11050163 - 19 May 2025
Cited by 1 | Viewed by 1801
Abstract
Image super-resolution (SR) models based on the generative adversarial network (GAN) face challenges such as unnatural facial detail restoration and local blurring. This paper proposes an improved GAN-based model to address these issues. First, a Multi-scale Hybrid Attention Residual Block (MHARB) is designed, [...] Read more.
Image super-resolution (SR) models based on the generative adversarial network (GAN) face challenges such as unnatural facial detail restoration and local blurring. This paper proposes an improved GAN-based model to address these issues. First, a Multi-scale Hybrid Attention Residual Block (MHARB) is designed, which dynamically enhances feature representation in critical face regions through dual-branch convolution and channel-spatial attention. Second, an Edge-guided Enhancement Block (EEB) is introduced, generating adaptive detail residuals by combining edge masks and channel attention to accurately recover high-frequency textures. Furthermore, a multi-scale discriminator with a weighted sub-discriminator loss is developed to balance global structural and local detail generation quality. Additionally, a phase-wise training strategy with dynamic adjustment of learning rate (Lr) and loss function weights is implemented to improve the realism of super-resolved face images. Experiments on the CelebA-HQ dataset demonstrate that the proposed model achieves a PSNR of 23.35 dB, a SSIM of 0.7424, and a LPIPS of 24.86, outperforming classical models and delivering superior visual quality in high-frequency regions. Notably, this model also surpasses the SwinIR model (PSNR: 23.28 dB → 23.35 dB, SSIM: 0.7340 → 0.7424, and LPIPS: 30.48 → 24.86), validating the effectiveness of the improved model and the training strategy in preserving facial details. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

19 pages, 2806 KB  
Article
SP-IGAN: An Improved GAN Framework for Effective Utilization of Semantic Priors in Real-World Image Super-Resolution
by Meng Wang, Zhengnan Li, Haipeng Liu, Zhaoyu Chen and Kewei Cai
Entropy 2025, 27(4), 414; https://doi.org/10.3390/e27040414 - 11 Apr 2025
Cited by 1 | Viewed by 808
Abstract
Single-image super-resolution (SISR) based on GANs has achieved significant progress. However, these methods still face challenges when reconstructing locally consistent textures due to a lack of semantic understanding of image categories. This highlights the necessity of focusing on contextual information comprehension and the [...] Read more.
Single-image super-resolution (SISR) based on GANs has achieved significant progress. However, these methods still face challenges when reconstructing locally consistent textures due to a lack of semantic understanding of image categories. This highlights the necessity of focusing on contextual information comprehension and the acquisition of high-frequency details in model design. To address this issue, we propose the Semantic Prior-Improved GAN (SP-IGAN) framework, which incorporates additional contextual semantic information into the Real-ESRGAN model. The framework consists of two branches. The main branch introduces a Graph Convolutional Channel Attention (GCCA) module to transform channel dependencies into adjacency relationships between feature vertices, thereby enhancing pixel associations. The auxiliary branch strengthens the correlation between semantic category information and regional textures in the Residual-in-Residual Dense Block (RRDB) module. The auxiliary branch employs a pretrained segmentation model to accurately extract regional semantic information from the input low-resolution image. This information is injected into the RRDB module through Spatial Feature Transform (SFT) layers, generating more accurate and semantically consistent texture details. Additionally, a wavelet loss is incorporated into the loss function to capture high-frequency details that are often overlooked. The experimental results demonstrate that the proposed SP-IGAN outperforms state-of-the-art (SOTA) super-resolution models across multiple public datasets. For the X4 super-resolution task, SP-IGAN achieves a 0.55 dB improvement in Peak Signal-to-Noise Ratio (PSNR) and a 0.0363 increase in Structural Similarity Index (SSIM) compared to the baseline model Real-ESRGAN. Full article
Show Figures

Figure 1

17 pages, 3528 KB  
Article
Enhanced Detection of Small Unmanned Aerial System Using Noise Suppression Super-Resolution Detector for Effective Airspace Surveillance
by Jiho Yoo and Jeongho Cho
Appl. Sci. 2025, 15(6), 3076; https://doi.org/10.3390/app15063076 - 12 Mar 2025
Viewed by 1991
Abstract
Small unmanned aerial systems have become increasingly prevalent in various fields, including agriculture, logistics and the public sector, but concerns over misuse, such as military intrusions and terrorist attacks, highlight the necessity for effective aerial surveillance. Although conventional radar systems can detect large [...] Read more.
Small unmanned aerial systems have become increasingly prevalent in various fields, including agriculture, logistics and the public sector, but concerns over misuse, such as military intrusions and terrorist attacks, highlight the necessity for effective aerial surveillance. Although conventional radar systems can detect large areas, they face challenges in accurately identifying small drones. In contrast, vision sensors offer high-resolution identification but encounter challenges in long-range detection and real-time processing. To address these limitations, this study proposes a vision sensor-based detection framework, termed the noise suppression super-resolution detector (NSSRD). To ensure the reliability and real-time capability of small drone detection, NSSRD integrates image segmentation, noise suppression, super-resolution transformation, and efficient detection processes. NSSRD divides the surveillance area into uniform sections, applies a bilateral filter to suppress noise before passing the images to an object detection model, and uses a region of interest selection process to reduce the detection area and computational load. The experimental results demonstrate that NSSRD outperforms existing models, achieving a 24% improvement in the true positive rate and a 25% increase in recall at an altitude of 40 m, validating its superior performance. Full article
Show Figures

Figure 1

30 pages, 22071 KB  
Article
Analysis of Optical Errors in Joint Fabry–Pérot Interferometer–Fourier-Transform Imaging Spectroscopy Interferometric Super-Resolution Systems
by Yu Zhang, Qunbo Lv, Jianwei Wang, Yinhui Tang, Jia Si, Xinwen Chen and Yangyang Liu
Appl. Sci. 2025, 15(6), 2938; https://doi.org/10.3390/app15062938 - 8 Mar 2025
Viewed by 1435
Abstract
Fourier-transform imaging spectroscopy (FTIS) faces inherent limitations in spectral resolution due to the maximum optical path difference (OPD) achievable by its interferometer. To overcome this constraint, we propose a novel spectral super-resolution technology integrating a Fabry–Pérot interferometer (FPI) with FTIS, termed multi-component joint [...] Read more.
Fourier-transform imaging spectroscopy (FTIS) faces inherent limitations in spectral resolution due to the maximum optical path difference (OPD) achievable by its interferometer. To overcome this constraint, we propose a novel spectral super-resolution technology integrating a Fabry–Pérot interferometer (FPI) with FTIS, termed multi-component joint interferometric hyperspectral imaging (MJI-HI). This method leverages the FPI to periodically modulate the target spectrum, enabling FTIS to capture a modulated interferogram. By encoding high-frequency spectral interference information into low-frequency interference regions through FPI modulation, an advanced inversion algorithm is developed to reconstruct the encoded high-frequency components, thereby achieving spectral super-resolution. This study analyzes the impact of primary optical errors and tolerance thresholds in the FPI and FTIS on the interferograms and spectral fidelity of MJI-HI, along with proposing algorithmic improvements. Notably, certain errors in the FTIS and FPI exhibit mutual interference. The theoretical framework for error analysis is validated and discussed through numerical simulations, providing critical theoretical support for subsequent instrument development and laying a foundation for advancing novel spectral super-resolution technologies. Full article
(This article belongs to the Special Issue Spectral Detection: Technologies and Applications—2nd Edition)
Show Figures

Figure 1

Back to TopTop