Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (76)

Search Parameters:
Keywords = digital super resolution

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 8091 KiB  
Article
Leveraging Synthetic Degradation for Effective Training of Super-Resolution Models in Dermatological Images
by Francesco Branciforti, Kristen M. Meiburger, Elisa Zavattaro, Paola Savoia and Massimo Salvi
Electronics 2025, 14(15), 3138; https://doi.org/10.3390/electronics14153138 - 6 Aug 2025
Viewed by 319
Abstract
Teledermatology relies on digital transfer of dermatological images, but compression and resolution differences compromise diagnostic quality. Image enhancement techniques are crucial to compensate for these differences and improve quality for both clinical assessment and AI-based analysis. We developed a customized image degradation pipeline [...] Read more.
Teledermatology relies on digital transfer of dermatological images, but compression and resolution differences compromise diagnostic quality. Image enhancement techniques are crucial to compensate for these differences and improve quality for both clinical assessment and AI-based analysis. We developed a customized image degradation pipeline simulating common artifacts in dermatological images, including blur, noise, downsampling, and compression. This synthetic degradation approach enabled effective training of DermaSR-GAN, a super-resolution generative adversarial network tailored for dermoscopic images. The model was trained on 30,000 high-quality ISIC images and evaluated on three independent datasets (ISIC Test, Novara Dermoscopic, PH2) using structural similarity and no-reference quality metrics. DermaSR-GAN achieved statistically significant improvements in quality scores across all datasets, with up to 23% enhancement in perceptual quality metrics (MANIQA). The model preserved diagnostic details while doubling resolution and surpassed existing approaches, including traditional interpolation methods and state-of-the-art deep learning techniques. Integration with downstream classification systems demonstrated up to 14.6% improvement in class-specific accuracy for keratosis-like lesions compared to original images. Synthetic degradation represents a promising approach for training effective super-resolution models in medical imaging, with significant potential for enhancing teledermatology applications and computer-aided diagnosis systems. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

14 pages, 16969 KiB  
Article
FTT: A Frequency-Aware Texture Matching Transformer for Digital Bathymetry Model Super-Resolution
by Peikun Xiao, Jianping Wu and Yingjie Wang
J. Mar. Sci. Eng. 2025, 13(7), 1365; https://doi.org/10.3390/jmse13071365 - 17 Jul 2025
Viewed by 231
Abstract
Deep learning has shown significant advantages over traditional spatial interpolation methods in single image super-resolution (SISR). Recently, many studies have applied super-resolution (SR) methods to generate high-resolution (HR) digital bathymetry models (DBMs), but substantial differences between DBM and natural images have been ignored, [...] Read more.
Deep learning has shown significant advantages over traditional spatial interpolation methods in single image super-resolution (SISR). Recently, many studies have applied super-resolution (SR) methods to generate high-resolution (HR) digital bathymetry models (DBMs), but substantial differences between DBM and natural images have been ignored, which leads to serious distortions and inaccuracies. Given the critical role of HR DBM in marine resource exploitation, economic development, and scientific innovation, we propose a frequency-aware texture matching transformer (FTT) for DBM SR, incorporating global terrain feature extraction (GTFE), high-frequency feature extraction (HFFE), and a terrain matching block (TMB). GTFE has the capability to perceive spatial heterogeneity and spatial locations, allowing it to accurately capture large-scale terrain features. HFFE can explicitly extract high-frequency priors beneficial for DBM SR and implicitly refine the representation of high-frequency information in the global terrain feature. TMB improves fidelity of generated HR DBM by generating position offsets to restore warped textures in deep features. Experimental results have demonstrated that the proposed FTT has superior performance in terms of elevation, slope, aspect, and fidelity of generated HR DBM. Notably, the root mean square error (RMSE) of elevation in steep terrain has been reduced by 4.89 m, which is a significant improvement in the accuracy and precision of the reconstruction. This research holds significant implications for improving the accuracy of DBM SR methods and the usefulness of HR bathymetry products for future marine research. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

23 pages, 4237 KiB  
Article
Debris-Flow Erosion Volume Estimation Using a Single High-Resolution Optical Satellite Image
by Peng Zhang, Shang Wang, Guangyao Zhou, Yueze Zheng, Kexin Li and Luyan Ji
Remote Sens. 2025, 17(14), 2413; https://doi.org/10.3390/rs17142413 - 12 Jul 2025
Viewed by 398
Abstract
Debris flows pose significant risks to mountainous regions, and quick, accurate volume estimation is crucial for hazard assessment and post-disaster response. Traditional volume estimation methods, such as ground surveys and aerial photogrammetry, are often limited by cost, accessibility, and timeliness. While remote sensing [...] Read more.
Debris flows pose significant risks to mountainous regions, and quick, accurate volume estimation is crucial for hazard assessment and post-disaster response. Traditional volume estimation methods, such as ground surveys and aerial photogrammetry, are often limited by cost, accessibility, and timeliness. While remote sensing offers wide coverage, existing optical and Synthetic Aperture Radar (SAR)-based techniques face challenges in direct volume estimation due to resolution constraints and rapid terrain changes. This study proposes a Super-Resolution Shape from Shading (SRSFS) approach enhanced by a Non-local Piecewise-smooth albedo Constraint (NPC), hereafter referred to as NPC SRSFS, to estimate debris-flow erosion volume using single high-resolution optical satellite imagery. By integrating publicly available global Digital Elevation Model (DEM) data as prior terrain reference, the method enables accurate post-disaster topography reconstruction from a single optical image, thereby reducing reliance on stereo imagery. The NPC constraint improves the robustness of albedo estimation under heterogeneous surface conditions, enhancing depth recovery accuracy. The methodology is evaluated using Gaofen-6 satellite imagery, with quantitative comparisons to aerial Light Detection and Ranging (LiDAR) data. Results show that the proposed method achieves reliable terrain reconstruction and erosion volume estimates, with accuracy comparable to airborne LiDAR. This study demonstrates the potential of NPC SRSFS as a rapid, cost-effective alternative for post-disaster debris-flow assessment. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Figure 1

18 pages, 6678 KiB  
Article
HIEN: A Hybrid Interaction Enhanced Network for Horse Iris Super-Resolution
by Ao Zhang, Bin Guo, Xing Liu and Wei Liu
Appl. Sci. 2025, 15(13), 7191; https://doi.org/10.3390/app15137191 - 26 Jun 2025
Viewed by 299
Abstract
Horse iris recognition is a non-invasive identification method with great potential for precise management in intelligent horse farms. However, horses’ natural vigilance often leads to stress and resistance when exposed to close-range infrared cameras. This behavior makes it challenging to capture clear iris [...] Read more.
Horse iris recognition is a non-invasive identification method with great potential for precise management in intelligent horse farms. However, horses’ natural vigilance often leads to stress and resistance when exposed to close-range infrared cameras. This behavior makes it challenging to capture clear iris images, thereby reducing recognition performance. This paper addresses the challenge of generating high-resolution iris images from existing low-resolution counterparts. To this end, we propose a novel hybrid-architecture image super-resolution (SR) network. Central to our approach is the design of Paired Asymmetric Transformer Block (PATB), which incorporates Contextual Query Generator (CQG) to efficiently capture contextual information and model global feature interactions. Furthermore, we introduce an Efficient Residual Dense Block (ERDB), specifically engineered to effectively extract finer-grained local features inherent in the image data. By integrating PATB and ERDB, our network achieves superior fusion of global contextual awareness and local detail information, thereby significantly enhancing the reconstruction quality of horse iris images. Experimental evaluations on our self-constructed dataset of horse irises demonstrate the effectiveness of the proposed method. In terms of standard image quality metrics, it achieves the PSNR of 30.5988 dB and SSIM of 0.8552. Moreover, in terms of identity-recognition performance, the method achieves Precision, Recall, and F1-Score of 81.48%, 74.38%, and 77.77%, respectively. This study provides a useful contribution to digital horse farm management and supports the ongoing development of smart animal husbandry. Full article
Show Figures

Figure 1

16 pages, 1830 KiB  
Article
Super-Resolution Contrast-Enhanced Ultrasound Examination Down to the Microvasculature Enables Quantitative Analysis of Liver Lesions: First Results
by Ulrich Kaiser, Ursula Vehling-Kaiser, Fabian Kück, Mia Gilanschah, Friedrich Jung and Ernst Michael Jung
Life 2025, 15(7), 991; https://doi.org/10.3390/life15070991 - 20 Jun 2025
Viewed by 735
Abstract
This article investigates the first use of dynamic microvascularization of liver lesions in clinical evaluation using real-time super-resolution contrast-enhanced ultrasound (SR CEUS). A retrospective analysis of SR CEUS examinations of liver lesions was performed. All examinations were conducted using an SC7-1U convex probe [...] Read more.
This article investigates the first use of dynamic microvascularization of liver lesions in clinical evaluation using real-time super-resolution contrast-enhanced ultrasound (SR CEUS). A retrospective analysis of SR CEUS examinations of liver lesions was performed. All examinations were conducted using an SC7-1U convex probe after the bolus injection of the ultrasound contrast agent. Digital cine loops were stored for independent evaluation. The evaluation was performed with respect to parallel measuring lines (the diameter corresponded to the capillary density) in the area of the reference lesion, liver tissue, and liver capsule using statistical analysis. In total, 65 patients (female n = 28; male n = 37; average age 57.8 ± 17.2 years) were evaluated. The examined liver lesions were mostly benign masses (n = 46). Mild liver fibrosis (stage F1) was present in most cases (n = 35). The lesions examined were located at an average tissue depth of 6.07 ± 2.47 cm. The highest number of lesion measurements was observed in the malignant lesion group. Significant differences in the measurements were found when comparing benign lesions with the capillary region (p < 0.001) and normal liver tissue (p < 0.01). The use of SR CEUS opens up new possibilities for the quantification of neovascularization, assessment of microvascular changes, and evaluation of the follow-up of intrahepatic interventions. Full article
(This article belongs to the Section Medical Research)
Show Figures

Figure 1

42 pages, 4633 KiB  
Article
Resolution-Aware Deep Learning with Feature Space Optimization for Reliable Identity Verification in Electronic Know Your Customer Processes
by Mahasak Ketcham, Pongsarun Boonyopakorn and Thittaporn Ganokratanaa
Mathematics 2025, 13(11), 1726; https://doi.org/10.3390/math13111726 - 23 May 2025
Viewed by 787
Abstract
In modern digital transactions involving government agencies, financial institutions, and commercial enterprises, reliable identity verification is essential to ensure security and trust. Traditional methods, such as submitting photocopies of ID cards, are increasingly susceptible to identity theft and fraud. To address these challenges, [...] Read more.
In modern digital transactions involving government agencies, financial institutions, and commercial enterprises, reliable identity verification is essential to ensure security and trust. Traditional methods, such as submitting photocopies of ID cards, are increasingly susceptible to identity theft and fraud. To address these challenges, this study proposes a novel and robust identity verification framework that integrates super-resolution preprocessing, a convolutional neural network (CNN), and Monte Carlo dropout-based Bayesian uncertainty estimation for enhanced facial recognition in electronic know your customer (e-KYC) processes. The key contribution of this research lies in its ability to handle low-resolution and degraded facial images simulating real-world conditions where image quality is inconsistent while providing confidence-aware predictions to support transparent and risk-aware decision making. The proposed model is trained on facial images resized to 24 × 24 pixels, with a super-resolution module enhancing feature clarity prior to classification. By incorporating Monte Carlo dropout, the system estimates predictive uncertainty, addressing critical limitations of conventional black-box deep learning models. Experimental evaluations confirmed the effectiveness of the framework, achieving a classification accuracy of 99.7%, precision of 99.2%, recall of 99.3%, and an AUC score of 99.5% under standard testing conditions. The model also demonstrated strong robustness against noise and image blur, maintaining reliable performance even under challenging input conditions. In addition, the proposed system is designed to comply with international digital identity standards, including the Identity Assurance Level (IAL) and Authenticator Assurance Level (AAL), ensuring practical applicability in regulated environments. Overall, this research contributes a scalable, secure, and interpretable solution that advances the application of deep learning and uncertainty modeling in real-world e-KYC systems. Full article
(This article belongs to the Special Issue Advanced Studies in Mathematical Optimization and Machine Learning)
Show Figures

Figure 1

30 pages, 19284 KiB  
Article
A Multi-Path Feature Extraction and Transformer Feature Enhancement DEM Super-Resolution Reconstruction Network
by Mingqiang Guo, Feng Xiong, Ying Huang, Zhizheng Zhang and Jiaming Zhang
Remote Sens. 2025, 17(10), 1737; https://doi.org/10.3390/rs17101737 - 16 May 2025
Cited by 1 | Viewed by 592
Abstract
In recent years, great progress has been made in the field of super-resolution (SR) reconstruction based on deep learning techniques. Although image SR techniques show strong potential in image reconstruction, the effective application of these techniques to SR reconstruction of digital elevation models [...] Read more.
In recent years, great progress has been made in the field of super-resolution (SR) reconstruction based on deep learning techniques. Although image SR techniques show strong potential in image reconstruction, the effective application of these techniques to SR reconstruction of digital elevation models (DEMs) remains an important research challenge. The complexity and diversity of DEMs limits existing methods to capture subtle changes and features of the terrain, thus affecting the quality of reconstruction. To solve this problem, a DEM SR reconstruction network based on multi-path feature extraction and transformer feature enhancement is proposed in this paper. The network structure has three parts: feature extraction, image reconstruction, and feature enhancement. The feature extraction component consists of three feature extraction blocks, and each feature extraction block contains multiple multi-path feature residuals to enhance the interaction between spatial information and semantic information, so as to fully extract image features. In addition, the transformer feature enhancement module uses an encoder and decoder based design, leveraging the correlation between low- and high-dimensional features to further improve network performance. Through repeated testing and improvement, the model shows excellent performance in high-resolution DEM image reconstruction, and can generate more accurate DEMs. In terms of elevation and slope evaluation indexes, the model was 3.41% and 1.11% better compared with the existing reconstruction methods, which promotes the application of SR reconstruction technology in terrain data. Full article
Show Figures

Figure 1

29 pages, 2763 KiB  
Review
A Review of Computer Vision Technology for Football Videos
by Fucheng Zheng, Duaa Zuhair Al-Hamid, Peter Han Joo Chong, Cheng Yang and Xue Jun Li
Information 2025, 16(5), 355; https://doi.org/10.3390/info16050355 - 28 Apr 2025
Viewed by 1949
Abstract
In the era of digital advancement, the integration of Deep Learning (DL) algorithms is revolutionizing performance monitoring in football. Due to restrictions on monitoring devices during games to prevent unfair advantages, coaches are tasked to analyze players’ movements and performance visually. As a [...] Read more.
In the era of digital advancement, the integration of Deep Learning (DL) algorithms is revolutionizing performance monitoring in football. Due to restrictions on monitoring devices during games to prevent unfair advantages, coaches are tasked to analyze players’ movements and performance visually. As a result, Computer Vision (CV) technology has emerged as a vital non-contact tool for performance analysis, offering numerous opportunities to enhance the clarity, accuracy, and intelligence of sports event observations. However, existing CV studies in football face critical challenges, including low-resolution imagery of distant players and balls, severe occlusion in crowded scenes, motion blur during rapid movements, and the lack of large-scale annotated datasets tailored for dynamic football scenarios. This review paper fills this gap by comprehensively analyzing advancements in CV, particularly in four key areas: player/ball detection and tracking, motion prediction, tactical analysis, and event detection in football. By exploring these areas, this review offers valuable insights for future research on using CV technology to improve sports performance. Future directions should prioritize super-resolution techniques to enhance video quality and improve small-object detection performance, collaborative efforts to build diverse and richly annotated datasets, and the integration of contextual game information (e.g., score differentials and time remaining) to improve predictive models. The in-depth analysis of current State-Of-The-Art (SOTA) CV techniques provides researchers with a detailed reference to further develop robust and intelligent CV systems in football. Full article
(This article belongs to the Special Issue AI-Based Image Processing and Computer Vision)
Show Figures

Figure 1

26 pages, 20953 KiB  
Article
Optimization-Based Downscaling of Satellite-Derived Isotropic Broadband Albedo to High Resolution
by Niko Lukač, Domen Mongus and Marko Bizjak
Remote Sens. 2025, 17(8), 1366; https://doi.org/10.3390/rs17081366 - 11 Apr 2025
Viewed by 426
Abstract
In this paper, a novel method for estimating high-resolution isotropic broadband albedo is proposed, by downscaling satellite-derived albedo using an optimization approach. At first, broadband albedo is calculated from the lower-resolution multispectral satellite image using standard narrow-to-broadband (NTB) conversion, where the surfaces are [...] Read more.
In this paper, a novel method for estimating high-resolution isotropic broadband albedo is proposed, by downscaling satellite-derived albedo using an optimization approach. At first, broadband albedo is calculated from the lower-resolution multispectral satellite image using standard narrow-to-broadband (NTB) conversion, where the surfaces are considered Lambertian with isotropic reflectance. The high-resolution true orthophoto for the same location is segmented with the deep learning-based Segment Anything Model (SAM), and the resulting segments are refined with a classified digital surface model (cDSM) to exclude small transient objects. Afterwards, the remaining segments are grouped using K-means clustering, by considering orthophoto-visible (VIS) and near-infrared (NIR) bands. These segments present surfaces with similar materials and underlying reflectance properties. Next, the Differential Evolution (DE) optimization algorithm is applied to approximate albedo values to these segments so that their spatial aggregate matches the coarse-resolution satellite albedo, by proposing two novel objective functions. Extensive experiments considering different DE parameters over an 0.75 km2 large urban area in Maribor, Slovenia, have been carried out, where Sentinel-2 Level-2A NTB-derived albedo was downscaled to 1 m spatial resolution. Looking at the performed spatiospectral analysis, the proposed method achieved absolute differences of 0.09 per VIS band and below 0.18 per NIR band, in comparison to lower-resolution NTB-derived albedo. Moreover, the proposed method achieved a root mean square error (RMSE) of 0.0179 and a mean absolute percentage error (MAPE) of 4.0299% against ground truth broadband albedo annotations of characteristic materials in the given urban area. The proposed method outperformed the Enhanced Super-Resolution Generative Adversarial Networks (ESRGANs), which achieved an RMSE of 0.0285 and an MAPE of 9.2778%, and the Blind Super-Resolution Generative Adversarial Network (BSRGAN), which achieved an RMSE of 0.0341 and an MAPE of 12.3104%. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Graphical abstract

20 pages, 2203 KiB  
Article
PixelCraftSR: Efficient Super-Resolution with Multi-Agent Reinforcement for Edge Devices
by M. J. Aashik Rasool, Shabir Ahmed, S. M. A. Sharif, Mardieva Sevara and Taeg Keun Whangbo
Sensors 2025, 25(7), 2242; https://doi.org/10.3390/s25072242 - 2 Apr 2025
Viewed by 1027
Abstract
Single-image super-resolution imaging methods are increasingly being employed owing to their immense applicability in numerous domains, such as medical imaging, display manufacturing, and digital zooming. Despite their widespread usability, the existing learning-based super-resolution (SR) methods are computationally expensive and inefficient for resource-constrained IoT [...] Read more.
Single-image super-resolution imaging methods are increasingly being employed owing to their immense applicability in numerous domains, such as medical imaging, display manufacturing, and digital zooming. Despite their widespread usability, the existing learning-based super-resolution (SR) methods are computationally expensive and inefficient for resource-constrained IoT devices. In this study, we propose a lightweight model based on a multi-agent reinforcement-learning approach that employs multiple agents at the pixel level to construct super-resolution images by following the asynchronous actor–critic policy. The agents iteratively select a predefined set of actions to be executed within five time steps based on the new image state, followed by the action that maximizes the cumulative reward. We thoroughly evaluate and compare our proposed method with existing super-resolution methods. Experimental results illustrate that the proposed method can outperform the existing models in both qualitative and quantitative scores despite having significantly less computational complexity. The practicability of the proposed method is confirmed further by evaluating it on numerous IoT platforms, including edge devices. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

12 pages, 16440 KiB  
Communication
Resolution Enhancement of Geometric Phase Self-Interference Incoherent Digital Holography Using Synthetic Aperture
by Youngrok Kim, KiHong Choi, Keehoon Hong and Sung-Wook Min
Photonics 2024, 11(12), 1170; https://doi.org/10.3390/photonics11121170 - 12 Dec 2024
Viewed by 1211
Abstract
Incoherent digital holography is a promising solution for acquiring three-dimensional information in the form of interference without a laser. Self-interference is the principal phenomenon of incoherent holography which splits the incident light wave into two waves and interferes with each one. The characteristics [...] Read more.
Incoherent digital holography is a promising solution for acquiring three-dimensional information in the form of interference without a laser. Self-interference is the principal phenomenon of incoherent holography which splits the incident light wave into two waves and interferes with each one. The characteristics of incoherent holograms are defined by the optical configuration of wavefront dividing units. In this paper, we propose the synthetic aperture method to achieve super-resolution of geometric phase lens-based self-interference incoherent digital holography. We analyze the resolution of incoherent holograms acquired from geometric phase lenses and experimentally demonstrate the resolution enhancement via the synthetic-aperture method. Moreover, the sparse synthetic-aperture method is proposed to ensure effective data capturing and the numerical optimization method is also proposed to fill the gap between the empty space of sparsely acquired holograms. Full article
(This article belongs to the Special Issue Optical Imaging Innovations and Applications)
Show Figures

Figure 1

32 pages, 7202 KiB  
Article
Learn from Simulations, Adapt to Observations: Super-Resolution of Isoprene Emissions via Unpaired Domain Adaptation
by Antonio Giganti, Sara Mandelli, Paolo Bestagini and Stefano Tubaro
Remote Sens. 2024, 16(21), 3963; https://doi.org/10.3390/rs16213963 - 24 Oct 2024
Cited by 5 | Viewed by 9922
Abstract
Plants emit biogenic volatile organic compounds (BVOCs), such as isoprene, significantly influencing atmospheric chemistry and climate. BVOC emissions estimated from bottom-up (BU) approaches (derived from numerical simulations) usually exhibit denser and more detailed spatial information compared to those estimated through top-down (TD) approaches [...] Read more.
Plants emit biogenic volatile organic compounds (BVOCs), such as isoprene, significantly influencing atmospheric chemistry and climate. BVOC emissions estimated from bottom-up (BU) approaches (derived from numerical simulations) usually exhibit denser and more detailed spatial information compared to those estimated through top-down (TD) approaches (derived from satellite observations). Moreover, numerically simulated emissions are typically easier to obtain, even if they are less reliable than satellite acquisitions, which, being derived from actual measurements, are considered a more trustworthy instrument for performing chemistry and climate investigations. Given the coarseness and relative lack of satellite-derived measurements, fine-grained numerically simulated emissions could be exploited to enhance them. However, simulated (BU) and observed (TD) emissions usually differ regarding value range and spatiotemporal resolution. In this work, we present a novel deep learning (DL)-based approach to increase the spatial resolution of satellite-derived isoprene emissions, investigating the adoption of efficient domain adaptation (DA) techniques to bridge the gap between numerically simulated emissions and satellite-derived emissions, avoiding the need for retraining a specific super-resolution (SR) algorithm on them. For this, we propose a methodology based on the cycle generative adversarial network (CycleGAN) architecture, which has been extensively used for adapting natural images (like digital photographs) of different domains. In our work, we depart from the standard CycleGAN framework, proposing additional loss terms that allow for better DA and emissions’ SR. We extensively demonstrate the proposed method’s effectiveness and robustness in restoring fine-grained patterns of observed isoprene emissions. Moreover, we compare different setups and validate our approach using different emission inventories from both domains. Eventually, we show that the proposed DA strategy paves the way towards robust SR solutions even in the case of spatial resolution mismatch between the training and testing domains and in the case of unknown testing data. Full article
Show Figures

Figure 1

27 pages, 9255 KiB  
Article
DDMAFN: A Progressive Dual-Domain Super-Resolution Network for Digital Elevation Model Based on Multi-Scale Feature Fusion
by Bing He, Xuebing Ma, Bo Kong, Bingchao Wang and Xiaoxue Wang
Electronics 2024, 13(20), 4078; https://doi.org/10.3390/electronics13204078 - 17 Oct 2024
Cited by 2 | Viewed by 1407
Abstract
This paper examines the multi-scale super-resolution challenge of digital elevation models in remote sensing. A dual-domain multi-scale attention fusion network is proposed, which reconstructs digital elevation image details step-by-step using cascading sub-networks. This model incorporates components like the wavelet guidance and separation module, [...] Read more.
This paper examines the multi-scale super-resolution challenge of digital elevation models in remote sensing. A dual-domain multi-scale attention fusion network is proposed, which reconstructs digital elevation image details step-by-step using cascading sub-networks. This model incorporates components like the wavelet guidance and separation module, multi-scale attention fusion blocks, dilated convolutional inception module, and edge enhancement module to improve feature extraction and fusion capabilities. A new loss function is designed to enhance the model’s robustness and stability. Experiments indicate that the proposed model outperforms 15 benchmark models in PSNR, RMSE, MAE, RMSEslope, and RMSEaspect metrics. In HMA data, The proposed model’s PSNR increases by 0.89 dB (~1.81%), and RMSE decreases by 1.22 m (~8.6%) compared to a state-of-the-art model. Compared to EDEM, which has the best elevation index, RMSEslope decreases by 0.79° (~16%). Additionally, the effectiveness and contribution of each DDMAFN component were verified through ablation experiments. Finally, on the SRTM dataset, The proposed model demonstrates superior performance even with interpolated degradation. Full article
(This article belongs to the Special Issue Advances in AI Technology for Remote Sensing Image Processing)
Show Figures

Figure 1

16 pages, 6200 KiB  
Communication
A Transformer-Unet Generative Adversarial Network for the Super-Resolution Reconstruction of DEMs
by Xin Zheng, Zhaoqi Xu, Qian Yin, Zelun Bao, Zhirui Chen and Sizhu Wang
Remote Sens. 2024, 16(19), 3676; https://doi.org/10.3390/rs16193676 - 1 Oct 2024
Cited by 3 | Viewed by 1679
Abstract
A new model called the Transformer-Unet Generative Adversarial Network (TUGAN) is proposed for super-resolution reconstruction of digital elevation models (DEMs). Digital elevation models are used in many fields, including environmental science, geology and agriculture. The proposed model uses a self-similarity Transformer (SSTrans) as [...] Read more.
A new model called the Transformer-Unet Generative Adversarial Network (TUGAN) is proposed for super-resolution reconstruction of digital elevation models (DEMs). Digital elevation models are used in many fields, including environmental science, geology and agriculture. The proposed model uses a self-similarity Transformer (SSTrans) as the generator and U-Net as the discriminator. SSTrans, a model that we previously proposed, can yield good reconstruction results in structurally complex areas but has little advantage when the surface is simple and smooth because too many additional details have been added to the data. To resolve this issue, we propose the novel TUGAN model, where U-Net is capable of multilayer jump connections, which enables the discriminator to consider both global and local information when making judgments. The experiments show that TUGAN achieves state-of-the-art results for all types of terrain details. Full article
Show Figures

Figure 1

13 pages, 900 KiB  
Article
Single-Snapshot Direction of Arrival Estimation for Vehicle-Mounted Millimeter-Wave Radar via Fast Deterministic Maximum Likelihood Algorithm
by Hong Liu, Han Xie, Zhen Wang, Xianling Wang and Donghang Chai
World Electr. Veh. J. 2024, 15(7), 321; https://doi.org/10.3390/wevj15070321 - 20 Jul 2024
Viewed by 1444
Abstract
As one of the fundamental vehicular perception technologies, millimeter-wave radar’s accuracy in angle measurement affects the decision-making and control of vehicles. In order to enhance the accuracy and efficiency of the Direction of Arrival (DoA) estimation of radar systems, a super-resolution angle measurement [...] Read more.
As one of the fundamental vehicular perception technologies, millimeter-wave radar’s accuracy in angle measurement affects the decision-making and control of vehicles. In order to enhance the accuracy and efficiency of the Direction of Arrival (DoA) estimation of radar systems, a super-resolution angle measurement strategy based on the Fast Deterministic Maximum Likelihood (FDML) algorithm is proposed in this paper. This strategy sequentially uses Digital Beamforming (DBF) and Deterministic Maximum Likelihood (DML) in the Field of View (FoV) to perform a rough search and precise search, respectively. In a simulation with a signal-to-noise ratio of 20 dB, FDML can accurately determine the target angle in just 16.8 ms, with a positioning error of less than 0.7010. DBF, the Iterative Adaptive Approach (IAA), DML, Fast Iterative Adaptive Approach (FIAA), and FDML are subjected to simulation with two targets, and their performance is compared in this paper. The results demonstrate that under the same angular resolution, FDML reduces computation time by 99.30% and angle measurement error by 87.17% compared with the angular measurement results of two targets. The FDML algorithm significantly improves computational efficiency while ensuring measurement performance. It provides more reliable technical support for autonomous vehicles and lays a solid foundation for the advancement of autonomous driving technology. Full article
Show Figures

Figure 1

Back to TopTop