Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (283)

Search Parameters:
Keywords = super pixel

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 37608 KB  
Article
ZoomPatch: An Adaptive PTZ Scheduling Framework for Small Object Video Analytics
by Shutong Chen, Binhua Liang and Yan Chen
Appl. Sci. 2026, 16(6), 2934; https://doi.org/10.3390/app16062934 - 18 Mar 2026
Viewed by 30
Abstract
Accurate detection of small objects in video analytics is limited by low pixel resolution and insufficient visual cues. While software-based enhancements often fail to recover missing details, Pan–Tilt–Zoom (PTZ) cameras can physically increase spatial resolution through optical zoom. However, mechanical latency and configuration [...] Read more.
Accurate detection of small objects in video analytics is limited by low pixel resolution and insufficient visual cues. While software-based enhancements often fail to recover missing details, Pan–Tilt–Zoom (PTZ) cameras can physically increase spatial resolution through optical zoom. However, mechanical latency and configuration complexity hinder their real-time applicability. We propose ZoomPatch, a real-time video analytics framework tailored for small object detection. ZoomPatch actively schedules PTZ adjustments to capture optically enhanced subframes of regions of interest (ROIs) and fuses inference results back to the global reference frame. Specifically, it introduces a dynamic Cycle Length Proposer to adapt analysis cycles based on scene motion, and a Mixed Integer Linear Programming (MILP)-based Configuration Decider to determine the optimal sequence of pan, tilt, and zoom adjustments under time budget constraints. Simulation-based experimental evaluations across diverse workloads demonstrate that ZoomPatch significantly outperforms fixed-perspective, super-resolution (SR), and greedy baselines. Notably, in the detection task using YOLOv10, ZoomPatch improves the F1-score from 0.33 to 0.47 (a 42% increase) compared to the fixed-perspective baseline. Furthermore, ZoomPatch yields performance gains of 30% and 7% over the SR baseline (0.36) and the greedy baseline (0.44). Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

17 pages, 3717 KB  
Article
Improving Astrometric Precision with MLP-Driven Super-Resolution of Star Maps
by Yi Lu, Xiping Xu, Juncen Yan, Ning Zhang and Yaowen Lv
Sensors 2026, 26(6), 1769; https://doi.org/10.3390/s26061769 - 11 Mar 2026
Viewed by 243
Abstract
Aiming at the star centroid positioning error in dynamic star simulators, a super-resolution star map correction method is proposed based on a multi-layer perceptron (MLP). A complete technical chain of “system calibration–aberration field modeling–network correction” is constructed to establish a data-driven end-to-end framework [...] Read more.
Aiming at the star centroid positioning error in dynamic star simulators, a super-resolution star map correction method is proposed based on a multi-layer perceptron (MLP). A complete technical chain of “system calibration–aberration field modeling–network correction” is constructed to establish a data-driven end-to-end framework for unified modeling and compensation of optical aberrations, assembly deviations, and device discreteness. Experimental results show that the proposed method achieves sub-pixel accuracy: the maximum star centroid and inter-star angular distance errors are reduced by 22.9% and 37.5% on average, respectively, which is significantly superior to traditional methods. This work provides a reliable technical approach for high-precision star map display and star sensor ground calibration, with clear engineering application value. Full article
(This article belongs to the Special Issue Optical Sensors: Instrumentation, Measurement and Metrology)
Show Figures

Figure 1

21 pages, 5982 KB  
Article
Evaluating Geostationary Satellite-Based Approaches for NDVI Gap Filling in Polar-Orbiting Satellite Observations
by Han-Sol Ryu, Sung-Joo Yoon, Jinyeong Kim and Tae-Ho Kim
Sensors 2026, 26(5), 1731; https://doi.org/10.3390/s26051731 - 9 Mar 2026
Viewed by 276
Abstract
The Normalized Difference Vegetation Index (NDVI) derived from polar-orbiting satellites is widely used for vegetation monitoring; however, its temporal continuity is often limited by cloud contamination and fixed revisit cycles. To address this limitation, this study investigates the feasibility of using geostationary satellite [...] Read more.
The Normalized Difference Vegetation Index (NDVI) derived from polar-orbiting satellites is widely used for vegetation monitoring; however, its temporal continuity is often limited by cloud contamination and fixed revisit cycles. To address this limitation, this study investigates the feasibility of using geostationary satellite observations to enhance the spatial completeness of Sentinel-2 NDVI at its standard revisit intervals through cloud gap-filling applications. Geostationary Ocean Color Imager II (GOCI-II) data (250 m) was used as input, while Sentinel-2 Multispectral Instrument (MSI) NDVI (10 m) served as the reference dataset. To enable cross-sensor integration, a data-driven transformation framework was developed to convert GOCI-II NDVI into MSI-like NDVI while preserving dominant spatial variation patterns rather than pursuing strict pixel-level super-resolution. The transformed NDVI was assessed through spatial comparisons and statistical metrics, including correlation coefficient, mean absolute error, root mean square error (RMSE), normalized RMSE, and structural similarity index measure. Results show that geostationary-derived NDVI captures broad spatial organization and field-scale variability observed in MSI NDVI. Building on this cross-scale consistency, cloud gap-filling experiments demonstrate that temporally adjacent transformed NDVI scenes maintain consistent variation patterns, supporting their complementary use for compensating cloud-induced gaps. Although reduced contrast and magnitude-dependent biases remain, primarily due to the large spatial resolution difference and sub-pixel heterogeneity, an intermediate-resolution (80 m) sensitivity analysis indicates improved stability when the resolution gap is reduced. Overall, these findings highlight the practical potential of integrating geostationary and polar-orbiting observations to improve NDVI spatial continuity in cloud-prone regions. Full article
(This article belongs to the Special Issue Remote Sensing Technology for Agricultural and Land Management)
Show Figures

Figure 1

26 pages, 12237 KB  
Article
SAMCM-SR: Applying SAM3 Under Data-Scarce Conditions for Cross-Modal Segmentation of Power Equipment Infrared Images with Super-Resolution Enhancement
by Junchao Wang, Xiang Wu, Tianrui Yang, Yin Wang, Mengru Xiao and Gaoxing Zheng
Appl. Sci. 2026, 16(5), 2351; https://doi.org/10.3390/app16052351 - 28 Feb 2026
Viewed by 188
Abstract
Infrared thermography is a significant and extensively utilized method for assessing the operational condition of power equipment. Nonetheless, the constrained spatial resolution of infrared imaging systems, imaging noise, and the inadequate representational capacity of single-modality data render the precise segmentation of power equipment [...] Read more.
Infrared thermography is a significant and extensively utilized method for assessing the operational condition of power equipment. Nonetheless, the constrained spatial resolution of infrared imaging systems, imaging noise, and the inadequate representational capacity of single-modality data render the precise segmentation of power equipment targets difficult, particularly in intricate backdrops and settings with weak structures. Simultaneously, obtaining high-quality pixel-level annotations for power equipment is expensive and laborious, leading to a scarcity of training samples and thus diminishing the efficacy of conventional supervised segmentation techniques. This research offers a super-resolution guided cross-modal segmentation strategy to tackle these issues in data-scarce circumstances and examines the applicability of the general-purpose segmentation model Segment Anything Model 3 (SAM3) for infrared image segmentation of power equipment. A super-resolution reconstruction framework based on a high-order degradation model is built to enhance low-resolution infrared images collected in real-world contexts. An Enhanced Super-Resolution Generative Adversarial Networks (ESRGAN) -based network incorporating residual-in-residual dense blocks (RRDB) is utilized to reconstruct infrared thermograms, hence improving structural features and boundary representations. Secondly, the concurrently obtained visible-light images are improved by low-light enhancement methods, and an anchor-free object detection framework is employed to ensure accurate localization of power equipment targets. The identified areas in visible images are aligned with the coordinate system of infrared super-resolution images via cross-modal geometric transformation, establishing a cross-modal spatial prior that efficiently limits the search space for infrared segmentation and mitigates background interference. The general-purpose segmentation model SAM3 is introduced, utilizing cross-modal detection boxes as prompts to facilitate precise segmentation of power equipment targets in infrared super-resolution images, achieving high-accuracy segmentation without the necessity for extensive task-specific annotated data. The experimental results demonstrate that our proposed approach significantly improves both the accuracy and robustness of infrared image segmentation for power equipment under complex conditions, attaining a Jaccard index of 89.86% and a Dice coefficient of 91.12%, thereby validating its efficacy and practical applicability in data-scarce environments. Full article
Show Figures

Figure 1

12 pages, 6299 KB  
Communication
Lensless Quantitative Phase Imaging with Bayer-Filtered Color Sensors Under Sequential RGB-LED Illumination
by Jiajia Wu, Yining Li, Yuheng Luo, Leiting Pan, Pengming Song and Qiang Xu
J. Imaging 2026, 12(3), 101; https://doi.org/10.3390/jimaging12030101 - 26 Feb 2026
Viewed by 282
Abstract
Lensless on-chip microscopy enables high-throughput, wide-FOV imaging; however, the Bayer color filter array (CFA) in standard color sensors spatially multiplexes spectral channels, introducing sub-sampling and spectral crosstalk that degrade phase retrieval. We propose a Wirtinger Poly-Gradient Solver (WPGS) for quantitative phase reconstruction with [...] Read more.
Lensless on-chip microscopy enables high-throughput, wide-FOV imaging; however, the Bayer color filter array (CFA) in standard color sensors spatially multiplexes spectral channels, introducing sub-sampling and spectral crosstalk that degrade phase retrieval. We propose a Wirtinger Poly-Gradient Solver (WPGS) for quantitative phase reconstruction with Bayer-filtered color sensors under sequential Red–Green–Blue Light-Emitting Diode (RGB-LED) illumination. The method combines Transport of Intensity Equation (TIE)-based initialization with polychromatic Wirtinger optimization to suppress CFA-induced artifacts and enable pixel super-resolution (PSR). Experiments resolve a 2.76 μm linewidth using a 1.85 μm pixel-pitch sensor, exceeding the nominal Nyquist limit imposed by pixel sampling. We further demonstrate label-free imaging of HeLa cells and unstained tissue sections, supporting high-throughput digital pathology and offering potential for longitudinal biological observation. Full article
(This article belongs to the Section Computational Imaging and Computational Photography)
Show Figures

Figure 1

16 pages, 6965 KB  
Article
FISH-Dist: An Automated Pipeline for 3D Genomic Spatial Distance Quantification in FISH Imaging
by Benoit Aigouy, Emmanuelle Caturegli, Bernard Charroux, Carla Silva Martins, Thomas Gregor and Benjamin Prud’homme
Bioengineering 2026, 13(3), 268; https://doi.org/10.3390/bioengineering13030268 - 26 Feb 2026
Viewed by 397
Abstract
Accurate quantification of spatial distances between fluorescent signals in multi-channel 3D microscopy is essential for understanding genomic organization and gene regulation. However, chromatic aberration introduces systematic spatial offsets between channels that significantly bias distance measurements, particularly at short genomic distances. We present FISH-Dist, [...] Read more.
Accurate quantification of spatial distances between fluorescent signals in multi-channel 3D microscopy is essential for understanding genomic organization and gene regulation. However, chromatic aberration introduces systematic spatial offsets between channels that significantly bias distance measurements, particularly at short genomic distances. We present FISH-Dist, an automated computational pipeline for quantitative distance measurements in 3D fluorescence in situ hybridization (FISH) experiments acquired on standard confocal microscopes. Our method combines deep learning-based spot segmentation, 3D Gaussian fitting for sub-pixel localization, and two complementary chromatic aberration correction approaches: affine (ACC) and linear (LCC). We validated the pipeline by measuring the lengths of DNA origami nanorulers and systematically evaluated FISH probe design parameters, including probe spacing, density, and target sequence length. FISH-Dist achieves sub-pixel accuracy in signal detection and substantially reduces inter-channel distance measurement errors. This enables a reproducible quantification of spatial relationships in 3D FISH datasets. Unlike existing tools optimized for long-range chromosomal interactions or requiring super-resolution microscopy, FISH-Dist specifically addresses the technical challenges of standard confocal imaging at short genomic distances, where chromatic aberration has a proportionally greater impact on measurement accuracy. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Figure 1

28 pages, 11762 KB  
Article
A Coarse-to-Fine Optical-SAR Image Registration Algorithm for UAV-Based Multi-Sensor Systems Using Geographic Information Constraints and Cross-Modal Feature Consistency Mapping
by Xiaoyong Sun, Zhen Zuo, Xiaojun Guo, Xuan Li, Peida Zhou, Runze Guo and Shaojing Su
Remote Sens. 2026, 18(5), 683; https://doi.org/10.3390/rs18050683 - 25 Feb 2026
Viewed by 250
Abstract
Optical and synthetic aperture radar (SAR) image registration faces challenges from nonlinear radiometric distortions and geometric deformations caused by different imaging mechanisms. This paper proposes a coarse-to-fine registration algorithm integrating geographic information constraints with cross-modal feature consistency mapping. The coarse stage employs imaging [...] Read more.
Optical and synthetic aperture radar (SAR) image registration faces challenges from nonlinear radiometric distortions and geometric deformations caused by different imaging mechanisms. This paper proposes a coarse-to-fine registration algorithm integrating geographic information constraints with cross-modal feature consistency mapping. The coarse stage employs imaging geometry-based coordinate transformation with airborne navigation data to eliminate scale and rotation differences. The fine stage constructs a multi-scale phase congruency-based feature response aggregation model combined with rotation-invariant descriptors and global-to-local search for sub-pixel alignment. Experiments on integrated airborne optical/SAR datasets demonstrate superior performance with an average RMSE of 2.00 pixels, outperforming both traditional handcrafted methods (3MRS, OS-SIFT, POS-GIFT, GLS-MIFT) and state-of-the-art deep learning approaches (SuperGlue, LoFTR, ReDFeat, SAROptNet) while reducing execution time by 37.0% compared with the best-performing baseline. The proposed coarse registration also serves as an effective preprocessing module that improves SuperGlue’s matching rate by 167% and LoFTR’s by 109%, with a hybrid refinement strategy achieving 1.95 pixels RMSE. The method demonstrates robust performance under challenging conditions, enabling real-time UAV-based multi-sensor fusion applications. Full article
Show Figures

Figure 1

13 pages, 13581 KB  
Article
POEMMA–Balloon with Radio: A Balloon-Borne Multi- Messenger Multi-Detector Observatory
by Giuseppe Osteria, Johannes Eser and Angela Olinto
Particles 2026, 9(1), 19; https://doi.org/10.3390/particles9010019 - 16 Feb 2026
Viewed by 251
Abstract
The Probe Of Extreme Multi-Messenger Astrophysics (POEMMA) is a proposed dual-satellite mission to observe Ultra-High-Energy Cosmic Rays (UHECRs), increase the statistics at the highest energies, and observe Very-High-Energy Neutrinos (VHENs) following multi-messenger alerts of astrophysical transient events, such as gamma-ray bursts and gravitational [...] Read more.
The Probe Of Extreme Multi-Messenger Astrophysics (POEMMA) is a proposed dual-satellite mission to observe Ultra-High-Energy Cosmic Rays (UHECRs), increase the statistics at the highest energies, and observe Very-High-Energy Neutrinos (VHENs) following multi-messenger alerts of astrophysical transient events, such as gamma-ray bursts and gravitational wave events, throughout the universe. POEMMA–Balloon with radio (PBR) is a small-scale version of the POEMMA design, adapted to be flown as a payload on one of NASA’s suborbital Super Pressure Balloons (SPBs) circling over the Southern Ocean for more than 20 days after a launch from Wanaka, New Zealand. The main science objectives of PBR are: (1) to observe UHECRs via the fluorescence technique from suborbital space; (2) to observe horizontal high-altitude air showers (HAHAs) with energies above the cosmic ray knee (E > 3PeV) using optical and radio detection for the first time; and (3) to follow astrophysical event alerts in the search of VHENs. The PBR instrument consists of a 1.1 m aperture Schmidt telescope similar to the POEMMA design, with two cameras on its focal surface: a Fluorescence Camera (FC) and a Cherenkov Camera (CC). In addition, PBR has a Radio Instrument (RI) optimized for detecting EASs (covering the 60–660 Mhz range). The FC observes UHECR-induced EASs in the ultraviolet (UV) spectrum using an array of 9216-pixel Multi-Anode Photo-Multiplier Tubes (MAPMTs) imaged every 1 μs. The CC uses a 2048-pixel Silicon Photo-Multiplier (SiPM) imager to observe cosmic-ray-induced HAHAs and search for neutrino-induced upward-going EASs. The CC covers a spectral range of 320–900 nm, with an integration time of 10 ns. This contribution provides an overview of PBR instruments and their current status. Full article
Show Figures

Figure 1

21 pages, 5152 KB  
Article
Mapping Paddy Rice Using Segmentation Techniques and Phenological Metrics Derived from Sentinel-2 Time Series in Senegal
by Fama Mbengue, Mamadou Adama Sarr, Egor Prikaziuk, Gayane Faye, Mamadou Simina Dramé and Abdoul Aziz Diouf
Geomatics 2026, 6(1), 20; https://doi.org/10.3390/geomatics6010020 - 14 Feb 2026
Viewed by 316
Abstract
Rice field mapping is essential for effective agricultural and water resource management due to high land pressure. This study aims to map paddy rice by combining segmentation techniques and phenological metrics derived from optical time series. Thus, a crop segmentation-based approach was developed [...] Read more.
Rice field mapping is essential for effective agricultural and water resource management due to high land pressure. This study aims to map paddy rice by combining segmentation techniques and phenological metrics derived from optical time series. Thus, a crop segmentation-based approach was developed using Sentinel-2 imagery (2018–2019) to assess the paddy rice extent in the Senegal River Delta (SRD). Two super-pixel segmentation algorithms were evaluated to optimize the identification of rice plots by integrating spectral and spatial characteristics from the green, red, and near-infrared (NIR) bands. In this study, the Felzenszwalb outperformed the Quickshift algorithm, achieving a median intersection over union (IoU) of 0.25 compared to 0.20 for the segmentation of rice fields. The analysis of NDVI time series enabled the identification of key stages in the rice phenological cycle. Two machine learning algorithms (i.e., Random Forest and XGBoost) were compared for rice crop detection. Random Forest delivered a better performance (AUC = 0.93, OA = 0.98, F1-score = 0.98) than the XGBoost (AUC = 0.92, OA = 0.98, F1-score = 0.98). Overall, the results indicated that the approach could accurately identify paddy rice fields, and thus improve decision making and support food security management in the region. Full article
Show Figures

Figure 1

22 pages, 4890 KB  
Article
Super-Resolution Reconstruction and Detector Geometric Error Correction for Parallel-Beam Low-Resolution Multi-Detector SPECT: A Proof of Concept
by Zhibiao Cheng, Jun Zhang, Ping Chen and Junhai Wen
Tomography 2026, 12(2), 23; https://doi.org/10.3390/tomography12020023 - 12 Feb 2026
Viewed by 350
Abstract
Objectives: Due to collimator limitations, Single-Photon Emission Computed Tomography (SPECT) suffers from relatively low spatial resolution, which hampers the detection of small lesions. This study proposes a super-resolution (SR) reconstruction algorithm for a parallel-beam, low-resolution (LR) multi-detector SPECT system and employs a neural [...] Read more.
Objectives: Due to collimator limitations, Single-Photon Emission Computed Tomography (SPECT) suffers from relatively low spatial resolution, which hampers the detection of small lesions. This study proposes a super-resolution (SR) reconstruction algorithm for a parallel-beam, low-resolution (LR) multi-detector SPECT system and employs a neural network to estimate and correct for geometric errors in the LR detectors. Methods: A parallel-beam LR multi-detector SPECT system is presented, in which the detectors perform relative sub-pixel shifts. At each sampling angle, an SR reconstruction algorithm synthesizes high-resolution (HR) SPECT images from LR projections acquired by four offset LR detectors. To correct for geometric errors among these detectors, a randomly distributed gamma point source was designed to generate training data. A neural network was then employed to estimate the geometric errors, thereby refining the SR reconstruction. Results: Numerical simulation demonstrated that the proposed neural network could accurately identify the displacement-based geometric errors of the LR detectors. Utilizing these estimated parameters to correct the SR reconstruction process yielded results comparable to those obtained from direct reconstruction of HR projections, achieving a two-fold resolution improvement. Conclusions: Preliminary proof-of-principle for SR reconstruction in a parallel-beam LR multi-detector SPECT system was established. Further validation of the hardware performance is warranted. Full article
Show Figures

Figure 1

21 pages, 6229 KB  
Article
A Spatial–Spectral Decoupled Transformer Framework for Super-Resolution of Low-Earth-Orbit Multispectral Satellite Imagery
by Duhui Yun and Seok-Teak Yun
Appl. Sci. 2026, 16(4), 1674; https://doi.org/10.3390/app16041674 - 7 Feb 2026
Viewed by 270
Abstract
Multispectral (MS) satellite imagery provides rich spectral information for surface and atmospheric interpretation, yet its spatial resolution is often limited by sensor design. In this study, we propose a Transformer-based MS super-resolution framework that uses high-resolution panchromatic (PAN) imagery to supply complementary spatial [...] Read more.
Multispectral (MS) satellite imagery provides rich spectral information for surface and atmospheric interpretation, yet its spatial resolution is often limited by sensor design. In this study, we propose a Transformer-based MS super-resolution framework that uses high-resolution panchromatic (PAN) imagery to supply complementary spatial detail cues for MS reconstruction and explicitly separates spatial enhancement from spectral preservation. In the spatial branch, PAN features are aligned to the MS grid via Pixel-Unshuffle and encoded with shifted-window self-attention to capture long-range spatial dependencies efficiently. In the spectral branch, spectral self-attention treats bands as tokens to learn inter-band correlations and maintain spectral consistency. The two representations are fused through channel concatenation and a 1 × 1 convolutional module, followed by a reconstruction head that upsamples the fused features to generate high-resolution MS outputs. For training, low-resolution MS inputs are synthesized from KOMPSAT-3A MS imagery using a degradation pipeline that combines modulation transfer function-based blur, downsampling, and additive Gaussian noise; the operation order is randomly permuted to emulate diverse acquisition conditions. In addition, Bayesian optimization is employed to explore network configurations through jointly considering the normalized mean absolute error and inference time. Experiments demonstrate that the proposed approach attains 46.23 dB PSNR, 0.9735 SSIM, and 3.12 ERGAS with approximately 167.4 K parameters, achieving a high restoration quality and computational efficiency across diverse degradation settings. Full article
Show Figures

Figure 1

27 pages, 18987 KB  
Article
YOLO11s-UAV: An Advanced Algorithm for Small Object Detection in UAV Aerial Imagery
by Qi Mi, Jianshu Chao, Anqi Chen, Kaiyuan Zhang and Jiahua Lai
J. Imaging 2026, 12(2), 69; https://doi.org/10.3390/jimaging12020069 - 6 Feb 2026
Viewed by 700
Abstract
Unmanned aerial vehicles (UAVs) are now widely used in various applications, including agriculture, urban traffic management, and search and rescue operations. However, several challenges arise, including the small size of objects occupying only a sparse number of pixels in images, complex backgrounds in [...] Read more.
Unmanned aerial vehicles (UAVs) are now widely used in various applications, including agriculture, urban traffic management, and search and rescue operations. However, several challenges arise, including the small size of objects occupying only a sparse number of pixels in images, complex backgrounds in aerial footage, and limited computational resources onboard. To address these issues, this paper proposes an improved UAV-based small object detection algorithm, YOLO11s-UAV, specifically designed for aerial imagery. Firstly, we introduce a novel FPN, called Content-Aware Reassembly and Interaction Feature Pyramid Network (CARIFPN), which significantly enhances small object feature detection while reducing redundant network structures. Secondly, we apply a new downsampling convolution for small object feature extraction, called Space-to-Depth for Dilation-wise Residual Convolution (S2DResConv), in the model’s backbone. This module effectively eliminates information loss caused by strided convolution or pooling operations and facilitates the capture of multi-scale context. Finally, we integrate a simple, parameter-free attention module (SimAM) with C3k2 to form Flexible SimAM (FlexSimAM), which is applied throughout the entire model. This improved module not only reduces the model’s complexity but also enables efficient enhancement of small object features in complex scenarios. Experimental results demonstrate that on the VisDrone-DET2019 dataset, our model improves mAP@0.5 by 7.8% on the validation set (reaching 46.0%) and by 5.9% on the test set (increasing to 37.3%) compared to the baseline YOLO11s, while reducing model parameters by 55.3%. Similarly, it achieves a 7.2% improvement on the TinyPerson dataset and a 3.0% increase on UAVDT-DET. Deployment on the NVIDIA Jetson Orin NX SUPER platform shows that our model achieves 33 FPS, which is 21.4% lower than YOLO11s, confirming its feasibility for real-time onboard UAV applications. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

25 pages, 4607 KB  
Article
CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution
by Xia Li, Haicheng Sun and Tie-Qiang Li
Sensors 2026, 26(2), 738; https://doi.org/10.3390/s26020738 - 22 Jan 2026
Cited by 1 | Viewed by 381
Abstract
Magnetic resonance imaging (MRI) super-resolution (SR) enables high-resolution reconstruction from low-resolution acquisitions, reducing scan time and easing hardware demands. However, most deep learning-based SR models are large and computationally heavy, limiting deployment in clinical workstations, real-time pipelines, and resource-restricted platforms such as low-field [...] Read more.
Magnetic resonance imaging (MRI) super-resolution (SR) enables high-resolution reconstruction from low-resolution acquisitions, reducing scan time and easing hardware demands. However, most deep learning-based SR models are large and computationally heavy, limiting deployment in clinical workstations, real-time pipelines, and resource-restricted platforms such as low-field and portable MRI. We introduce CHARMS, a lightweight convolutional–Transformer hybrid with attention regularization optimized for MRI SR. CHARMS employs a Reverse Residual Attention Fusion backbone for hierarchical local feature extraction, Pixel–Channel and Enhanced Spatial Attention for fine-grained feature calibration, and a Multi-Depthwise Dilated Transformer Attention block for efficient long-range dependency modeling. Novel attention regularization suppresses redundant activations, stabilizes training, and enhances generalization across contrasts and field strengths. Across IXI, Human Connectome Project Young Adult, and paired 3T/7T datasets, CHARMS (~1.9M parameters; ~30 GFLOPs for 256 × 256) surpasses leading lightweight and hybrid baselines (EDSR, PAN, W2AMSN-S, and FMEN) by 0.1–0.6 dB PSNR and up to 1% SSIM at ×2/×4 upscaling, while reducing inference time ~40%. Cross-field fine-tuning yields 7T-like reconstructions from 3T inputs with ~6 dB PSNR and 0.12 SSIM gains over native 3T. With near-real-time performance (~11 ms/slice, ~1.6–1.9 s per 3D volume on RTX 4090), CHARMS offers a compelling fidelity–efficiency balance for clinical workflows, accelerated protocols, and portable MRI. Full article
(This article belongs to the Special Issue Sensing Technologies in Digital Radiology and Image Analysis)
Show Figures

Figure 1

29 pages, 6047 KB  
Article
Robust Multi-Resolution Satellite Image Registration Using Deep Feature Matching and Super Resolution Techniques
by Yungyo Im and Yangwon Lee
Appl. Sci. 2026, 16(2), 1113; https://doi.org/10.3390/app16021113 - 21 Jan 2026
Viewed by 494
Abstract
This study evaluates the effectiveness of integrating a Residual Shifting (ResShift)-based deep learning super-resolution (SR) technique with the Robust Dense Feature Matching (RoMa) algorithm for high-precision inter-satellite image registration. The key findings of this research are as follows: (1) Enhancement of Structural Details: [...] Read more.
This study evaluates the effectiveness of integrating a Residual Shifting (ResShift)-based deep learning super-resolution (SR) technique with the Robust Dense Feature Matching (RoMa) algorithm for high-precision inter-satellite image registration. The key findings of this research are as follows: (1) Enhancement of Structural Details: Quadrupling image resolution via the ResShift SR model significantly improved the distinctness of edges and corners, leading to superior feature matching performance compared to original resolution data. (2) Superiority of Dense Matching: The RoMa model consistently delivered overwhelming results, maintaining a minimum of 2300 correct matches (NCM) across all datasets, which substantially outperformed existing sparse matching models such as SuperPoint + LightGlue (SPLG) (minimum 177 NCM) and SuperPoint + SuperGlue (SPSG). (3) Seasonal Robustness: The proposed framework demonstrated exceptional stability, maintaining registration errors below 0.5 pixels even in challenging summer–winter image pairs affected by cloud cover and spectral variations. (4) Geospatial Reliability: Integration of SR-derived homography with RoMa achieved a significant reduction in geographic distance errors, confirming the robustness of the dense matching paradigm for multi-sensor and multi-temporal satellite data fusion. These findings validate that the synergy between diffusion-based SR and dense feature matching provides a robust technological foundation for autonomous, high-precision satellite image registration. Full article
(This article belongs to the Special Issue Applications of Deep and Machine Learning in Remote Sensing)
Show Figures

Figure 1

17 pages, 4767 KB  
Article
Adaptive Low-Resolution Combination Search for Reference-Independent Image Super-Resolution
by Ye Tian
Sensors 2026, 26(2), 725; https://doi.org/10.3390/s26020725 - 21 Jan 2026
Viewed by 205
Abstract
Accurately reconstructing high-resolution (HR) images remains challenging in scenarios where HR observations cannot be captured due to optical, hardware, or cost constraints. To address this limitation, we introduce an image super-resolution (SR) framework that reconstructs HR content solely from multiple low-resolution (LR) measurements, [...] Read more.
Accurately reconstructing high-resolution (HR) images remains challenging in scenarios where HR observations cannot be captured due to optical, hardware, or cost constraints. To address this limitation, we introduce an image super-resolution (SR) framework that reconstructs HR content solely from multiple low-resolution (LR) measurements, without relying on any HR reference images. The proposed method formulates a unified degradation model that describes how HR pixels contribute to LR observations under subpixel shifts and anisotropic downsampling. Based on this model, we develop an adaptive search algorithm capable of identifying the minimal and most informative combination of LR images required to equivalently represent the latent HR image. The selected LR images are then used to construct a solvable linear system whose solution directly yields the HR pixel values. Experiments conducted on the USAF 1951 resolution target demonstrate that the proposed approach improves Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity (SSIM) by 27.33% and 44.64%, respectively, achieving a resolvable spatial frequency of 228 line pairs per millimeter. In semiconductor chip inspection, PSNR and SSIM increase by 22.36% and 40.38%. These results verify that the proposed LR-combination-based strategy provides a physically interpretable and highly practical alternative for applications in which HR reference images cannot be obtained. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Back to TopTop