Next Article in Journal
Research on the Elastic Loss Characteristics of Acoustic Echoes from Underwater Corner Reflector
Previous Article in Journal
Transverse Electric Inverse Scattering of Conductors Using Artificial Intelligence
Previous Article in Special Issue
A System for Real-Time Detection of Abandoned Luggage
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PIABC: Point Spread Function Interpolative Aberration Correction

by
Chanhyeong Cho
,
Chanyoung Kim
and
Sanghoon Sull
*
School of Electrical Engineering, Korea University, 1 5Ga Anam-Dong, Seoul 02841, Republic of Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(12), 3773; https://doi.org/10.3390/s25123773
Submission received: 6 May 2025 / Revised: 4 June 2025 / Accepted: 9 June 2025 / Published: 17 June 2025
(This article belongs to the Special Issue Sensors for Pattern Recognition and Computer Vision)

Abstract

:
Image quality in high-resolution digital single-lens reflex (DSLR) systems is degraded by Complementary Metal-Oxide-Semiconductor (CMOS) sensor noise and optical imperfections. Sensor noise becomes pronounced under high-ISO (International Organization for Standardization) settings, while optical aberrations such as blur and chromatic fringing distort the signal. Optical and sensor-level noise are distinct and hard to separate, but prior studies suggest that improving optical fidelity can suppress or mask sensor noise. Upon this understanding, we introduce a framework that utilizes densely interpolated Point Spread Functions (PSFs) to recover high-fidelity images. The process begins by simulating Gaussian-based PSFs as pixel-wise chromatic and spatial distortions derived from real degraded images. These PSFs are then encoded into a latent space to enhance their features and used to generate refined PSFs via similarity-weighted interpolation at each target position. The interpolated PSFs are applied through Wiener filtering, followed by residual correction, to restore images with improved structural fidelity and perceptual quality. We compare our method—based on pixel-wise, physical correction, and densely interpolated PSF at pre-processing—with post-processing networks, including deformable convolutional neural networks (CNNs) that enhance image quality without modeling degradation. Evaluations on DIV2K and RealSR-V3 confirm that our strategy not only enhances structural restoration but also more effectively suppresses sensor-induced artifacts, demonstrating the benefit of explicit physical priors for perceptual fidelity.

1. Introduction

High-resolution image sensors now extend beyond digital single-lens reflex (DSLR) cameras to smartphones and drones. This resolution increase reveals optical aberrations and sensor noise, diminishing image quality. Lens imperfections, including spherical aberration, chromatic dispersion, and misalignment errors, combined with high-ISO (International Organization for Standardization) sensor noise, degrade image sharpness and color fidelity.
Image quality depends not only on optical characteristics but also on sensor-level noise. As demonstrated by Healey and Kondepudy (1994) [1], sensor noise exhibits strong signal dependency, where photon shot noise dominates in high-signal regions while read noise and dark current become significant in low-signal areas. Foi et al. (2008) [2] further established that sensor noise follows heteroscedastic Gaussian distributions with intensity-dependent parameters. This signal-dependent noise becomes particularly problematic when combined with optical aberrations, as areas with reduced light transmission due to aberrations experience lower signal-to-noise ratios. El Gamal and Eltoukhy (2005) [3] showed that Complementary Metal-Oxide-Semiconductor (CMOS) sensor quantum efficiency interacts directly with light distribution patterns affected by lens aberrations, suggesting that optical correction can improve sensor utilization efficiency.
The cascading effects of optical aberrations through modern image processing pipelines compound these issues. Zhang et al. (2011) [4] demonstrated that demosaicing algorithms struggle with optically degraded inputs, generating additional artifacts. Similarly, Tian and Chen (2012) [5] found that poor optical quality necessitates aggressive sharpening, introducing new artifacts. These cumulative degradations underscore the need for addressing optical aberrations as a foundational step in improving overall image quality.
Sensor noise and optical aberrations arise from different sources, yet our research confirms that Point Spread Function (PSF) correction produces cleaner outputs without explicit noise modeling. This aligns with Hasinoff et al. (2010) [6], who established the theoretical relationship between optical efficiency and signal-to-noise ratio in digital imaging systems. By correcting optical aberrations, we improve photon collection efficiency, indirectly enhancing the signal component of the signal-to-noise equation.
To recover the original scene, we must reverse this process using knowledge of the optical blur—often modeled as a PSF. A widely used solution is the Wiener filter, which assumes that the observed image is the result of convolving the original image with a PSF and adding noise. The Wiener filter aims to suppress noise while recovering sharp details, making it an optimal choice under this model. However, it relies on an accurate PSF estimate and moderate noise levels to perform well. To support this step, we first estimate a physically meaningful PSF map and then apply the Wiener filter to restore the image.
In optics, PSF describes how light spreads across sensors through lenses. While ideal systems maintain uniform PSFs, real lenses create position-dependent variations in PSF shape, size, and orientation. These variations resist exact mathematical formulation, necessitating approximation methods. Gaussian-based PSF models, though not an exact physical model, are widely used in computational imaging due to their computational simplicity and their practical approximation of local optical blur.
Pixel-level distortion simulation captures optical aberration changes that are difficult to capture in patch-based methods. Instead of directly modifying image pixels, and assuming uniform patch characteristics, we use parametric approximations of blur intensity, color variation, and orientation characteristics to generate physically simulated PSFs that reflect both chromatic and spatial variations based on adjacent structural features.
Since images are formed by convolution with spatially varying PSFs, direct PSF interpolation better preserves optical properties. This method compensates for color and focus errors and better preserves the structure of the image even in the presence of strong distortion.
In this paper, we propose an interpolation method using pixel-wise chromatic-spatial simulated PSFs, guided by feature similarities across locations. To simulate these PSFs, we apply a simple circular blur function with increasing standard deviation as the distance from the image center increases, simulating field-dependent blur. Chromatic shifts between red, green, and blue channels are also introduced according to lens curvature, reflecting axial dispersion. These simulated PSFs preserve typical optical distortions found in consumer lenses and are used to guide interpolation at each location.
After generating a PSF at each target location, we use a deep learning method to generate a PSF that better reflects local features in a high-dimensional latent space, measure how similar it is to four adjacent PSFs at the required point, and combine them into a new interpolated PSF, weighted by similarity. This results in a PSF that better reflects local features and is more accurate than using basic interpolation alone. Interpolation by the generated PSFs more accurately captures optical aberrations than conventional patch-based or region-based blur models and maintains physical consistency better than pixel-level reconstruction methods. In this approach, the process precisely estimates the PSF characteristics at each target location.
Previous studies such as Mildenhall et al. (2018) [7] and Sitzmann et al. (2018) [8] explored joint optimization of optical and digital components, but these methods required specific models for each lens configuration. In contrast, our framework requires no neural network retraining for different lenses, achieving adaptability through PSF adjustment alone. Similarly, Li et al. (2021) [9] and Chen et al. (2021) [10] proposed methods relying on pre-computed PSFs—using calibration or simulation—achieving high performance, yet they remained constrained by lens-specific setups and computational complexity. In contrast, our framework requires no neural network retraining for different lenses, achieving adaptability through PSF adjustment alone. This accommodates optical system variations across lens types and wavelength ranges, addressing a key limitation in existing approaches.
We validate our approach using the DIV2K synthetic dataset and RealSR Canon2-LR 2× real-world dataset. These test environments isolate optical correction effects and demonstrate performance under realistic conditions. For benchmark comparison, we adopted previous methods [9,10] adapted for our setting, where explicit lens models or optical prescriptions are unavailable. Compared to Deformable ResU-Net (Chen et al. (2021) [10]), which focuses on optical aberration correction in postprocessing, our preprocessing-based approach showed superior performance. Quantitative results demonstrate +2.7 dB PSNR, +0.039 SSIM, and −0.0535 LPIPS improvements on DIV2K and +1.1 dB PSNR, +0.019 SSIM, and −0.0428 LPIPS on RealSR, confirming that our optical correction approach enhances both signal quality and perceptual image fidelity.
As a result, in addition to optical distortion, it improves control over a variety of factors, including photon shot noise, contrast loss due to stray light, dynamic range suppression, and sharpening artifacts, which are advantageous for sensor-level signal preservation in optically degraded environments.
Our proposed PSF Interpolation Aberration Correction (PIABC) framework comprises:
  • Dense PSF representation interpolation using a shallow transformer;
  • Wiener filtering with the interpolated PSFs;
  • Spatial signal refinement through a deep residual U-Net.
Our transformer architecture features:
  • Dense PSF estimation through pixel-level PSF embedding;
  • Autoencoder (AE) latent space PSF interpolation using learned representations;
  • Cross-attention-based PSF refinement between patches;
  • Long-range dependency capture ensuring global correction coherence.
The PIABC framework demonstrates consistent performance across optical systems, enabling accurate image restoration under complex aberrations through adaptive pixel-level PSF interpolation. Unlike previous approaches that address optical aberrations in isolation from sensor characteristics, our method implicitly improves overall signal quality, thereby reducing the impact of sensor noise without direct noise modeling.

2. Relative Works

Recent developments in computational imaging have produced a wide range of aberration correction strategies. We classify and summarize the most relevant approaches below.

2.1. Interrelation Between Optical Aberration Correction and System-Level Artifacts

Optical aberrations not only reduce the sharpness and color accuracy of raw images but also introduce challenges for downstream image processing. Studies have shown that algorithms such as demosaicing, denoising, and sharpening are sensitive to optical quality, and their performance deteriorates when the input is degraded by aberrations. Zhang et al. (2011) [11] highlighted that demosaicing under aberrated conditions amplifies color artifacts, while Tian and Chen (2012) [5] observed that aggressive sharpening in poorly focused images introduces ringing and aliasing. These effects accumulate and lower overall image quality, especially in compact devices where processing pipelines are highly compressed.
Heide et al. (2014) [12] proposed a unified processing pipeline, FlexISP, to jointly model optical and digital components, demonstrating that optimizing the optics-processing combination yields superior results. Schuler et al. (2016) [13] and Shih et al. (2012) [14] further emphasized that poor optical inputs constrain the effectiveness of image restoration algorithms. These findings suggest that addressing aberrations at the source not only improves native image quality but also mitigates secondary distortions introduced during enhancement steps.

2.2. Correction of Sensor-Induced Degradation and Optical Aberrations

Sensor noise and optical aberrations are the two principal sources of image degradation in modern digital imaging systems. Electronic noise—including thermal, read, and quantization noise—interacts with lens-induced blur, producing spatially heterogeneous distortions across the sensor plane. This interaction becomes more pronounced in low-light scenes or with high-ISO (International Organization for Standardization) settings (Kondepudy, 1994 [1]; Foi et al., 2008 [2]). As CMOS (complementary metal-oxide-semiconductor) quantum efficiency and photon collection are directly affected by optical transmission (El Gamal and Eltoukhy, 2005 [3]), mitigating lens aberrations improves sensor utilization efficiency and reduces signal-dependent noise effects.
These insights motivate the integration of optical modeling into image correction frameworks, especially in mobile or low-cost sensors where hardware limitations amplify these interactions. In this context, we adopt a simulation-based approach that approximates chromatic and spatial blur using parametric Gaussian point spread functions (PSFs). Our model uses radial functions in polar coordinates to replicate the smooth variation of blur and color shifts, reflecting common patterns in consumer-grade optics. This method avoids the need for lens-specific calibration and remains compatible with real-world degraded input.
The simulated PSFs are embedded in a latent space via a lightweight autoencoder (AE), enabling structured and continuous interpolation. These PSFs are then applied through Wiener filtering, followed by residual correction using residual U-Net. By modeling degradation before correction, our approach addresses both optical and sensor-level artifacts at the source, improving structural fidelity and perceptual quality across diverse conditions.

2.3. Comparison with Existing PSF Modeling Approaches

Aberration correction methods can be broadly categorized into analytical PSF modeling based on physical optics, patch-based statistical correction, and learning-based interpolation frameworks.
Classical approaches derive PSFs using Zernike or Seidel expansions under ideal or calibrated wavefront assumptions (Goodman, 2017 [15]). While interpretable, they require precise calibration and are typically limited to axis-symmetric distortions. Kee et al. (2011) [16], for instance, proposed a parametric blur model fitted from calibration images, expressing PSF variation as a function of spatial coordinates and lens parameters. Although accurate, such methods require dense, lens-specific calibration and generalize poorly to mobile or unknown systems. Wavefront-based approaches often employ Zernike polynomials to represent wavefront errors (Stallinga, 2010 [17]; Siddik et al., 2023 [18]), providing a physically accurate description but at the cost of high computational complexity and calibration requirements.
In contrast, PIABC approximates spatial and chromatic PSF variation using radial functions in polar coordinates, avoiding calibration while maintaining interpretability.
Patch-based methods, such as non-blind deconvolution (Zhang, 2019 [19]) or kernel prediction (Mildenhall et al., 2018 [7]), estimate local blur from image statistics but ignore global optical structure, often treating patches independently.
Recent learning-based approaches embed PSFs into latent spaces. Jing et al. (2021) [20] employed rank-minimizing autoencoders for structured blur representation, and Chen et al. (2021) [10] used raytraced wavefronts to simulate spatially varying PSFs with chromatic shifts, achieving 44.77 PSNR on DSLR data. Monakhova et al. (2020) [21] applied deep models with simulated PSFs for tomographic reconstruction but were limited to static setups.
Gaussian-based PSF models are widely used in computational imaging frameworks due to their computational simplicity and tractability (Chouzenoux et al., 2019 [22]). While efficient, Gaussian-based PSFs may not fully capture higher-order off-axis aberrations or complex wavefront structures present in real lenses.
PIABC initially employs isotropic Gaussian PSFs with radially varying blur and chromatic displacement, capturing realistic RGB shift patterns (e.g., r < g < b) observed in consumer optics. This isotropic baseline is chosen for its computational simplicity before introducing anisotropic Gaussian PSFs at the interpolation stage. These are interpolated through a transformer-based attention mechanism and refined by residual learning. Unlike prior methods, PIABC models PSF at the pixel level without calibration, bridging interpretable physics and scalable restoration while remaining robust to sensor-level noise variations.

3. Methods

We conceptually illustrate the overall framework in Figure 1, highlighting that our transformer-based interpolation module uses only the autoencoder (AE) embedding of sub-patch point spread functions (PSFs) as input. Notably, image sub-patches are not used in the transformer but are directly processed via Wiener filtering. This preserves the modular design and allows the transformer to focus exclusively on modeling spatially varying PSFs without interference from image content. However, this design choice also limits the transformer’s ability to capture interactions between the PSF and image content.
The proposed model is described in the following stages, and the overall framework is illustrated in Figure 2:
  • Formulation;
  • Generating chromatic-spatial simulated PSF (CS-sim PSF);
  • PSF interpolation on latent space using AE and vision transformer (ViT);
  • Patch-wise image restoration using a Wiener filter;
  • Patch assembling and full-image compilation;
  • Deep residual correction using a residual U-Net (Res U-Net);
  • Training objects and loss functions.

3.1. Formulation

The primary challenge in aberration correction is estimating the PSF at every point in the lens field—an expensive process if done via direct physical computation. As discussed in Section 2.3, autoencoders can learn a latent space that supports interpolation between measured or simulated PSFs (Jia et al. [23]), reducing reliance on system-specific calibration. Inspired by that data-driven perspective, we propose not to compute the exact physical function at position p but rather to approximate it with an encoder–decoder framework, refined through interpolation. We denote vectors in boldface and scalars in regular font. This convention is applied consistently throughout all interpolation and PSF modeling steps.
Under this assumption system (Schuler et al. [24]), aberration correction can be formulated as an inverse problem, where the goal is to recover the original image x · from its degraded observation y p at position p , considering contributions from other pixel position p P that influence the observed degradation:
y p , c = p P x p , c h p , c p + η p , c
where h p , c p is transfer function at pixel position p to the other pixel position p , c denotes channel or wavelength, and represents the convolution operator. The key challenge is that blur kernel h is typically unknown, making the problem ill-posed. Non-blind methods assume that h is known and applies it with deconvolution techniques to estimate x , whereas blind methods jointly estimate both h and x directly during the reconstruction process (Levin et al. [25], Yue et al. [26], Li et al. [27]). Let H be the true optical PSF, and f a b e r ( p ^ ) the estimated PSF at the interpolated location p ^ of observed locations p j * j = 1 m , using interpolation weights α j :
p ^ = j = 1 n α j · p j * , where   j = 1 n α j = 1 , α j R
Given this formulation, we approximate the H ( p ^ ) as
H ( p ^ ) f a b e r p ^ ; p j * j = 1 m , φ , c
= f d e c f e n c H ~ φ p ^ ; p j * j = 1 m , c , φ
f d e c f ^ e n c p ^ , f e n c H ~ φ p j * , c j = 1 m
= f d e c j = 1 m f s i m p j * , p ^ ; φ , c · f e n c H ~ φ p j * , c
where m is the number of nearest observed positions; φ is the aberration such as distortion, chromatic, coma, etc.; H ~ φ · is the initial CS-sim PSF for transformer input (Section 3.2); f e n c · is the encoder function; f d e c · is the decoder function; and f s i m · is theposition similarity function.
Exploiting the natural smoothness of PSF variations between nearby spatial positions, we approximate unmeasured PSFs through weighted interpolation in the autoencoder’s latent space, avoiding the complexity of exact physical wavefront modeling.
The first approximation of Equation (3a) indicates that the proposed PSF depends on PSF patch interpolation, wavelength, and aberration as functions. Equation (3b) replaces direct computation of the physical PSF, including aberration φ at position p , with an encoder–decoder model embedded in latent space. This approximation ensures that the model output aligns with the actual PSF characteristics. The second approximation in Equation (3c) assumes interpolation within the latent space. While the encoder and decoder contain non-linear activation functions and attention weights, PSF variations between adjacent pixels can still be approximated using weighted sums in the decoder in (3d). We incorporate Equations (3b)–(3d) with a transformer to manage the positional relationships and patch similarities, yielding a flexible architecture capable of capturing complex aberrations (including chromatic and rotational factors, as outlined in Section 3.2. Subsequent sections detail how these interpolated PSFs feed into our Wiener filter (Section 3.4.1) and eventually a residual U-Net for high-frequency restoration (Section 3.4.3).

3.2. Generating Chromatic-Spatail Simulated PSF (CS-sim PSF)

In this section, we formulate a pixel-wise blur simulation that approximates both geometric and chromatic aberration effects observed in lens-based imaging systems. Rather than relying on physical calibration or optical measurements, we analytically construct spatially varying PSFs using parameterized Gaussian kernels. We refer to the Gaussian-based PSF model in Chouzenoux et al. (2019) [22] as a basis and apply it with pixel-wise chromatic and spatial variation in our transformer-based interpolation framework.
For each pixel p with center coordinate ( o x , o y ), we define the reference isotropic Gaussian kernel (representing baseline blur) as:
G o u t , c x , y ; x 0 , y 0 , σ o u t , c = 1 2 π σ o u t , c 2 exp x x 0 2 + y y 0 2 2 π σ o u t , c 2 ,
where G o u t , c R H × W , c R , G , B is the channel; σ o u t , c denotes the outer (global) deviations, respectively; and ( x 0 ,   y 0 ) is the center coordinate.
To introduce anisotropy and directional blur, we define the elliptical Gaussian kernel with chromatic and spatial distortion:
G i n , c x , y ; x 0 , y 0 , σ i n , c , τ c , ρ c = 1 2 π σ i n , c 2 exp Δ T Σ i n , c Δ ,
Σ i n , c = R θ · ρ c σ i n , c 0 0 ρ c 1 σ i n , c · R θ T ,   θ = a r c t a n y x ,
Δ = x x 0 c T ,   c R , G , B
R = k R · f d · n x , y , G = k G · f d · n x , y , B = k B · f d · n x , y ,
where Δ R ,   Δ G , Δ B R , are distances of pixels due to chromatic aberration, k R ,   k G , k B are coefficients of chromatic aberration, and f d is a distance-based transformation function. n x , y denotes the normalized vector from the optical center to location ( x , y ) . f d and n x , y are defined respectively as:
f d = d i s t d m a x γ
n x , y = x x 0 , y y 0 T x x 0 , y y 0 ,
where d i s t = x x 0 2 + y y 0 2 represents the Euclidean distance from the center, d m a x Maximum distance, and γ is the decay factor. If we denote the variable blur due to independent changes in position per channel by index naming dynamic to the Gaussian standard deviation, the Gaussian kernel can be rewritten as:
P S F R G B x , y ; x 0 , y 0 , Δ R , Δ G , Δ B = G x x 0 Δ R x , y y 0 Δ R y ; Σ R G x x 0 Δ G x , y y 0 Δ G y ; Σ G G x x 0 Δ B x , y y 0 Δ B y ; Σ B
where P S F R G B ( · , ·   ; x 0 , y 0 ) R H × W × 3 , Σ c is the standard deviation of each channel. To reflect the fact that the amount of blur increases with distance from the lens center, the Gaussian standard deviation is redefined as follows:
Σ f i n a l ,   c x , y = σ o u t , c Σ i n , c · 1 + α c f d ,
where Σ f i n a l , c R H × W × ( 2 × 2 ) , α c is distance-sensitive modulation coefficient.
Finally, by integrating the elements, the initial PSF originated from ( x 0 , y 0 ) is formulated as:
H ~ φ p ^ i , c = P S F i n i t , c x , y = W c · G x x 0 Δ c x , y y 0 Δ c y ; Σ f i n a l ,   c ,
where P S F i n i t , c R H × W × ( H × W ) denotes the initial PSF generated through a chromatic-spatial PSF simulation process. W c is the weight that modulates the contribution of each color channel ( R , G , B ) to the final aberration correction, and G is the spatially varying Gaussian function with chromatic displacement Δ c .
This process enables pixel-specific variation of the PSF, reflecting how aberration differs at each location and across color channels. It serves as a simplified yet physically interpretable initialization for learning-based PSF interpolation.

3.3. Spatially Varying PSF Interpolation for Noise Reduction Using Deep Learning

3.3.1. Query Patch Configuration for Spatially Varying PSF Interpolation

To support PSF interpolation across spatially varying regions, we design two types of query configurations based on patch size. For 256 × 256 patches, five interpolation points are used between patch centers (see Figure 3a); for 128 × 128 patches, a single central point is used (Figure 3b). These key query points determine the spatial density of interpolated PSFs, enabling locally adaptive aberration correction. A trade-off exists between interpolation precision and computational cost, which we evaluate in Section 4.
Figure 3. Query configuration for interpolation in Cartesian coordinate system: (a) Five query positions t i 1 ~ t i 5 are placed at the midpoints between neighboring PSF centers in a 256 × 256 patch, which may cause overlapping windows, depending on query settings; this overlap case is illustrated here for clarity; (b) A single query point t i is at the center of a 128 × 128 patch; Subpatch centers and queries are marked by blue and red dots, respectively, and interpolation windows by red boxes.
Figure 3. Query configuration for interpolation in Cartesian coordinate system: (a) Five query positions t i 1 ~ t i 5 are placed at the midpoints between neighboring PSF centers in a 256 × 256 patch, which may cause overlapping windows, depending on query settings; this overlap case is illustrated here for clarity; (b) A single query point t i is at the center of a 128 × 128 patch; Subpatch centers and queries are marked by blue and red dots, respectively, and interpolation windows by red boxes.
Sensors 25 03773 g003

3.3.2. PSF Patch Embedding with Autoencoder

To embed each CS-sim PSF patch into a structured and compact latent space, we employ a vanilla autoencoder (AE). Each PSF patch h ~ j p is subdivided into sub-patches indexed by j { a , b , c , d } . For each sub-patch, a feature embedding:
h ˘ i j s p = P a t c h E m b e d d i n g ( h ~ i j s p ) R 768
which flattens the 33 × 33 × 3 sub-patch into a 3267-dimensional vector and then projects it to 768 dimensions.
This embedding preserves the essential PSF structure while enabling subsequent transformer-based interpolation. During AE training, Gaussian noise with σ = max V a r ( h ˘ i j s p ,   5 × 10 2 is added to each embedding to yield a denoising effect.
The decoder head is implemented as a multi-layer perceptron (MLP) consisting of three fully connected layers (768 → 1536 (ReLU) → 1536 (ReLU) → 3267) with intermediate ReLU activations. Specifically, the architecture first expands the latent feature from 768 to 1536, applies ReLU activation, then maintains this hidden dimension through another 1536-unit layer with ReLU activation, and finally projects the feature vector into the target PSF dimension (e.g., 33 × 33 × 3). After decoding, a softplus activation ensures non-negativity, followed by normalization so that the output PSF sums to one, consistent with physical PSF energy constraints. Note: For clarity, the sub-patch index s p is omitted in Figure 2 but is explicitly included hereafter for consistency.

3.3.3. Transformer-Based PSF Interpolation for Enhanced Accuracy

We adopt a vision transformer (ViT) architecture consisting of an encoder and a decoder to perform spatial interpolation of PSFs while maintaining physical consistency.
Unlike typical ViTs that use image patches as inputs, our method feeds only the AE-embedded PSF sub-patches to the encoder, focusing the transformer solely on PSF modeling. Each encoder feature vector is constructed by combining the AE embedding of a sub-patch PSF with local and global position encodings. (see Figure 4 and Table 1)
  • Encoder: Applies self-attention to aggregated PSF embeddings with positional encodings. Each sub-patch PSF embedding h ˘ i j s p is combined with local position embedding P E m j s p and global position embedding G l o b a l P E c i . The encoder input query is then formulated as:
    Q j e n c = h ˘ i j s p + L i n e a r P r o j e c t i o n ( L o c a l   P E m j s p + G l o b a l P E ( c i )
    where j indexes sub-patches within the PSF patch i . Although layer normalization (LN) is not explicitly shown here, it is applied internally within the transformer encoder block. This step preserves both local and global spatial context in the encoder tokens. m j s p R 2 denotes the center coordinates of each sub-patch.
  • Decoder: For each target query position t i l R 2 , the decoder query vector vector q i l d e c is computed as:
    q i , l d e c = L i n e a r P r o j e c t i o n t i l + G l o b a l P E ( c i ) ,
    This decoder query interacts with the encoder outputs via a cross-attention block, producing attention weights γ i j and aggregated latent features h ˘ i j s p :
    γ i j = s o f t m a x q i , l d e c · k j T d k
    where k j represents the encoder key vector, and d k is the key dimension. This mechanism captures the contribution of each sub-patch embedding to the interpolation target.
  • Position Similarity: Position similarity between the query position and each sub-patch center is defined as:
    α i j = s o f t m a x ( t i , l m i j s p 2 ) ,
    where j { a , b , c , d } indexes the neighboring sub-patches. This term quantifies the geometric proximity between the interpolation target and each sub-patch center.
  • Latent Aggregation: The position similarity and attention weights are combined to produce the weighted latent feature:
    z i j = α i j · γ i j ,
  • Interpolation Step: The final interpolated PSF at the query position t i l is then obtained as:
    h ^ i t l =   j β j · z i j ,
    where β j denotes the bilinear interpolation weight derived from the relative distances between the target position t i l and the sub-patch centers m i j s p :
    β j = d { x , y } ( 1 t i d m i j d ) ,
    Finally, a linear projection is applied to z i to generate the interpolated PSF H ^ :
    H ^ i = L i n e a r ( z i )
This step aggregates sub-patch contributions into a single PSF at location t i l . This formulation decouples spatial attention (via γ i j ) from position similarity (via α i j ), allowing stable and structured interpolation without relying on raw coordinate distances.
Figure 4. Overview of the proposed PSF-aware transformer architecture: (a) Encoder: Each PSF patch is linearly embedded and combined with positional encodings (local and global) to form attention tokens. Attention is computed across sub-patches to capture spatial relationships. AE: Autoencoder. (b) Decoder: Queries at target positions are projected into the embedding space and attend to encoder features. In the decoder, the query at position i attends to encoder sub-patches at positions j , and the attention weights γ i j determine their relative contributions to the aggregated latent feature z i . Attention-weighted aggregation yields latent PSF representations, which are linearly transformed to generate spatially interpolated PSFs. Query configuration follows Figure 3a. Sub-patch index ‘sp’ is omitted in Figure 2 for clarity but included here for detail. Layer Normalization (LN) is applied internally within the encoder. Decoder queries do not directly incorporate Local PE; instead, Global PE is indirectly propagated via attention; Case (a) in Figure 2 is used to illustrate the configuration. The bluish shaded blocks in the upper-right indicate subpatch-wise contributions to the interpolated result, based on the composite weight α i j · γ i j where α i j denotes position similarity and γ i j is the attention weight learned by the transformer; Relevant positional encodings and notations are indicated by shade intensity in the figure and fully defined in Table 1.
Figure 4. Overview of the proposed PSF-aware transformer architecture: (a) Encoder: Each PSF patch is linearly embedded and combined with positional encodings (local and global) to form attention tokens. Attention is computed across sub-patches to capture spatial relationships. AE: Autoencoder. (b) Decoder: Queries at target positions are projected into the embedding space and attend to encoder features. In the decoder, the query at position i attends to encoder sub-patches at positions j , and the attention weights γ i j determine their relative contributions to the aggregated latent feature z i . Attention-weighted aggregation yields latent PSF representations, which are linearly transformed to generate spatially interpolated PSFs. Query configuration follows Figure 3a. Sub-patch index ‘sp’ is omitted in Figure 2 for clarity but included here for detail. Layer Normalization (LN) is applied internally within the encoder. Decoder queries do not directly incorporate Local PE; instead, Global PE is indirectly propagated via attention; Case (a) in Figure 2 is used to illustrate the configuration. The bluish shaded blocks in the upper-right indicate subpatch-wise contributions to the interpolated result, based on the composite weight α i j · γ i j where α i j denotes position similarity and γ i j is the attention weight learned by the transformer; Relevant positional encodings and notations are indicated by shade intensity in the figure and fully defined in Table 1.
Sensors 25 03773 g004
Table 1 summarizes the stages and notations involved in the transformer-based PSF interpolation pipeline. Importantly, at the interpolation stage, the query index l is aggregated through weighted sums into the final position index i (see Figure 2). This notation aligns with Table 1 and ensures seamless integration between encoder–decoder outputs.

3.4. Image Restoration Using Image Restoration Using Wiener Filtering and Deep Correction

For high-resolution image restoration, we employ a Wiener filter for frequency-domain restoration, followed by patch assembly and compilation. This integrated approach, which will be depicted in Figure 5, ensures the effective enhancement of fine details. As will be discussed in Section 3.4.3, ResU-Net [28] further improves high-frequency detail recovery.

3.4.1. Patch Image Restoration Using Wiener Filter

While the Wiener filter is widely used for frequency-domain restoration, it often introduces ringing artifacts near sharp edges. To address this, we employ PSFs interpolated by the transformer model, which incorporate spatial variations and suppress artifacts during high-frequency restoration. The Wiener filter-based image restoration process for a patch k at location p is formulated as follows:
X ^ k u = H k , c * u H k , c u 2 + K k , c · Y k u ,
where Y k u is the observed patch, H k , c u is the PSF in the frequency domain, K k , c is a constant reflecting signal-to-noise characteristics, and H k , c * is the complex conjugate frequency representation of the interpolated PSF. The result X ^ k u is transformed back to the spatial domain as:
x k r e c u = F 1 X ^ k u
This approach allows local variations in PSFs to be incorporated into the restoration process, reducing artifacts compared to global Wiener filtering. Consequently,
x ^ k r e c = F 1 H ^ i * H ^ i 2 + K F P k H ^ ,
x ~ k r e c = F 1 H ~ i * H ~ i 2 + K F P k H ~ ,
where, P k H ^ and P k H ~ represent the patch domains used for each Wiener filtering step, combining the main support region P a ,   b ,   c ,   d and the corresponding PSF query positions P H ~ . The terms x ^ k r e c and x ~ k r e c denote the restored patches using PSFs H ^ i and H ~ i respectively, where H ^ i * and H ~ i * are conjugate frequency-domain representations. F and F 1 denote the Fourier transform and its inverse, respectively, and K is the noise control factor.
In Equations (23)–(25b), the scalar parameter K represents the ratio between the noise and signal power spectral density, i.e., K = σ n o i s e 2 / σ s i g n a l 2 which controls the trade-off between deblurring and noise amplification in Wiener filtering. A smaller K preserves fine details at the risk of amplifying noise, while a larger K suppresses noise more aggressively but may over-smooth image features.
In real-world deployments, noise characteristics vary significantly across optical designs, sensor types, and lighting conditions. Using a fixed K avoids the complexity and potential instability of dynamic noise estimation while enabling consistent and computationally efficient preprocessing.
Moreover, modern architectures such as the residual U-Net demonstrate strong robustness to fixed preprocessing patterns. Through their hierarchical feature representation learning, they effectively compensate for suboptimal filtering artifacts, making the precise tuning of K less critical for overall restoration performance. This empirical value is later validated through dataset-wide sensitivity analysis (see Appendix A).

3.4.2. Patch Assembly Using Transformer-Interpolated PSFs

Figure 5 illustrates this process, summarizing the overall patch assembly and correction pipeline.
Figure 5. Patch assembly and compilation pipeline using Wiener Filter image restoration: The left part illustrates the input aberrated image patches and the interpolated PSFs generated by the Transformer for each query configuration (see Figure 3). The center part shows how the interpolated PSFs are combined with the Wiener filter to reconstruct image patches under each setting. The right part depicts α-weighted fusion of reconstructed results from CS-sim and interpolated PSFs, followed by patch integration, deep feature correction using a residual U-Net (see Section 3.4.3), and final full-size image restoration.
Figure 5. Patch assembly and compilation pipeline using Wiener Filter image restoration: The left part illustrates the input aberrated image patches and the interpolated PSFs generated by the Transformer for each query configuration (see Figure 3). The center part shows how the interpolated PSFs are combined with the Wiener filter to reconstruct image patches under each setting. The right part depicts α-weighted fusion of reconstructed results from CS-sim and interpolated PSFs, followed by patch integration, deep feature correction using a residual U-Net (see Section 3.4.3), and final full-size image restoration.
Sensors 25 03773 g005
Once the patches are restored using the Wiener filter, they are combined to form a full-resolution image using a learned weighting strategy. The filtered outputs based on different PSF configurations are first merged as:
x ¯ k r e c = w k x ^ k r e c + 1 w k x ~ k r e c ,
where w k is a learned weight that balances the contributions of the two filtering results.
To assemble the full image from overlapping patches, the restored patches are spatially blended using attention-based weighting:
X ¯ r e c x , y = k γ k x , y · x ¯ k r e c m , n ,
where γ k x , y s the transformer-derived spatial attention weight that determines each patch’s contribution, normalized as:
k γ k x , y = 1 ,
which determines the contribution of each patch to the final restored image. To stabilize this weighting process, we apply a softplus activation to the attention output before normalization:
z i = s o f t p l u s V i d e c ,
where V i d e c is the decoder output from the transformer, as described in Table 1.
To further ensure seamless transitions between patches, a boundary blending mask is applied:
X ¯ r e c x , y = X ¯ r e c x , y · M x , y + X ¯ b o r d e r x , y · 1 M x , y
with M x , y defined as a normalized blending mask:
M x , y = k γ k x , y k γ k x , y + ϵ ,
where ϵ is a small numerical constant to prevent division by zero. Finally, the Gaussian weighting function used in the mask is given by:
G x , y = e x p x x 0 2 + y y 0 2 2 σ 2 x , y e x p x x 0 2 + y y 0 2 2 σ 2
where ( x 0 ,   y 0 ) represents the center of each patch, and σ controls the smoothness of the blending transition.
The sliding window approach ensures that overlapping regions are blended smoothly, avoiding discontinuities and improving spatial coherence. This step enables deep-corrected patches to be integrated into a fully restored, high-quality image.

3.4.3. Deep Feature Correction Using Residual U-Net

The initial PSFs H ~ and interpolated PSFs H ^ exhibit complementary spatial characteristics: H ~ is locally sharper and exhibits higher peak intensity due to its direct observation, while H ^ tends to be spatially broader and smoother owing to latent space-based interpolation. These differences, while beneficial for robustness in patch-wise PSF coverage, result in spectral inconsistency and potential high-frequency attenuation during synthesis and Wiener filtering. Therefore, the combined image may still exhibit blurred fine structures and edge softening.
To address this, we employ residual U-Net [28] to restore lost high-frequency details by learning spatial priors over multiple PSF-corrected patches, as described below. The restored image from the previous steps suffers from high-frequency loss during the PSF synthesis and Wiener filtering processes.
As illustrated in Figure 6, this architecture extends vanilla U-Net [29] by introducing residual connections, improving gradient flow and feature refinement, minimizing computational overhead while effectively restoring fine structures. We implement several optimizations, such as:
  • Adaptive Receptive Fields: Given the variation in image dimensions within the DIV2K HR Train [30] and RealSR-V3 [31] datasets, we configured adaptive receptive fields to accommodate patch sizes up to 734 × 734 pixels with 17 × 17 padding.
  • Addition-Based Skip Connections: Reduce computational complexity while preserving critical feature information, enhancing segmentation performance in deep learning models such as ResU-Net-a [32] and HRNet (High-Resolution Network) [33].
  • Enhanced Decoder with Residual Additions: Residual connections are added within the decoder to improve feature refinement and gradient flow stability.

3.5. Training Objects and Loss Functions

It is optimized using two main objectives: PSF interpolation fidelity and final image restoration quality. The total loss is defined as:
L t o t a l = L r e c P S F + L r e c ,
The PSF reconstruction loss L r e c P S F ensures that interpolated PSFs remain consistent with observed PSFs. While BCE loss is commonly used for binary classification, it is well-suited here because our interpolated PSF maps can be interpreted as soft assignment weights normalized between 0 and 1. This facilitates stable training and interpretable interpolation:
L r e c P S F = B C E H ~ i , H ^ i ,
The image restoration loss L r e c enforces pixel-wise similarity between the ground truth and restored image using the l 1 -norm:
L r e c = X g t X ¯ r e c 1
L t o t a l represents the overall optimization objective, L r e c P S F regularizes PSF interpolation, while L r e c focuses on perceptual image quality. For models using explicit PSF embeddings, both losses are applied; for non-interpolative variants, only L r e c is used.

4. Experiments and Results

In this section, we present the experimental results for the point spread function (PSF) Interpolative transformer, restoration filter, and deep restoration methods. We begin by describing the datasets and evaluation metrics, followed by a comparison with benchmark studies. Finally, we discuss ablation studies on parameters and data configurations for the proposed model and evaluate computational overhead.

4.1. Experimental Setup

4.1.1. Training Dataset

We employed two datasets for training and evaluation: DIV2K high resolution (DIV2K HR) and RealSR-V3, each serving a distinct experimental purpose.
DIV2K HR was used to evaluate the isolated impact of our simulated PSFs under conditions that minimize sensor influence, enabling validation of the physical consistency and spatial sensitivity of the proposed PSF formulation. The DIV2K dataset contains 800 high-resolution (2K) images, which were degraded using our CS-sim PSF model. This model applies spatially and chromatically varying anisotropic Gaussian blur across the image plane, allowing us to generate diverse aberration patterns without introducing sensor-related artifacts.
RealSR-V3, on the other hand, was employed to validate the model’s generalizability on real sensor-captured images, which inherently include complex and uncontrolled degradations such as chromatic aberration, defocus, geometric distortion, sensor noise, and demosaicing artifacts. We used the Canon2-LR subset, captured with a Canon EOS 5D Mark III (Canon Inc., Tokyo, Japan) and a 24–105 mm f/4.0 zoom lens. Although the dataset provides multiple magnification levels, we selected ×2 scaling, which preserves optical aberration characteristics without introducing aliasing artifacts that commonly occur at higher magnifications.
This Canon2 subset is known to exhibit relatively strong and spatially diverse optical aberrations—particularly in the outer field—due to the zoom lens configuration and wide field of view. These characteristics make it a well-suited benchmark for evaluating pixel-wise PSF estimation and correction under real-world, sensor-integrated conditions.
To maintain consistent lens behavior across both datasets, we applied a common protocol: each distortion case was assigned to a batch of 10 images—5 used for training and 5 for testing—assuming a shared lens profile within each batch.

4.1.2. Training Parameter Configuration

The training process (Equation (28)) was conducted using the hyperparameters based on vanilla transformer [34], vision transformer (ViT) [35], and Lee et al. [36], but with additional modifications to adapt the spatial weighting factor μ that controls how PSF z is applied to nearby query locations by adjusting the attention bias based on spatial distance. The spatial weighting factor ( μ ) is integrated into the cross-attention computation as:
γ i j = s o f t m a x ( Q i W Q ) ( K j W K ) T d k + μ · P i j ,
where γ i j denotes the cross-attention weight between query Q i and key K j ; W Q   a n d   W K are projection matrices; d k is the dimensionality of the key vectors; P i j encodes the relative spatial distance between the query and known PSF positions; and μ R is regularized using the softplus function to ensure positivity and numerical stability during training.
As training proceeds, μ is adaptively learned to modulate spatial bias, enabling the model to prioritize local PSF information that is most relevant to the observed image content. This adaptive mechanism improves the ability to characterize fine-scale optical aberrations that vary spatially across the sensor field.
The detailed architectural and training configuration of our transformer module, including optimizer setup, positional encoding, and loss function structure, is summarized in Table 2.

4.1.3. Residual U-Net Configuration

We define the hyperparameters for training Equation (30) based on those used in residual U-Net [28]. To achieve more robust and fine-grained learning, we introduce the following modifications to the loss function:
  • Loss Function: l 1 -norm loss;
  • Optimizer: Adam optimizer;
  • Learning Rate: 0.0001;
  • Activation Function: ReLU;
  • Batch Size (B): 1.
The proposed approach integrates consistent training and fine-tuning strategies, leveraging stable loss functions and augmentation techniques. These measures ensure robustness across various lens models and distortion conditions.

4.1.4. Chromatic-Spatial Simulated PSF (CS-sim PSF) Setup

We apply our PSF simulation process to both the DIV2K and RealSR-V3 datasets, where PSFs are synthetically generated for training and evaluation.
We simulate the initial PSFs under three different optical conditions across 400 images, considering refractive index variations caused by lens position and wavelength dispersion—two major contributors to high-resolution optical aberrations. Spatially varying aberrations are modeled based on position shifts from the center to the lens edge, while the variation of chromatic aberration with wavelength models the effect of refractive index dispersion in the optical system with wavelength (Section 3.2). To implement these, we design our PSF simulation with the following components:
  • Gaussian blur applied uniformly using an isotropic Gaussian function;
  • Radial anisotropic transformations to simulate spatially varying distortions;
  • Wavelength-dependent distortion modeling (e.g., r < g < b or b < g < r ) with varying inner and outer sigma intensities.
In our experiments, lens position variations are applied at 2-pixel intervals from −5 to 15 pixels. Chromatic aberration effects are controlled via outer sigma, which determines global blur intensity, and inner sigma, which modulates local blur based on pixel distance from the center. For the outer sigma, we set it to range from 1.0 to 2.5, while the inner sigma is sampled between 0.8 and 2.0. These parameters create low, medium, and high aberration levels, enabling training under diverse optical conditions.
These simulated aberrations, combined with the diversity and high-resolution conditions of the DIV2K HR dataset, ensure that the generated PSFs closely reflect real-world aberration variation patterns for robust framework validation.

4.1.5. Fine-Tuning Configuration

Unlike the RealSR-V3 dataset, which consists of images captured using a fixed lens configuration (e.g., Canon, Nikon), the DIV2K dataset lacks explicit lens metadata and is synthetically augmented to simulate diverse aberration patterns. To evaluate the model’s adaptability to such variations, we implemented a fine-tuning strategy tailored to DIV2K. Each session consisted of 300 iterations per simulated lens group and was applied to batches of 10 images (5 for training, 5 for testing), assuming shared optical properties. Hyperparameters were consistent with the general training phase (learning rate: 0.0005; step size: 1000). To stabilize training and smooth loss fluctuations, a moving average with a window size of 50 was applied every 100 iterations.

4.1.6. K Parameter Pre-Evaluation for Robust Filtering

We chose a fixed value of K = 0.01 based on a preliminary evaluation over a wide range of values [0.0002, 0.1] using the full test sets of DIV2K (n = 400) and RealSR-V3 Canon 2 LR (n = 50). While this value was not optimal for all scenarios, it showed robust and consistent performance across both synthetic and real-world conditions. Detailed sensitivity analysis is presented in Appendix A.

4.1.7. Computing Environment

All experiments were conducted on a system equipped with an Intel i7 12th Gen CPU, 64 GB RAM, and an Nvidia RTX 3090 GPU with 24 GB VRAM. The experiments were implemented using Python 3.12, PyTorch 2.2.2 (https://pytorch.org) and CUDA 12.2 on Ubuntu 20.04.

4.2. Benchmark Test

4.2.1. Performance Metrics

For evaluation, we use SSIM (Structural Similarity), PSNR (Peak Signal-to-Noise Ratio), and LPIPS (Learned Perceptual Image Patch Similarity). SSIM measures the preservation of structural features, indicating whether the restored image retains the original structural details. SSIM assesses whether PSF correction via high-dimensional positional information accurately restores image structures. PSNR evaluates image clarity by quantifying the reduction of blurring and distortion, reflecting how accurately positional information is applied during PSF correction. LPIPS, a perceptual metric based on deep features from pre-trained networks, to assess image similarity from a human-vision perspective. LPIPS has been widely adopted in recent image restoration studies and complements SSIM and PSNR by capturing perceptual quality beyond pixel-level similarity.

4.2.2. Benchmark Models

For methodological comparison, we adapted two hybrid computational optics methods—UABC by Li et al. [9] and Chen et al.’s image simulation framework [10]—to fit our testing environment, maintaining their core principles. Both methods originally used Zemax and lens prescriptions for PSF simulation, unavailable in our setup, so we used synthetic PSFs validated against real aberration patterns and calibrated to mimic real-world distributions (Section 4.4).
  • Wiener-Prior UABC-net (adapted from Li et al. [9]): The original UABC uses a PSF-aware deep network, pre-trained on Zemax-simulated PSFs and fine-tuned for specific lenses. We replaced PSF reconstruction with regulated Wiener filtering using synthetic PSFs, keeping the U-Net refinement to reduce artifacts. This preserves the PSF-aware correction focus;
  • Patch-based Deformable ResU-net (adapted from Chen et al. [10]): Chen et al. (2021) proposes an imaging simulation system that computes spatially variant PSFs using lens data and raytracing for post-processing, paired with a spatial-adaptive CNN featuring Deformable ResBlocks. We preserve the Deformable ResBlock’s role in handling non-uniform distortions. This ensures the method’s spatial adaptability is retained.
Both adaptations align with our experimental framework while using validated synthetic PSFs to compensate for missing optical priors. While these implementations differ from the original Zemax-based approaches, we mitigated potential discrepancies through empirical validation, acknowledging that our preprocessing-focused PIABC represents a fundamentally different approach to the same optical correction challenge.

4.2.3. Quantitative and Qualitative Comparison on DIV2K Dataset

To evaluate our physically simulated PSFs in isolation from sensor effects, we conducted quantitative and qualitative comparisons on the DIV2K dataset. These high-resolution images were captured under conditions that minimize sensor influence [30], enabling us to highlight the spatial sensitivity and correction accuracy of our method.
Table 3 presents the PSNR and SSIM results for two patch sizes, 256 × 256 and 128 × 128, evaluated under both globally trained and batch-wise fine-tuned conditions. The benchmark includes our proposed method (PIABC) alongside the two adapted baseline models introduced in Section 4.2.2.
Our method outperformed both baselines across all settings. Compared to the fine-tuned patch-based deformable ResU-net, PIABC achieved up to 9.6% higher PSNR and 4.5% improvement in SSIM, showcasing its capacity to generalize spatial information without relying on handcrafted priors. In addition, PIABC achieved an LPIPS score of 0.2036, representing a decrease of 0.0535 compared to the fine-tuned patch-based deformable-net (0.2571), highlighting improved perceptual similarity in the restored images. This performance stems from the use of high-dimensional positional embeddings and shallow self-attention modules within the autoencoder architecture.
Notably, while fine-tuning improved results for most models, PIABC experienced slight performance degradation post-fine-tuning. This suggests that the fine-tuning process, which is tailored to lens-specific distortions, may overfit and inadvertently reduce generalization. Our findings support the conclusion that PIABC’s design benefits more from broad, diverse training data than from batch-specific adaptation. This observation can be further explained by the fact that lens-specific characteristics are already incorporated through the pre-filtered PSFs applied in the Wiener filtering stage, effectively embedding lens-specific information before residual U-Net training. Consequently, additional lens-specific fine-tuning in the residual U-Net does not provide substantial new information but rather risks reinforcing batch-specific features, which can compromise the generalizable latent representations learned during global training.
Due to the patch-based training structure of the residual U-Net and the incorporation of lens-specific priors via Wiener filtering, fine-tuning often fails to produce noticeable fluctuations in validation loss curves or distinct visual artifacts, making such plots less informative for identifying overfitting in our framework. Instead, performance trends were best captured through quantitative metrics, as reflected in Table 3.
Figure 7 illustrates qualitative differences between the models. PIABC successfully corrects chromatic aberration and recovers fine structures, particularly in edge and texture regions. Compared to the baselines, it produces sharper, cleaner reconstructions with better color fidelity.

4.2.4. Quantitative and Qualitative Comparison Using RealSR-V3 Canon 5D3 Dataset

To evaluate the applicability of our method under real-world conditions, we conducted experiments on the RealSR-V3 Canon2-LR dataset, which includes low-resolution images captured using a zoom-lens DSLR camera Canon EOS 5D Mark III (Canon Inc., Tokyo, Japan) under uncontrolled settings. These images contain diverse degradations, such as optical blur, chromatic aberration, and sensor noise, arising from real lenses and imaging hardware [31].
We specifically selected the image Canon_038_LR2.png, which depicts a scene frequently used in prior benchmark works [28], and employed the ×2 magnification version to preserve the spatial variation of optical aberrations. While the RealSR-V3 dataset does not explicitly annotate aberration severity per image, Canon2 images are known to exhibit relatively strong field-dependent distortions due to DSLR lens characteristics. This observation aligns with prior reports that show similar aberration profiles in datasets acquired using low-end mobile and consumer-grade optics (Chen et al., 2021) [10]. PSFs were estimated directly from the LR image using our proposed model and used in subsequent Wiener filtering and deep residual correction stages. This setup allows us to validate the robustness of our method in the presence of sensor-captured degradations, without relying on synthetic blur assumptions.
Benchmark results are summarized in Table 4. Compared to the baseline methods adapted from [9,10], PIABC achieved the highest PSNR and SSIM values under both 256 × 256 and 128 × 128 patch configurations. Notably, without lens-specific fine-tuning, PIABC attained 30.0258 dB PSNR and 0.8851 SSIM at 256 × 256 resolution, and 30.5716 dB PSNR and 0.8935 SSIM at 128 × 128 resolution. PIABC also achieved LPIPS scores of 0.2143 (256 × 256 patch) and 0.2073 (128 × 128 patch), reflecting reduced perceptual differences relative to the high-resolution ground-truth images. Fine-tuning slightly degraded performance, which is consistent with our observations on DIV2K, suggesting that the pre-learned generalized PSF representations are more robust across diverse real-world variations. This performance drop is likely due to batch-specific adaptation to the limited Canon Test2 subset (50 images). Notably, no distinct visual artifacts were observed, which can be attributed to the patch-based training strategy of the residual U-Net and the pre-filtered PSFs that already encode lens-specific information via Wiener filtering.
Despite known acquisition inconsistencies—having been captured at a different focal length and with a higher magnification—the HR image provided in RealSR-V3 remains a standard reference across recent studies [31,37].
Figure 8 shows the qualitative comparison of restored outputs. PIABC restores sharper edges with fewer artifacts and reduced chromatic shifts. The interpolated PSFs used in our model exhibit smoother spatial structures than those used in other methods, contributing to more effective high-frequency restoration. These results confirm that PSF interpolation, combined with attention-based spatial modeling, enables consistent restoration across both simulated and sensor-acquired real-world images, without requiring prior knowledge of the optical system.

4.3. Ablation Study

4.3.1. Analysis of Stage-Wise Effects and Embedding on DIV2K Dataset

To evaluate the effect of positional and PSF embeddings throughout the correction pipeline, we performed stage-wise analysis across four key steps, as summarized in Table 5.
This experiment was conducted only on the DIV2K dataset, where ground-truth PSF parameters are known. In contrast, the RealSR-V3 dataset does not provide explicit degradation models, making stage-wise attribution infeasible despite the availability of HR-LR pairs. These results quantify the gains at each stage and underscore the contribution of embedding-based PSF interpolation to final restoration quality. Qualitative comparisons corresponding to each stage are illustrated in Figure 9.
Sequential performance gains reveal the relative impact of each module:
  • Applying the Wiener filter to the aberrated image improves PSNR and SSIM by 1.4123 dB and 0.0787, respectively.
  • Replacing observed PSFs with interpolated PSFs yields a marginal improvement (ΔPSNR = 0.1373 dB, ΔSSIM = 0.0024), suggesting interpolation accuracy aligns closely with actual measurements.
  • Compared to using only observed PSFs, the residual U-Net adds a substantial gain of 5.479 dB in PSNR and 0.1161 in SSIM, highlighting the effectiveness of deep refinement.
These results indicate that PSF interpolation meaningfully enhances the performance of the residual U-Net, confirming that the interpolated PSFs contain physically relevant cues that guide the network toward more accurate restorations.

4.3.2. Stage-Wise PSF Modeling and Fine-Tuning Effects on DIV2K Dataset

To examine the effects of PSF embedding and lens-specific fine-tuning on aberration correction performance, we compared four configurations: with/without positional and PSF embedding and before/after fine-tuning. This experiment was performed only on the DIV2K dataset, which provides controlled and consistent PSF labels necessary for isolating embedding and fine-tuning effects. In contrast, RealSR-V3 lacks ground-truth PSFs, making quantitative attribution of such factors infeasible. The results, measured in PSNR and SSIM after both Wiener filtering and residual U-Net deep feature correction, are summarized in Table 6.
Figure 10 illustrates the embedding methods for evaluation.
Across all configurations, PSF embedding consistently led to modest but measurable improvements in PSNR and SSIM. Notably, the best performance was achieved with positional and PSF embedding without fine-tuning. This indicates that embedding high-dimensional PSF features contributes positively to spatial modeling during transformer-based interpolation.
In contrast, the effect of fine-tuning varies depending on the correction stage. When evaluated after using the Wiener filter, fine-tuning slightly improved performance. This suggests that adaptation to lens-specific distortions can benefit early-stage restoration, where correction is shallow and relies on PSF accuracy.
However, after deep correction using the residual U-Net, this trend reversed: models trained without fine-tuning outperformed their fine-tuned counterparts. We interpret this as a possible overfitting effect. Since lens-specific information is already embedded via the PSFs at the Wiener filtering stage, further fine-tuning may introduce redundancy or degrade generalization in the deep learning pipeline.
These findings imply that while embedding enhances restoration through better feature representation, lens-specific fine-tuning should be applied, particularly in architecture where the PSF embedding already encodes optical priors.

4.3.3. Stage-Wise Fine-Tuning Effects on RealSR-V3 Canon 5D3 Dataset

In fine-tuning, additional learning is performed only for data subjected to distortions from the same lens parameters, grouping every 10 images based on presumed similarity.
Unlike conventional SR models, our framework learns specifically for each discrete location, inherently limiting data diversity at each position. Consequently, as shown in Table 7, the combined performance of interpolated PSF and Wiener filtering based on PSF embeddings is reduced after fine-tuning.
PSNR and SSIM show a slight decline compared to the results without fine-tuning, both at the PSF interpolation and residual U-Net stages. This suggests that excessive specialization to limited data induces overfitting, particularly given the small number of available samples (i.e., 50 images in the RealSR-V3 Canon Test2 subset). This trend may also stem from the fact that lens-specific aberration characteristics are already encoded through the PSF-based Wiener filtering, and additional fine-tuning could override the generalizable latent features learned during global training. Consequently, overfitting not only hampers PSF interpolation but also propagates through residual feature refinement stages.
Therefore, while PSF embedding learning, maintaining a wide diversity of PSFs across the latent space appears critical to ensuring generalizable correction.

4.4. Results

The performance of PIABC was evaluated on both synthetic (DIV2K) and real-world (RealSR-V3) datasets, allowing for a comprehensive comparison of its effectiveness across controlled and practical scenarios.

4.4.1. Performance on DIV2K (Synthetic LR)

On the DIV2K dataset, PIABC demonstrated significant improvements over baseline methods. As shown in Table 3, PIABC achieved PSNR values of 31.0044 dB (256 × 256 patch) and 31.4027 dB (128 × 128 patch) before fine-tuning, with SSIM values of 0.9045 and 0.9082, respectively. The corresponding LPIPS values were 0.2036 (256 × 256 patch) and 0.2077 (128 × 128 patch), indicating lower perceptual distortion compared to baseline models.
Compared to the fine-tuned patch-based deformable-net (PSNR 28.2875 dB, SSIM 0.8659, LPIPS 0.2571 for 256 × 256), PIABC yielded up to 9.6% improvement in PSNR, 4.5% in SSIM, and a reduction of 0.0535 in LPIPS.
Figure 7 further illustrates the qualitative superiority of PIABC, showcasing sharper edges, reduced chromatic aberrations, and improved color separation in restored images. These results highlight PIABC’s ability to effectively handle synthetic degradations under controlled conditions, leveraging its interpolative transformer and latent representation space PSF modeling to achieve robust restoration.

4.4.2. Performance on RealSR-V3 (Natural LR)

For the RealSR-V3 dataset, which consists of naturally aberrated low-resolution images captured by a Canon 5D3 DSLR, PIABC maintained strong performance despite the presence of complex sensor noise and optical distortions. Table 4 reports PSNR values of 30.0258 dB (256 × 256 patch) and 30.5716 dB (128 × 128 patch) before fine-tuning, with SSIM values of 0.8851 and 0.8935, respectively. LPIPS values of 0.2143 and 0.2073 (for 256 × 256 and 128 × 128 patches, respectively) further confirm that PIABC restores images with better perceptual fidelity compared to other methods. Compared to the fine-tuned patch-based deformable-net (PSNR 28.9409 dB, SSIM 0.8664 for 256 × 256, and LPIPS 0.2571 to 0.2777 for 256 × 256), PIABC consistently outperformed baselines. Figure 8, PIABC results (lower row, global and fine-tuned) demonstrates PIABC’s ability to restore sharp edges and reduce artifacts in Canon2_LR images, while Figure 8 highlights its effectiveness in recovering fine text details (“VIS CALLS 475”) from a real-world image, achieving significantly sharper text edges and reduced blur artifacts compared to Wiener-prior UABC-net and patch-based deformable-net.

4.4.3. Comparative Analysis and Implications

While PIABC exhibited slightly higher PSNR and SSIM on DIV2K (e.g., PSNR 31.0044 dB vs. 30.0258 dB for 256 × 256 patches) due to the controlled nature of synthetic degradations, its performance on RealSR-V3 remained robust, with only a marginal drop in metrics (e.g., SSIM 0.9045 to 0.8851 for 256 × 256). This consistency underscores PIABC’s generalization ability across diverse degradation types, from synthetic isotropic Gaussian blur to real-world optical aberrations and CMOS sensor noise. The qualitative results further confirm this, as PIABC consistently produced sharper details and fewer artifacts in both datasets (Figure 7 and Figure 8). Notably, the text restoration in Figure 8 emphasizes PIABC’s practical utility in real-world scenarios, where sensor-induced noise and complex aberrations are prevalent, aligning with the needs of CMOS sensor-based systems like the Canon 5D3.
Unlike post-processing approaches that rely on raytraced PSFs from Zemax and lens prescription data [9,10], PIABC takes a pre-processing perspective by simulating chromatic and spatially varying PSFs based on generic optical priors. This is achieved using autoencoder-driven latent embeddings and transformer-based interpolation, without requiring proprietary software or lens-specific metadata. The result is a more adaptable and physically interpretable solution applicable to a wide range of optical systems, including DSLR devices in RealSR-V3.
The effectiveness of this approach is validated by consistent improvements even with low-resolution inputs, as shown in Table 4. Figure 9 and Table 6 further highlight the benefits of our stage-wise embedding and correction process, demonstrating SSIM improvements of up to 4.5% and visibly reduced artifacts in dark regions. Overall, PIABC proves to be a robust computational solution for correcting optical degradation and sensor-related noise, particularly in uncontrolled real-world DSLR imaging environments.

5. Discussion

The results demonstrate that point spread function (PSF) interpolation significantly enhances the performance of the residual U-Net (ResU-Net), indicating that interpolated PSFs provide physically relevant cues that guide the network toward improved restoration accuracy. This chromatic-spatial aberration modeling approach, supported by transformer-based techniques, enables robust correction of lens-induced aberrations, particularly effective for lens-induced degradations under spatially smooth degradation conditions; any observed benefit for CMOS sensor data is an indirect result of improved optical correction. Specifically, the transformer interpolation leverages high-dimensional space embeddings derived from an autoencoder (AE), enabling dense and spatially varying PSF modeling. This AE-driven embedding facilitates the transformer’s ability to capture structured optical and sensor behavior, further improving the accuracy of aberration correction. The method’s reliance on PSF-domain interpolation—rather than noisy pixel intensities—enhances optical correction and overall image quality, with any observed reduction of sensor noise being an incidental outcome rather than a directly modeled effect. This was validated on datasets like RealSR-V3, captured with a Canon EOS 5D Mark III (Canon Inc., Tokyo, Japan), (reported as ISO 100 in 90% of images, up to 400 [31]).
The framework’s effectiveness is validated with high-resolution datasets (e.g., DIV2K HR), where fine textures enable precise patch-wise evaluation. Ablation studies on DIV2K and RealSR-V3 confirm the method’s generalization capability, with RealSR-V3’s sensor-acquired low-resolution images (2× zoom) likely exhibiting authentic optical aberrations.
While fine-tuning was expected to improve adaptation to lens-specific distortions, our experiments showed slight performance degradation (Table 3 and Table 4), as pre-filtered PSFs in the Wiener filtering stage already encoded lens-specific information. Consequently, residual U-Net fine-tuning may reinforce batch-specific features at the expense of generalization, especially with limited data diversity. No noticeable fluctuations in validation loss were observed during fine-tuning, and no distinct visual artifacts emerged, which suggests that performance trends might be better interpreted through quantitative metrics rather than visual comparisons. These results underscore the importance of balancing fine-tuning with broad training data to ensure robust generalization.
Although explicit ground-truth PSF maps are unavailable, the imaging setup—combined with prior literature on lens-based distortion in consumer DSLRs—supports the presence of spatially heterogeneous aberrations in Canon2.
This CS-sim PSF strategy captures structured distortion patterns, some of which may be related to sensor characteristics. However, the framework is primarily designed to model and correct lens-induced degradations. Any incidental improvement in sensor-originated artifacts is a secondary effect of improved optical correction.
While the Gaussian-based PSF model enabled efficient interpolation in this study, we recognize that future work should examine how higher-order aberrations might influence restoration accuracy.
Architecturally, CNN–transformer hybrids leverage attention-based feature fusion, while Deformable ResU-Net enhances local spatial adaptability through deformable convolutions. Both aim to improve spatial feature representation, albeit through different mechanisms. In our framework, the transformer interpolator captures spatial and chromatic dependencies during PSF modeling (pre-deblurring), while the residual U-Net refines high-frequency details after physics-based restoration. Adding another transformer in the post-processing stage may duplicate spatial modeling already addressed in the PSF domain, potentially reducing interpretability and increasing redundancy. Our design separates structural alignment and detail enhancement to maintain clarity and robustness.
For post-processing, while recent diffusion-based generative approaches (e.g., DDPM, DDNM) and hybrid CNN–transformer architectures have shown impressive performance in blind restoration, our study adopts a complementary strategy grounded in physically interpretable priors.

6. Conclusions

This paper presents a Chromatic-Spatial Simulated PSF (CS-sim PSF) framework for correcting spatially varying aberrations by integrating point spread function (PSF)-domain modeling with dense interpolation techniques. The approach employs autoencoder (AE)-driven latent space embeddings to facilitate transformer-based PSF interpolation, recovering high-resolution outputs with physical plausibility.
Validated across synthetic (DIV2K) and real-world (RealSR-V3) datasets, the method enhances CMOS sensor performance with significant quantitative improvements, without requiring lens-specific calibration. It supports scalable deployment with only low-resolution inputs at inference, affirming the practicality of the proposed technique for computational aberration correction in diverse optical systems.

7. Limitations and Future Work

The current framework employs a simplified point spread function (PSF) simulation model that addresses dominant spatial and chromatic degradations but is insufficient to fully capture like coma or astigmatism. Additionally, training relies on high-resolution images, limiting applicability to datasets without high resolution (HR) references. While effective, the use of autoencoder (AE)-driven latent space embeddings for transformer-based PSF interpolation introduces computational complexity due to high-dimensional space processing, potentially restricting real-time sensor applications.
While the Gaussian-based PSF model was chosen for its computational efficiency in this study, its limitations in representing complex wavefront structures are acknowledged. Future work will explore physically grounded models to better capture higher-order aberrations.
We also acknowledge that the current similarity check step does not explicitly account for large field angle PSFs or significant off-axis aberrations, where skew angle variations between adjacent PSFs could reduce the effectiveness of local interpolation. Incorporating more physically informed criteria for similarity checking will be an important area for future study.
Furthermore, the Gaussian-based PSF model, while computationally efficient, may not fully capture complex wavefront structures or higher-order off-axis aberrations.
Future efforts will expand PSF modeling to include a wider range of aberrations and wavefront-based effects. We also plan to optimize the autoencoder–transformer integration to reduce computational cost while preserving the benefits of spatially adaptive PSF interpolation, which is designed to improve optical correction and thereby enhance overall signal quality, potentially reducing the perceptual impact of sensor noise without directly modeling it. Furthermore, developing direct PSF estimation from low-resolution inputs will reduce HR dependency, enhancing applicability under constrained data conditions and improving scalability for practical imaging systems.

Author Contributions

Conceptualization, C.C. and C.K.; methodology, C.C., C.K., and S.S.; software, C.K. and C.C.; validation, C.C. and C.K. and S.S.; formal analysis, C.C. and C.K.; investigation, C.C., C.K., and S.S.; writing—original draft, C.K. and C.C.; writing—review and editing, C.C., C.K., and S.S.; supervision, S.S.; project administration, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study will be openly available in FigShare at https://doi.org/10.6084/m9.figshare.29328275.

Acknowledgments

Generative AI (ChatGPT by OpenAI) was used to assist with language editing and improve the clarity and readability of the text. However, AI did not generate any part of the scientific content, including research design, data analysis, or conclusions. The authors have reviewed and verified all content and take full responsibility for the work.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. K Value Sensitivity Analysis for Robust Wiener Filtering

This appendix supplements Section 4.1.6 by providing expanded visualizations and empirical justification for our fixed choice in the Wiener filtering stage. Full-range sensitivity analysis was conducted on both synthetic (DIV2K HR) and real-world (RealSR-V3 Canon2) datasets. Each dataset evaluation used the full test set to ensure statistical reliability: 400 images for DIV2K HR and 50 images for RealSR-V3.
Figure A1. Sensitivity analysis of K value: (a) for synthetic dataset DIV2K; (b) for real-world datasets RealSR(V3) Canon 2 LR; (Left) PSNR vs. K. (Middle) SSIM vs. K. (Right) Distribution of estimated K values from MMSE criterion.
Figure A1. Sensitivity analysis of K value: (a) for synthetic dataset DIV2K; (b) for real-world datasets RealSR(V3) Canon 2 LR; (Left) PSNR vs. K. (Middle) SSIM vs. K. (Right) Distribution of estimated K values from MMSE criterion.
Sensors 25 03773 g0a1aSensors 25 03773 g0a1b
To rigorously validate our empirical choice of K = 0.01, we conducted extensive sensitivity analysis across two distinct datasets representing different degradation characteristics. K values were evaluated, and performance was measured using PSNR and SSIM, with error bars representing 95% confidence intervals across full test sets (n = 400 and n = 50, respectively).
Table A1. Quantitative results for K sensitivity analysis based on full test sets.
Table A1. Quantitative results for K sensitivity analysis based on full test sets.
DatasetK ValuePSNR (dB)SSIM
DIV2K
HR Train
0.00029.81 ± 2.040.0815 ± 0.0765
0.001011.30 ± 2.650.1857 ± 0.1294
0.005013.39 ± 3.280.3413 ± 0.1684
0.010014.53 ± 3.540.4197 ± 0.1747
0.050017.70 ± 3.790.6026 ± 0.1580
0.100018.69 ± 3.200.6703 ± 0.1397
RealSR (V3)
Canon 2 LR
0.000214.06 ± 2.700.2305 ± 0.0754
0.001016.47 ± 3.060.4152 ± 0.0994
0.005018.69 ± 3.370.5819 ± 0.1038
0.010019.58 ± 3.450.6372 ± 0.1006
0.050021.42 ± 3.180.7355 ± 0.0843
0.100021.34 ± 2.540.7664 ± 0.0763
For each K value, we measured PSNR and SSIM metrics across multiple test samples, with error bars representing 95% confidence intervals. While the optimal K varies with data characteristics (K = 0.1947 for DIV2K and K = 0.0602 for RealSR), K = 0.01 provides stable and competitive results (within 2–4 dB of the optimal) across both domains.
Table A2. Comparison of optimal and empirical K values across datasets. The deviation factor quantifies the ratio between the MMSE-optimal and fixed.
Table A2. Comparison of optimal and empirical K values across datasets. The deviation factor quantifies the ratio between the MMSE-optimal and fixed.
DatasetOptimal KK = 0.01 PerformanceDeviation Factor
DIV2K HR Train0.1947PSNR: 14.53 ± 3.54 dB,
SSIM: 0.4197 ± 0.1747
19.5×
RealSR (V3) Canon 2 LR0.0602PSNR: 19.58 ± 3.45 dB,
SSIM: 0.6372 ± 0.1006
6.0×
Note: The estimated K values are obtained using an MMSE-based method that leverages Laplacian variance to estimate noise power (Immerkaer, 1996 [38]), widely adopted for unsupervised Wiener filter tuning.
The results demonstrate the following insights:
  • Validity of K = 0.01:
    • The histogram of MMSE-estimated K values shown in Figure A1 (a, right) and Figure A1 (b, right) shows a dominant concentration in the range of 0.02 to 0.12, with a mean of 0.1947 and 0.0602, respectively. This supports the view that our empirical choice (K = 0.01) is a conservative yet well-grounded setting.
    • Our empirical choice of K = 0.01 lies near the lower bound of this range, representing a conservative yet reasonable selection grounded in signal preservation.
  • Sensitivity Trends:
    • As K increases from 0.0002 to 0.1000, both PSNR and SSIM improve steadily:
      DIV2K: PSNR improves from 9.81 dB to 18.69 dB, SSIM from 0.0815 to 0.6703,
      RealSR: PSNR improves from 14.06 dB to 21.34 dB, SSIM from 0.2305 to 0.7664.
    • It shows a robust, quasi-logarithmic relationship between K and restoration quality.
  • Robustness of Fixed K:
    • Despite not being the MMSE-optimal, K = 0.01 yields strong intermediate performance across datasets.
    • RealSR shows only 1.76 dB loss (vs. optimal K); DIV2K shows 4.16 dB loss.
    • The consistent error bar trends validate the reliability of this fixed setting across real and synthetic domains.
Therefore, the empirical selection of K = 0.01 represents a robust engineering compromise that maintains strong restoration performance across both synthetic and real-world scenarios. Its stability, combined with the deep network’s ability to compensate for residual noise, justifies its use as a fixed hyperparameter.

Appendix B. Illustration of CS-sim-PSF

To complement Section 3.2, we provide an illustration of the CS-sim-PSF map generation process using a 128 × 128-pixel patch. Figure A2 shows four cropped PSF patches from the upper-right corner of the image, each depicting a different stage of simulated optical degradation: (a) isotropic Gaussian blur, (b) additional blur, (c) chromatic distortion, and (d) elliptical distortion. The lower image shows a zoomed-in view of the PSFs to illustrate the progression of degradation steps more clearly.
Figure A2. Progressive illustration of a CS-sim-PSF map using 128 × 128-pixel patch. (a) Initial isotropic Gaussian PSF; (b) addition of Gaussian blur; (c) incorporation of chromatic distortion; (d) application of elliptical distortion. These steps correspond to the sequential combination described in Equations (4)–(13) in Section 3.2.
Figure A2. Progressive illustration of a CS-sim-PSF map using 128 × 128-pixel patch. (a) Initial isotropic Gaussian PSF; (b) addition of Gaussian blur; (c) incorporation of chromatic distortion; (d) application of elliptical distortion. These steps correspond to the sequential combination described in Equations (4)–(13) in Section 3.2.
Sensors 25 03773 g0a2
These progressive degradations are generated using the parametric models defined in Equations (4)–(13) of Section 3.2. In particular, the PSF at each pixel is defined as a composition of isotropic and elliptical Gaussian functions, along with chromatic displacements, based on this formulation.
This formulation allows us to approximate local aberration patterns, including both chromatic displacement and elliptical distortions, without requiring direct wavefront measurements or global Zernike decompositions. It effectively captures spatially varying degradations and supports patch-wise training of the restoration model. While our method does not explicitly validate PSFs against Zernike polynomials, future work may explore such integration to bridge traditional optical models and data-driven frameworks.

References

  1. Kondepudy, R.; Healey, G. Using Moment Invariants to Analyze 3-D Color Textures. In Proceedings of the 1st International Conference on Image Processing, Austin, TX, USA, 13–16 November 1994; Volume 2, pp. 61–65. [Google Scholar]
  2. Foi, A.; Trimeche, M.; Katkovnik, V.; Egiazarian, K. Practical Poissonian-Gaussian Noise Modeling and Fitting for Sin-gle-Image Raw-Data. IEEE Trans. Image Process. 2008, 17, 1737–1754. [Google Scholar] [CrossRef] [PubMed]
  3. El Gamal, A.; Eltoukhy, H. CMOS Image Sensors. IEEE Circuits Devices Mag. 2005, 21, 6–20. [Google Scholar] [CrossRef]
  4. Zhang, C.; Li, Y.; Wang, J.; Hao, P. Universal Demosaicking of Color Filter Arrays. IEEE Trans. Image Process. 2016, 25, 5173–5186. [Google Scholar] [CrossRef]
  5. Tian, J.; Chen, L. Adaptive Multi-Focus Image Fusion Using a Wavelet-Based Statistical Sharpness Measure. Signal Process. 2012, 92, 2137–2146. [Google Scholar] [CrossRef]
  6. Hasinoff, S.W.; Durand, F.; Freeman, W.T. Noise-Optimal Capture for High Dynamic Range Photography. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 553–560. [Google Scholar]
  7. Mildenhall, B.; Barron, J.T.; Chen, J.; Sharlet, D.; Ng, R.; Carroll, R. Burst Denoising with Kernel Prediction Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2502–2510. [Google Scholar]
  8. Sitzmann, V.; Diamond, S.; Peng, Y.; Dun, X.; Boyd, S.; Heidrich, W.; Heide, F.; Wetzstein, G. End-to-End Optimization of Optics and Image Processing for Achromatic Extended Depth of Field and Super-Resolution Imaging. ACM Trans. Graph. 2018, 37, 114. [Google Scholar] [CrossRef]
  9. Li, X.; Suo, J.; Zhang, W.; Yuan, X.; Dai, Q. Universal and Flexible Optical Aberration Correction Using Deep-Prior Based Deconvolution. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 2593–2601. [Google Scholar]
  10. Chen, S.; Feng, H.; Pan, D.; Xu, Z.; Li, Q.; Chen, Y. Optical Aberration Correction in Postprocessing Using Imaging Simulation. ACM Trans. Graph. 2021, 40, 192. [Google Scholar] [CrossRef]
  11. Zhang, L.; Zhang, H.; Shen, H.; Li, P. A Super-Resolution Reconstruction Algorithm for Surveillance Images. Signal Process. 2010, 90, 848–859. [Google Scholar] [CrossRef]
  12. Heide, F.; Steinberger, M.; Tsai, Y.-T.; Rouf, M.; Pająk, D.; Reddy, D.; Gallo, O.; Liu, J.; Heidrich, W.; Egiazarian, K.; et al. FlexISP: A Flexible Camera Image Processing Framework. ACM Trans. Graph 2014, 33, 231. [Google Scholar] [CrossRef]
  13. Schuler, C.J.; Hirsch, M.; Harmeling, S.; Schölkopf, B. Learning to Deblur. Adv. Neural Inf. Process. Syst. 2014, 27, 1439–1451. [Google Scholar] [CrossRef]
  14. Shih, Y.; Guenter, B.; Joshi, N. Image Enhancement Using Calibrated Lens Simulations. In Computer Vision—ECCV 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7575, pp. 42–56. ISBN 978-3-642-33764-2. [Google Scholar]
  15. Goodman, J.W. Introduction to Fourier Optics, 4th ed.; W. H. Freeman: New York, NY, USA, 2017; ISBN 978-1-319-11916-4. [Google Scholar]
  16. Kee, E.; Paris, S.; Chen, S.; Wang, J. Modeling and Removing Spatially-Varying Optical Blur. In Proceedings of the 2011 IEEE International Conference on Computational Photography (ICCP), Pittsburgh, PA, USA, 8–10 April 2011; pp. 1–8. [Google Scholar]
  17. Stallinga, S. Accuracy of the Gaussian Point Spread Function Model in 2D Localization Microscopy. Opt. Express 2010, 18, 24461–24475. [Google Scholar] [CrossRef]
  18. Siddik, A.B.; Sandoval, S.; Voelz, D.; Boucheron, L.E.; Varela, L. Deep learning estimation of modified Zernike coefficients for image point spread functions. Opt. Express 2023, 31, 22903–22913. [Google Scholar] [CrossRef] [PubMed]
  19. Zhang, H.; Dai, Y.; Li, H.; Koniusz, P. Deep Stacked Hierarchical Multi-Patch Network for Image Deblurring. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5971–5979. [Google Scholar]
  20. Jing, L.; Zbontar, J.; LeCun, Y. Implicit Rank-Minimizing Autoencoder. Adv. Neural Inf. Process. Syst. 2020, 33, 14736–14746. [Google Scholar]
  21. Monakhova, K. Physics-Informed Machine Learning for Computational Imaging. Ph.D. Thesis, University of California, Berkeley, CA, USA, 2022. [Google Scholar]
  22. Chouzenoux, E.; Lau, T.T.K.; Lefort, C.; Pesquet, J.-C. Optimal Multivariate Gaussian Fitting with Applications to PSF Mod-eling in Two-Photon Microscopy Imaging. J. Math. Imaging Vis. 2019, 61, 1037–1050. [Google Scholar] [CrossRef]
  23. Jia, P.; Li, X.; Li, Z.; Wang, W.; Cai, D. Point Spread Function Modelling for Wide Field Small Aperture Telescopes with a Denoising Autoencoder. Mon. Not. R. Astron. Soc. 2020, 493, 651–660. [Google Scholar] [CrossRef]
  24. Schuler, C.J.; Hirsch, M.; Harmeling, S.; Schölkopf, B. Blind Correction of Optical Aberrations. In Computer Vision—ECCV 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7574, pp. 187–200. ISBN 978-3-642-33711-6. [Google Scholar]
  25. Levin, A.; Weiss, Y.; Durand, F.; Freeman, W.T. Understanding and Evaluating Blind Deconvolution Algorithms. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1964–1971. [Google Scholar]
  26. Yue, T.; Suo, J.; Wang, J.; Cao, X.; Dai, Q. Blind Optical Aberration Correction by Exploring Geometric and Visual Priors. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1684–1692. [Google Scholar]
  27. Li, W.; Yin, X.; Liu, Y.; Zhang, M. Computational Imaging through Chromatic Aberration Corrected Simple Lenses. J. Mod. Opt. 2017, 64, 2211–2220. [Google Scholar] [CrossRef]
  28. Zhang, Z.; Liu, Q.; Wang, Y. Road Extraction by Deep Residual U-Net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  30. Agustsson, E.; Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1122–1131. [Google Scholar]
  31. Cai, J.; Zeng, H.; Yong, H.; Cao, Z.; Zhang, L. Toward Real-World Single Image Super-Resolution: A New Benchmark and a New Model. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3086–3095. [Google Scholar]
  32. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A Deep Learning Framework for Semantic Segmentation of Remotely Sensed Data. ISPRS J. Photogramm. Remote Sens. 2020, 162, 94–114. [Google Scholar] [CrossRef]
  33. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 5693–5703. [Google Scholar] [CrossRef]
  34. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  35. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  36. Lee, S.H.; Lee, S.; Song, B.C. Vision Transformer for Small-Size Datasets. Sensors 2021, 21, 8448. [Google Scholar] [CrossRef]
  37. Saeed Rad, M.; Yu, T.; Musat, C.; Kemal Ekenel, H.; Bozorgtabar, B.; Thiran, J.-P. Benefiting from Bicubically Down-Sampled Images for Learning Real-World Image Super-Resolution. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 1589–1598. [Google Scholar]
  38. Immerkær, J. Fast Noise Variance Estimation. Comput. Vis. Image Underst. 1996, 64, 300–302. [Google Scholar] [CrossRef]
Figure 1. Proposed end-to-end pipeline for sensor-based image restoration: The unboxed area illustrates the conventional image acquisition process, where unknown blur (PSF, point spread function) and sensor-induced electrical noise degrade image quality. Improving optical quality often leads to simultaneous enhancements in sensor performance, such as increased efficiency, reduced photon shot noise, and an extended dynamic range. In the proposed PIABC framework, spatially varying PSFs are first estimated from degraded sensor signals using a transformer-based interpolation method. These PSFs are then applied via Wiener filtering and further refined through residual correction to restore high-frequency details. The entire framework provides interpretable correction for optical aberrations and also demonstrates that, as a secondary benefit, optical restoration can reduce sensor-related artifacts. Here, η is sensor noise, and y * and h denote the patch of the degraded observation Y * = Y + η and the corresponding patch of the interpolated PSF field H respectively.
Figure 1. Proposed end-to-end pipeline for sensor-based image restoration: The unboxed area illustrates the conventional image acquisition process, where unknown blur (PSF, point spread function) and sensor-induced electrical noise degrade image quality. Improving optical quality often leads to simultaneous enhancements in sensor performance, such as increased efficiency, reduced photon shot noise, and an extended dynamic range. In the proposed PIABC framework, spatially varying PSFs are first estimated from degraded sensor signals using a transformer-based interpolation method. These PSFs are then applied via Wiener filtering and further refined through residual correction to restore high-frequency details. The entire framework provides interpretable correction for optical aberrations and also demonstrates that, as a secondary benefit, optical restoration can reduce sensor-related artifacts. Here, η is sensor noise, and y * and h denote the patch of the degraded observation Y * = Y + η and the corresponding patch of the interpolated PSF field H respectively.
Sensors 25 03773 g001
Figure 2. Concept of PIABC. (a) CS-sim PSF Patch (initial and interpolated): Each image patch is subdivided into four regions, from which CS-sim PSFs are embedded into a latent space using a trained autoencoder. A transformer decoder computes attention between a query position (marked with an arrow) and these sub-patches to interpolate the corresponding PSF. Query position ti is provided independently as input and serves as the query in the transformer cross-attention. PSF interpolation block: The attention output ( γ i j ) from the cross-attention is combined with position similarity weight ( α i j ) to produce interpolated PSF vectors ( z i j a , b , c , d ) , shown vertically stacked in the figure. (b) Image restoration, patch assembly, and deep correction: interpolated PSFs are used sequentially for Wiener deconvolution and residual correction to reconstruct the final restored image. All steps are described in detail in Section 3.2, Section 3.3 and Section 3.4. For the overall query and correction structure, see Figure 3; for details of the transformer architecture, see Figure 4, and for the residual correction pipeline, see Figure 5. Note: Arrows are shown as examples, and only two are depicted here for clarity.
Figure 2. Concept of PIABC. (a) CS-sim PSF Patch (initial and interpolated): Each image patch is subdivided into four regions, from which CS-sim PSFs are embedded into a latent space using a trained autoencoder. A transformer decoder computes attention between a query position (marked with an arrow) and these sub-patches to interpolate the corresponding PSF. Query position ti is provided independently as input and serves as the query in the transformer cross-attention. PSF interpolation block: The attention output ( γ i j ) from the cross-attention is combined with position similarity weight ( α i j ) to produce interpolated PSF vectors ( z i j a , b , c , d ) , shown vertically stacked in the figure. (b) Image restoration, patch assembly, and deep correction: interpolated PSFs are used sequentially for Wiener deconvolution and residual correction to reconstruct the final restored image. All steps are described in detail in Section 3.2, Section 3.3 and Section 3.4. For the overall query and correction structure, see Figure 3; for details of the transformer architecture, see Figure 4, and for the residual correction pipeline, see Figure 5. Note: Arrows are shown as examples, and only two are depicted here for clarity.
Sensors 25 03773 g002
Figure 6. Deep Correction with residual U-Net for our PSF interpolative aberration correction: Input Processing: Fourier-restored image patches of varying sizes are passed through a 3 × 3 convolution filter, converting the input to B × 4 × H × W. Subsequently, it is transformed into 1 × 64 × H × W; Encoding: Involves three layers of stride convolutions, downsampling the input to a bottleneck layer; Bottleneck: Reduce features to 512 Ch, downsampling by a factor of 1/8 to retain critical information; Decoding: Performed through 3 stages of transpose convolution layers, upsampling the features back to the original dimensions; Residual Blocks: Each encoder and decoder block incorporates two 3 × 3 convolution layers and a batch normalization (BN) layer, followed by element-wise residual addition.
Figure 6. Deep Correction with residual U-Net for our PSF interpolative aberration correction: Input Processing: Fourier-restored image patches of varying sizes are passed through a 3 × 3 convolution filter, converting the input to B × 4 × H × W. Subsequently, it is transformed into 1 × 64 × H × W; Encoding: Involves three layers of stride convolutions, downsampling the input to a bottleneck layer; Bottleneck: Reduce features to 512 Ch, downsampling by a factor of 1/8 to retain critical information; Decoding: Performed through 3 stages of transpose convolution layers, upsampling the features back to the original dimensions; Residual Blocks: Each encoder and decoder block incorporates two 3 × 3 convolution layers and a batch normalization (BN) layer, followed by element-wise residual addition.
Sensors 25 03773 g006
Figure 7. Qualitative comparison of DIV2K dataset: (a) ground truth image; (b) aberrated image; (c) PIABC (Ours, before finetuning); (d) Wiener-prior UABC-net (local finetune); (e) patch-based deformable-net (local finetune).
Figure 7. Qualitative comparison of DIV2K dataset: (a) ground truth image; (b) aberrated image; (c) PIABC (Ours, before finetuning); (d) Wiener-prior UABC-net (local finetune); (e) patch-based deformable-net (local finetune).
Sensors 25 03773 g007
Figure 8. Qualitative comparison on RealSR-V3: Canon2 LR, ID: 038.png; Left: Ground-truth high-resolution image (Canon2-HR); Right: ROI on HR image showing text region (“VIS, CALLS, 475”); Upper: HR, LR, Bicubic, Wiener-prior UABC-net (globally trained and fine-tuned respectively); Lower: Patch-based deformable-net (globally trained and fine-tuned respectively), PIABC (before and after fine-tuning, respectively), PSF (initial CS-sim PSF and interpolated PSF, respectively).
Figure 8. Qualitative comparison on RealSR-V3: Canon2 LR, ID: 038.png; Left: Ground-truth high-resolution image (Canon2-HR); Right: ROI on HR image showing text region (“VIS, CALLS, 475”); Upper: HR, LR, Bicubic, Wiener-prior UABC-net (globally trained and fine-tuned respectively); Lower: Patch-based deformable-net (globally trained and fine-tuned respectively), PIABC (before and after fine-tuning, respectively), PSF (initial CS-sim PSF and interpolated PSF, respectively).
Sensors 25 03773 g008
Figure 9. Stage-wise qualitative comparison of restored images: Each row shows a test image from the DIV2K dataset processed through four stages of our restoration pipeline: (a) Ground Truth; (b) Aberrated image; (c) Final output after residual U-Net correction; (d) Wiener-filtered image using observed PSFs; (e) Patch-wise compilation using interpolated PSFs. Below each row, red-boxed regions are cropped and enlarged for detailed comparison. For (d,e), the applied PSFs are also visualized on the right. Interpolated PSFs appear smoother and more spatially coherent due to softmax-weighted blending and softplus regularization, which helps reduce aliasing at the cost of some high-frequency sharpness. This facilitates more effective deep correction using residual U-Net.
Figure 9. Stage-wise qualitative comparison of restored images: Each row shows a test image from the DIV2K dataset processed through four stages of our restoration pipeline: (a) Ground Truth; (b) Aberrated image; (c) Final output after residual U-Net correction; (d) Wiener-filtered image using observed PSFs; (e) Patch-wise compilation using interpolated PSFs. Below each row, red-boxed regions are cropped and enlarged for detailed comparison. For (d,e), the applied PSFs are also visualized on the right. Interpolated PSFs appear smoother and more spatially coherent due to softmax-weighted blending and softplus regularization, which helps reduce aliasing at the cost of some high-frequency sharpness. This facilitates more effective deep correction using residual U-Net.
Sensors 25 03773 g009
Figure 10. Transformer scheme for image patch input on Cartesian coordinate system: The configurations in (a,b) correspond to those in Figure 3. The transformer module is illustrated here in a simplified form, as shown in Figure 4; (a) None of embeddings are applied, only position information (coordinate) is used with 5 query positions; (b) Positional information and PSF embedding used with 1 interpolation query position.; The color shading on the left represents positional information; The letters ( a d , t 1 t 5 ) follow the patch and query index notation used in Figure 2, Figure 3 and Figure 4; In (b), A E e n c ( P S F j ) denotes the autoencoder’s encoder module applied to the PSF patch at sub-patch index j ( j { a , b , c , d } ).
Figure 10. Transformer scheme for image patch input on Cartesian coordinate system: The configurations in (a,b) correspond to those in Figure 3. The transformer module is illustrated here in a simplified form, as shown in Figure 4; (a) None of embeddings are applied, only position information (coordinate) is used with 5 query positions; (b) Positional information and PSF embedding used with 1 interpolation query position.; The color shading on the left represents positional information; The letters ( a d , t 1 t 5 ) follow the patch and query index notation used in Figure 2, Figure 3 and Figure 4; In (b), A E e n c ( P S F j ) denotes the autoencoder’s encoder module applied to the PSF patch at sub-patch index j ( j { a , b , c , d } ).
Sensors 25 03773 g010
Table 1. Overview of transformer-based PSF interpolation process.
Table 1. Overview of transformer-based PSF interpolation process.
StageInputOutputDimensionDescription
1CS-sim PSF
h ~ i j s p
Latent feature
h ˘ i j s p
( N , C i n ,   H P S F ,   W P S F ) ( N , 768 ) Patch-level PSFs are flattened and embedded as feature vector.
2 h ˘ i j s p , m j s p , c i ( = c i p ) Encoded vector
Q i e n c
N , 768 , N , 2 N , 2 ( N , 768 ) Encoder feature by combining AE output, local PE, and global PE.
3Query position
t i l
Decoder query vector
q i l d e c
( M , 2 ) ( M , 768 ) Decoder query tokens generated from 2D query positions.
4 q i l d e c , Q i e n c Aggregated latent
Feature   z i j
( M , 768 ) ( M , 768 ) Attention aggregates latent features for each query from encoder outputs.
5 z i j Intermediate PSF
h ^ i j t l
( 1 ,   768 ) ( 1 ,   H P S F ,   W P S F ) Decoder reshapes latent feature to generate interpolated PSF patch.
6 h ^ i j t l Final Interpolated PSF
H ^ i   p ^ i ,   c
( M ,   H P S F ,   W P S F ) ( 1 ,   H P S F ,   W P S F ) Weighted sum and smoothing produce final interpolated PSF.
i : sub-patch index; j : query index, j { a , b , c , d } ; m j s p : Sub-patch center position vector ( x ,   y ) ; c i ( = c i p ) : Patch center position vector x ,   y . (Note: In embedding-related notations (e.g., embedding indices), regular font is used); p denotes patch; q i , l d e c , Q i e n c : Decoder and encoder input features, respectively; t i : interpolated PSF target position (final index simplified to i for consistency); z i : Aggregated latent feature at target position; H ^ i : Interpolated PSF at target position, obtained via a weighted combination of sub-patch contributions. ( j index aggregated via α i j , γ i j weighted sum.); Self-attention in the encoder and cross-attention in the decoder each use 12 heads.
Table 2. Transformer training and architecture configuration.
Table 2. Transformer training and architecture configuration.
ComponentSpecificationComponentSpecificationComponentSpecification
Embedding
Dimension
768Activation
Function
ReLU
(PatchEmbed., MLP)
Batch Size1
Encoder
Layers
12NormalizationLayerNorm
(pre-norm config.)
Learning Rates10−4 (model params),
5 × 10−4 (spatial weight μ)
Decoder
Layers
6Positional
Encoding
Learnable 2D
(Cartesian coord.)
OptimizerAdam,
β₁ = 0.9, β₂ = 0.999
Attention
Heads
12Initialization
Scheme
Kaiming Uniform
(PyTorch default)
SchedulerStepLR, step size = 1000, decay γ = 0.9
Feed-Forward
Hidden Dimension
1536
(MLP ratio = 2.0)
Iterations
(global training)
20,000 Loss Function
(see Equations (27)–(29))
Total =
Rec-PSF + Rec1
Dropout Rate0.1Iterations
(Fine-tuning)
300Loss WeightsUniform
Rec1: Equation (29) defines the main image reconstruction loss using an l 1 -norm. During implementation, auxiliary terms such as masked region reconstruction and KL (Kullback-Leibler) regularization are added to strengthen spatial learning and stabilize latent representations; Training logs and image outputs are automatically generated during each run to facilitate reproducibility and fine-tuning validation; Data augmentation: Random cropping and horizontal flipping were applied to enhance generalization; Training iterations: 20,000 iterations across 400 training images, using 80 simulated lens models (40 for training and 40 for testing).
Table 3. PSNR and SSIM comparison to benchmark works on DiV2K (Average). (a) Five query positions with 256 × 256 patch size; (b) Single central query position with 128 × 128 patch size. (See Figure 3).
Table 3. PSNR and SSIM comparison to benchmark works on DiV2K (Average). (a) Five query positions with 256 × 256 patch size; (b) Single central query position with 128 × 128 patch size. (See Figure 3).
Benchmark Model(a) 256 × 256 Size Patch(b) 128 × 128 Size Patch
PSNRSSIMLPIPSPSNRSSIMLPIPS
Wiener-prior UABC-net 1
(4 global train)
20.92330.62730.373021.36390.64180.3672
Wiener-prior UABC-net 1
(5 local finetune)
22.10030.67730.342725.45060.78670.3382
Patch-based Deformable-net 2
(4 global train)
26.33580.81470.293025.95870.80720.3066
Patch-based Deformable-net 2
(5 local finetune)
28.28750.86590.257127.58920.85270.2777
PIABC (ours)
(before 6 fine-tuning) 3
31.00440.90450.203631.40270.90820.2077
PIABC (ours)
(after 6 fine-tuning) 3
30.76110.90000.207731.29930.90790.2058
1 Adapted from Li et al. [9]: Uses a PSF-aware deep network pre-trained on simulated lens data; 2 Adapted from Chen et al. [10]: Raytraced PSFs with a deformable CNN for spatial correction; 3 Ours (PIABC): Simulated PSF-based correction using transformer-driven latent interpolation; 4 globally trained: Trained without considering any variations in dataset (here, DIV2K HR train); 5 locally fine-tuned: Perform for every batch of 10 images (5 for training, 5 for testing) assuming those were captured with identical lens (Section 4.1.4); 6 Fine-tuning: The same as 5 locally fine-tuned; Bold numbers indicate the best performance in each column.
Table 4. RealSR-V3 benchmark results on Canon 2 LR (average). (a) Five query positions with 256×256 patch size; (b) Single central query position with 128 × 128 patch size. (See Figure 3).
Table 4. RealSR-V3 benchmark results on Canon 2 LR (average). (a) Five query positions with 256×256 patch size; (b) Single central query position with 128 × 128 patch size. (See Figure 3).
Benchmark Model(a) 256 × 256 Size Patch(b) 128 × 128 Size Patch
PSNRSSIMLPIPSPSNRSSIMLPIPS
Wiener-prior UABC-net 1
(4 global train)
25.43070.76230.373024.55120.74650.3672
Wiener-prior UABC-net 1
(5 local finetune)
26.43000.81090.342726.37500.81570.3382
Patch-based Deformable-net 2
(4 global train)
28.05150.84500.293027.98090.84620.3066
Patch-based Deformable-net 2
(5 local finetune)
28.94090.86640.257129.20010.87330.2777
PIABC (ours)
(before 6 fine-tuning) 3
30.02580.88510.214330.57160.89350.2073
PIABC (ours)
(after 6 fine-tuning) 3
28.91280.87680.220830.10900.88720.2120
1 Adapted from Li et al. [9]: Uses a PSF-aware deep network pre-trained on simulated lens data; 2 Adapted from Chen et al. [10]: Raytraced PSFs with a deformable CNN for spatial correction; 3 Ours (PIABC): Simulated PSF-based correction using transformer-driven latent interpolation; 4 globally trained: Trained without considering any variations in dataset (here, RealSR Canon 2 LR); 5 locally fine-tuned: Perform for every batch of 10 images (5 for training, 5 for testing) assuming those were captured with identical lens (Section 4.1.4); 6 Fine-tuning: The same as 5 locally fine-tuned; Bold numbers indicate the best performance in each column.
Table 5. PSNR and SSIM across different stages on DIV2K (Average): using Cartesian positional information and PSF embedding for 256 × 256 patches.
Table 5. PSNR and SSIM across different stages on DIV2K (Average): using Cartesian positional information and PSF embedding for 256 × 256 patches.
Process StepPSNR (dB)SSIM (dB)
Aberrated image.23.97580.7073
Wiener filter using observed PSFs.25.38810.7860
Compiled image using interpolated PSFs.25.52540.7884
Deeply corrected image.31.00440.9045
Bold numbers indicate the best performance in each column.
Table 6. PSNR and SSIM under four configurations (Average): with/without PSF embedding and before/after fine-tuning. All values are reported after Wiener filtering and residual U-Net correction. Fine-tuning was performed using image batches grouped in units of 10, assuming they share the same lens distortion characteristics.
Table 6. PSNR and SSIM under four configurations (Average): with/without PSF embedding and before/after fine-tuning. All values are reported after Wiener filtering and residual U-Net correction. Fine-tuning was performed using image batches grouped in units of 10, assuming they share the same lens distortion characteristics.
Fine-TuningEmbedding Type1 Wiener Restoration2 ResU-net Deep Correction
PSNRSSIMPSNRSSIM
Before3 Position only25.48400.787430.89060.9028
Before4 Positional + PSF embedding25.52540.788431.00440.9045
After3 Position only25.53410.789230.77360.9034
After4 Positional +PSF embedding25.53750.789330.71510.9031
1 Wiener: restoration using Wiener filtering; 2 ResU-Net: Deep correction using the residual U-Net; 3 Position only: Uses position information (coordinate) directly, without any explicit positional or PSF embeddings; 4 Positional + PSF: Incorporates both coordinate embeddings and PSF feature embeddings. Experiments in this table are conducted only on the DIV2K dataset due to the availability of ground-truth PSFs; Bold numbers indicate the best performance in each column.
Table 7. PSNR and SSIM under stage-wise and with/without fine-tuning configuration (average): All values are reported after Wiener filtering and residual U-Net correction. Fine-tuning was performed using image batches grouped in units of 10, assuming they share the same lens distortion characteristics. (a) Five query positions with 256 × 256 patch size; (b) Single central query position with 128 × 128 patch size. (See Figure 3).
Table 7. PSNR and SSIM under stage-wise and with/without fine-tuning configuration (average): All values are reported after Wiener filtering and residual U-Net correction. Fine-tuning was performed using image batches grouped in units of 10, assuming they share the same lens distortion characteristics. (a) Five query positions with 256 × 256 patch size; (b) Single central query position with 128 × 128 patch size. (See Figure 3).
Fine-TuningStage(a) 256 × 256 Size Patch(b) 128 × 128 Size Patch
PSNRSSIMPSNRSSIM
Before1 Wiener26.69830.803826.89370.8081
Before2 ResU-net30.02580.885130.57160.8935
After1 Wiener26.69150.803826.91910.8091
After2 ResU-net28.91280.876830.10900.8872
1 Wiener: restoration using Wiener filtering; 2 ResU-Net: Deep correction using the residual U-Net; Experiments in this table are conducted on the RealSR(V3)-Canon-LR dataset due to the availability of ground-truth PSFs.; Bold numbers indicate the best performance in each column.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cho, C.; Kim, C.; Sull, S. PIABC: Point Spread Function Interpolative Aberration Correction. Sensors 2025, 25, 3773. https://doi.org/10.3390/s25123773

AMA Style

Cho C, Kim C, Sull S. PIABC: Point Spread Function Interpolative Aberration Correction. Sensors. 2025; 25(12):3773. https://doi.org/10.3390/s25123773

Chicago/Turabian Style

Cho, Chanhyeong, Chanyoung Kim, and Sanghoon Sull. 2025. "PIABC: Point Spread Function Interpolative Aberration Correction" Sensors 25, no. 12: 3773. https://doi.org/10.3390/s25123773

APA Style

Cho, C., Kim, C., & Sull, S. (2025). PIABC: Point Spread Function Interpolative Aberration Correction. Sensors, 25(12), 3773. https://doi.org/10.3390/s25123773

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop