1. Introduction
Adaptive optics (AO) is an optoelectronic technique that compensates for wavefront distortion in optical systems. This technique was initially developed in the field of astronomical observation [
1,
2] and later applied to retinal imaging. AO retinal imaging is a widely used technique to explore the structure and function of the living retina [
3]. By compensating for wavefront distortions during real-time imaging, AO enables us to examine the architecture of the retina at a cellular scale. However, AO retinal images inherently contain information from both in-focus and out-of-focus planes, leading to resolution degradation in AO flood images. Therefore, accurate interpretation of these images requires advanced post-processing techniques. From a mathematical perspective, retinal imaging can be modeled as a three-dimensional (3D) convolution. When the real object demonstrates translational invariance along the optical axis, the original 3D model can be simplified to a two-dimensional (2D) representation.
In conventional 2D image restoration, the point spread function (PSF) is generally fully known or treated as completely unknown [
4]. However, AO retinal image restoration presents a distinctive scenario where the true PSF (and the blurring matrix A) is only partially known. This intermediate regime between fully known and completely unknown PSFs is formally classified as a mildly blind or myopic deconvolution problem. Specifically, the global PSF of AO retinal images can be represented through a model that combines unknown parameters and a few PSFs. This combination tends to create coupled terms, making the problem more complex. Therefore, this poses a challenge to the existing methods for AO retinal image restoration problems [
5]. The central difficulty lies in the inherent composite nature of the global PSF, which requires simultaneous parameter estimation during image restoration.
The mathematical formulation of myopic deconvolution for AO retinal imaging can be expressed as
where
denotes an observed image with noise,
represents additive noise, and
represents unknown real images.
is the blurring matrix defined by PSF, and
is an unknown parameter in the blurring matrix
.
is constructed as a linear combination of
p known blurring matrices
, formulated as
Regularization methods play a central role in solving such deconvolution problems. As a classical technique in image denoising, total variation (TV) regularization [
6] effectively balances noise suppression and detail preservation through its edge-preserving diffusion mechanism. TV regularization has demonstrated robust performance across various noise environments, including Gaussian noise, Poisson noise, and other complex noise scenarios [
7,
8,
9,
10]. To address the AO retinal image denoising challenge, Chen et al. developed a novel model incorporating isotropic TV regularization [
5]
where
denotes the data fidelity term.
denotes the isotropic TV regularizer, which has symmetry properties or rotational invariance.
denotes the weighting parameter to balance the two terms in the objective function,
satisfies certain constraints. The TV model achieves simultaneous optimization of image restoration and unknown parameter estimation.
However, the primary limitation of TV regularization lies in its tendency to produce step artifacts. To address this issue, numerous enhanced regularization methods have been developed, including fractional-order TV (FOTV) [
11,
12], higher-order TV (HOTV) [
13,
14,
15], non-local TV (NLTV) [
16,
17], total generalized variation (TGV) [
18,
19], total variation with overlapping group sparse (OGSTV) [
20,
21,
22], and so on. Among them, Selesnick et al. proposed OGSTV [
22] as an extension of TV, which has been successfully applied to image denoising in recent years by considering the group sparsity feature of the signal derivatives and utilizing the correlation of neighboring gradient values to retain the image detail information, which effectively reduces the step artifacts.
Specifically, the OGSTV regularizer divides the image gradient field into overlapping
pixel blocks (
K is a fixed constant) and extends them in all directions centered on the pixel blocks. This ensures that every gradient is jointly constrained by multiple overlapping groups. Mathematically, for the two-dimensional gradient operators,
and
, OGSTV constructs corresponding overlapping group matrices
and computes the sum of
norm of gradients within each group, enforcing the global consistency of local smoothness. The OGSTV regularizer has been widely applied to different noise removal scenarios. Liu et al. effectively eliminated salt-and-pepper noise using OGSTV [
23]. Yin et al. enhanced impulse noise suppression by integrating OGSTV with an
norm data fidelity term [
24]. Li et al. achieved high-quality restoration of Gaussian noise-degraded images through a hybrid framework combining generalized nonconvex nonsmooth fourth-order total variation with OGS regularization [
25].
Inspired by the AO retinal image model proposed by Chen et al. in [
5], and the better restoration effect of OGSTV in image restoration, we propose the image myopic deconvolution model with OGSTV regularization term
where
denotes the OGSTV regularization function. This is a non-convex model. To the best of our knowledge, there have been no attempts to apply OGSTV to AO retinal imaging to simultaneously perform image restoration and parameter estimation in the coupled problem.
Since the proposed model is non-convex, we adopt the alternating direction method of multipliers (ADMM) as the outer-layer optimization strategy. While ADMM was originally designed for 2-block convex optimization problems, recent theoretical advances have demonstrated its empirical effectiveness in handling non-convex objective functions or non-convex sets [
26,
27,
28,
29,
30,
31]. This motivates our application of ADMM to the AO retinal image restoration task. In the concrete steps, the computational architecture employs the ADMM as the outer-layer optimization framework. According to the special structure in our proposed AO retinal image restoration model, we also use the majorization–minimization (MM) method [
24] and the linearize and project (LAP) method [
32]. Among them, the MM method deals with the OGSTV regularizer that appears during each ADMM iteration, while the LAP method solves the tightly coupled subproblem.
The main contributions of this paper are outlined below:
- (i)
We propose a new myopic deconvolution model with an OGSTV regularization term arising from AO retinal image restoration. The proposed model achieves superior image restoration and enhanced parameter estimation accuracy.
- (ii)
To address the nonconvex nature of the model and the tight coupling between variables and , we propose an ADMM-MM-LAP method.
- (iii)
We conduct comprehensive numerical experiments to evaluate the performance of our proposed OGSTV model.
The following outlines the organization of this paper. First, we propose an AO retinal imaging model with the OGSTV regularization term in
Section 2. In
Section 3, we present the iterative framework of ADMM and provide a concise overview of the MM method and the LAP method. Then, we propose the ADMM-MM-LAP method. Theoretically, we also analyze the computational complexity.
Section 4 provides the numerical experiments, and
Section 5 offers the concluding remarks.
3. Optimization Schemes
In this section, we employ the alternating direction method of multipliers (ADMM) as an outer-layer optimization framework to solve the proposed model (
6). The ADMM subproblems are efficiently solved by two strategies: the majorization–minimization (MM) method is applied to address the OGSTV component, while a modified version of the linearize and project (LAP) method is implemented to address the tightly coupled subproblem. Based on the strategies above, we propose an ADMM-MM-LAP method.
3.1. ADMM Scheme
Considering the structure of the model (
6), we apply ADMM as an outer-layer algorithm to solve it. By introducing the artificial vector
, we rewrite (
6) in the following equivalent optimization problem
where
represents the indicator function
The augmented Lagrangian function for (
7) is given by
where
represents the Lagrangian multiplier and
represents the penalty parameter.
The main idea of ADMM for solving (
7) is to alternately update variables by minimizing the augmented Lagrangian function (
8). Then, we treat
as a block of variables and
as another block of variables. Given an initial point
, ADMM iteratively solves the following
-subproblem and
-subproblem
First, the
-subproblem can be formulated as
Second, for the
-subproblem,
let
and define
Then, the (
)-subproblem can be formulated as
Finally, we update the Lagrangian multiplier
Next, we will solve these two subproblems according to their respective structural characteristics.
3.2. The Majorization–Minimization Algorithm (MM)
To tackle the
-subproblem (
9), we adopt the MM method in [
24]. Before providing the solution to (
9), we specifically explore a more generalized problem formulation, i.e.,
where
and the function
is given by (
4). Rather than directly optimizing the complicated problem (
9), we employ the MM method to iteratively solve it. Specifically, we simply look for a sequence of convex functions:
. The principle underlying the MM method is that each
serves as an upper bound on
for any
and
and attains
at
. The updated iteration yields a reduced cost. However, the complexity within the MM framework lies in constructing an appropriate function
.
Notice that the inequality
for all
and
with equality when
. By plugging each group of
into (
13) and aggregating the results, we obtain a majorizor of
. Let
with
, provided that
. By collating,
can be reformulated as
where
represents a constant, which is independent on
, and
denotes a diagonal matrix whose diagonal elements are defined as
with
,
, the entries of
can be efficiently computed via convolution operation.
Then, the majorizer of
can be expressed as
We have
for any
,
, and
. In order to obtain the optimal
from
, the MM iterations can be performed as
where
I represents the identity matrix of the same dimension as
. We would like to point out that the inversion of
can be computed efficiently, as it requires only straightforward computations of its components. To sum up, Algorithm 1 shows the iterative format of the MM method for (
12).
Algorithm 1 The MM method for (12) |
- 1:
Input: . - 2:
Output: Set . - 3:
Initialize: , , , group size K, maximum iterations N. - 4:
Iterates - 5:
Calculate according to ( 15). - 6:
Calculate according to ( 16). - 7:
- 8:
until iteration N is satisfied.
|
3.3. The Linearize and Project (LAP)
To solve the
-subproblem (
10), we employ a modified version of the linearize and project (LAP) method. The original LAP framework is by Herring et al. [
32] for solving coupled optimization problems, incorporating fundamental principles derived from the Gauss–Newton method. By using linearization and projection steps to deal with coupled variables, this method can effectively deal with tightly coupled optimization problems and complex problems with pointwise boundary constraints.
The main step of the LAP method is to linearize the residuals first, and then eliminate a set of variables by projection operation to obtain the approximation problem. The variables can be classified into two distinct groups: active set variables and inactive set variables. Subsequently, the update steps
and
can be calculated within both sets. First, we formulate the set of feasible indicators as
The active set and inactive set are defined as
respectively.
Then, the variables can be divided into active set variables , and inactive set variables . The update step can be represented as and , respectively.
The computation method of
and
is similar to the unconstrained situation. It follows the idea of Newton’s method, the update steps
are given by
That is,
where
and
.
and
denote the Jacobian matrices with respect to the
block and the
block, respectively. However, it should be noted that the variables must be projected onto the inactive set. To be specific, the value of
at the present iteration point
can be computed by
where
,
denote
,
restricted to the inactive set, respectively. The reduced problem (
17) does not require a high-precision solution. For instance, in [
32], a stopping tolerance of
is employed for solving the reduced problem iteratively. Once the value of
has been obtained through the solving process, the value of
can then be calculated by
For the active set,
and
can be obtained through
where
,
denote the projection of
,
onto the active set, respectively.
Subsequently,
and
can be computed as a combination of
,
,
, and
in the following equation
the appropriate parameter
can be selected based on [
33].
After that, the update steps are integrated through the projection gradient descent method, with the active set and inactive set being iteratively updated to guarantee the point-wise boundary constraints. This comprehensive approach enables the effective handling of large-scale optimization problems with tightly coupled variables and point-wise boundary constraints. The LAP method is used to solve the
-subproblem inexactly. The residual is
For the stopping criterion, let
where
is a non-negative sequence satisfying
.
In summary, from the above analysis, we provide the iteration scheme for solving (
12) in Algorithm 2.
Algorithm 2 The LAP method for (12) |
- 1:
Input: . Let . - 2:
Output: . - 3:
Iterates - 4:
Apply the LAP method to compute and according to ( 17) and ( 18), respectively. - 5:
Apply the projection gradient descent method to compute and according to ( 19). - 6:
Combine according to ( 20). - 7:
Compute and with a projected Armijo line search. - 8:
Update the active set and the inactive set . - 9:
. - 10:
until the termination condition is satisfied.
|
3.4. ADMM-MM-LAP Method
Based on the representations and analysis above, we present the iterative scheme of the ADMM-MM-LAP method for solving the AO retinal image model (
7) with the OGSTV regularization term in Algorithm 3.
Algorithm 3 ADMM-MM-LAP method for (7) |
- 1:
Input: and parameters , group size K. - 2:
Output: . - 3:
Initialize: . - 4:
Apply Algorithm 1 to compute . - 5:
Apply Algorithm 2 to compute (). - 6:
Update the multipliers using ( 11). - 7:
. - 8:
until the termination condition is satisfied.
|
3.5. Complexity Analysis
In this subsection, the computational cost generated at each step of Algorithm 3 is considered, and the computational complexity analysis is given.
First, the -subproblem can be solved by Algorithm 1. Each iteration needs to compute the diagonal matrix , and the computational complexity of each iteration is by calculating diagonal elements through convolution operation. Then, the matrix inverse calculation can be computed efficiently by simple element-by-element calculation, and the overall computational complexity remains . Moreover, the computational cost is also related to the group size K, and K is a constant. Therefore, based on the above analysis, the complexity of -subproblem is .
Second, the -subproblem can be efficiently solved using Algorithm 2, whose computational cost comprises several main subcomponents, such as . The computation primarily involves three core components: matrix–vector product, vector norm computation, and conjugate gradient method. Then, we proceed with a detailed step-by-step complexity analysis. First, for computing the update step , it is necessary to calculate the product of matrix , vectors and , respectively. Due to the Jacobi matrix and each PSF matrix, supports vector multiplication via the Fast Fourier Transform (FFT), the computational complexity of multiplying with a vector is . Similarly, multiplying with a vector also achieves complexity by FFTs. So the matrix–vector product involving has the same complexity. Moreover, the computational cost for the matrix–vector product of is . Furthermore, the operations involving the expressions and can be computed via FFTs with a complexity of . Thus, the overall computational cost of is . Similarly, for , the computations of their matrix–vector products also demand acceleration by FFTs, such as , ,, etc. Therefore, the computational cost of calculating is .
In addition, update steps and mainly involve scalar–vector multiplication and vector summation calculation, while the projection Armijo line search involves the calculations of and the inner product. The computational cost is . In summary, the -subproblem complexity is .
Finally, in the Lagrangian multiplier update step, the computational cost is dominated by the calculations of and , each requiring operations. Therefore, we can conclude the following theorem that the computational complexity per iteration of Algorithm 3 is .
Theorem 1. The computational complexity of Algorithm 3 is .
4. Numerical Experiments
In this section, we present the numerical experimental results of the proposed model, which incorporates the OGSTV regularization term for myopic deconvolution of AO retinal images. All experimental results were implemented using MATLAB 2020a on a PC with 16.0 GB RAM and AMD Ryzen 5 4600U with Radeon Graphics at 2.10 GHz.
The restoration performance is evaluated by five principal metrics, including the relative errors of and , signal-to-noise ratio , peak signal-to-noise ratio , and structural similarity index . Specifically, the metrics are defined as follows:
, where denotes the true image and denotes the restored image.
, where is the real parameter, and is the calculated parameter.
, where denotes the true image, denotes the mean intensity value of , and denotes the restored image.
, where refers to the size of the image, and denote the true image and the restored image, respectively.
, where and are constants. denotes the covariance between the images and . , , and , denote the average and standard variance of the AO retinal image and , respectively.
The quantitative evaluation of retinal image restoration requires multiple metrics. Although PSNR provides computationally efficient distortion measurements via mean square error (MSE), its correlation with clinical diagnostic relevance remains limited due to the non-linear nature of human visual perception. In contrast, SSIM demonstrates superior alignment with subjective evaluations by modeling luminance, contrast, and structural correlations. For the comprehensive analysis, we employ both metrics alongside relative errors to ensure mathematical rigor and parameter estimation.
The parameters are carefully configured through experimental verification. In the OGSTV model, the group size
K is set to 3 for ensuring restoration performance [
12]. We set the parameter
so that the obtained parameter meets
. The parameter
is chosen as 9. In order to guarantee the non-negative sequence
satisfying
, we choose
, with
. We implement a unified stopping criterion
where
represents the function value at the
-th iteration. The maximum number of iterations for the ADMM is set to 50. For the MM method, the inner iteration number
N is set to 10.
In the context of AO retinal imaging, we use the TV model combined with the ADMM-LAP algorithm as the benchmark method [
5]. In the following numerical experiments, we present quantitative comparisons between the proposed OGSTV model and the classical TV model, including relative errors for both reconstructed images and estimated parameters, as well as SNR, PSNR, and SSIM.
4.1. Example 1
For the myopic deconvolution problem of AO retinal images,
is established in this example. The global PSF is synthesized by combining two distinct components: a focused PSF and a defocused PSF. The test image is a
segment extracted from an AO retinal image. The test problem is constructed using the regularization toolbox IR Tools [
34]. It can produce different BlurLevel indicators (‘mild’, ‘medium’, ‘severe’), and simulate spatially invariant Gaussian blur and out-of-focus blur using the
function and
function in the IR Tools, respectively. For example, the
function simulates a spatially invariant Gaussian blur to construct a PSF, while the
function simulates a spatially invariant out-of-focus blur for another PSF. In the example, three cases are designed. One PSF is fixed as a spatially invariant Gaussian blur generated by the
function with a ‘mild’ blur level, whereas the other PSF is a composite of two components: a ‘mild’ Gaussian blur from
, and a defocus blur from
with three distinct levels (‘mild’, ‘medium’, and ‘severe’). Next, we choose the parameter
and set
in the stopping criterion to
. Gaussian noise is added via the
function and the noise level is 0.01. The true parameter is set to
= [0.3; 0.7]. We choose the initial value of the parameter
= [0.5; 0.5]. The initial guess
is a random image with
pixels.
As an example, when the BlurLevel of
is set to ‘medium’, comparing the image restoration effects of our proposed OGSTV model and the classical TV model, it can be seen from
Figure 1c,d that both models restored the images relatively clearly. It is not easy to judge which one of the OGSTV model and the TV model restores better by the eye only. To show the differences more clearly,
Figure 1e–h presents an enlarged fragment of the corresponding images. In
Figure 1g, the image recovered by the OGSTV model exhibits enhanced clarity and contrast, while the background in
Figure 1h demonstrates a coarser background texture and has more residual noise. This indicates that the restored image obtained by the OGSTV model is more similar to the true image, and its restoration effect is better than that of the TV model.
For a quantitative comparison of Example 1,
Figure 2 and
Table 1 enumerate the
,
, SNR, PSNR, and SSIM of the restored images. In
Figure 2, we plot the curve of
and
changing with the number of iterations. We can see that the relative errors of the OGSTV model in the recovered image and estimating parameters are all lower than those of the TV model. It shows that our OGSTV model simultaneously enhances image restoration accuracy and improves parameter recovery precision. This validates our OGSTV model’s capability to outperform the TV model in AO retinal image processing.
In
Table 1, the data comparison between the two models under three Blurlevels is detailed. Our OGSTV model outperforms the TV model in the key metrics. Specifically, our proposed OGSTV achieves consistently lower relative errors
and
than the TV model, reflecting its enhanced precision in reconstructing AO retinal images and parameter estimation. Simultaneously, the OGSTV model’s accuracy is further corroborated by its superior performance in SNR, PSNR, and SSIM levels. Although the OGSTV model exhibits slightly higher computational time compared to the TV model, the disparity remains minor. More critically, the OGSTV model recovers AO retinal images of better quality within a similar computation time. Notably, under the ‘severe’
, the OGSTV achieves a
lower
and a
higher SNR compared to the TV model. It indicates that the OGSTV model demonstrates increasing gaps in computationally demanding cases, and the image restoration effect is getting better. Therefore, it again shows that the OGSTV model outperforms the TV model in this test problem.
4.2. Example 2
The myopic deconvolution problem of AO retinal imaging is also considered, and the image size is 256 × 256. In this example, we build a combined PSF using with a ‘mild’ BlurLevel and with three BlurLevels (‘mild’, ‘medium’, ‘severe’) in IR Tools. The other PSF is constructed using only with a ‘mild’ BlurLevel. We choose the parameter and set the stopping criterion with . Add Gaussian noise with a noise level of 0.01. Let the real parameter be = [0.3; 0.7]. To initiate the process, we select an arbitrary image as the initial value and select the initial value of the parameter = , where are constants in [0, 1] and satisfy
The experimental results are consistent with the results of Example 1. In this example, the BlurLevel of
is ‘mild’. We still need to observe the enlarged fragments of the two models in
Figure 3. It can be observed that the background region in
Figure 3g retains more complete texture details with significantly reduced residual noise levels. The images in
Figure 3 further indicate that the image restored by our proposed OGSTV model resembles the true image more closely.
Figure 4 shows the relative errors of the reconstructed image and the estimated parameters against the number of iterations. As shown in
Figure 4, although the TV model achieves a marginally lower image’s relative error at the fourth iteration, the OGSTV model demonstrates superior stability. Under identical stopping criteria, the OGSTV model attains a significantly smaller final error after additional iterations. Notably, the parameter estimation error of the OGSTV model converges to a negligible magnitude.
In
Table 2, a detailed comparison of the data obtained from the two methods under three distinct Blurlevels is presented. The OGSTV model demonstrates consistent superiority over the TV model in both image reconstruction quality and parameter estimation accuracy across all tested conditions. Meanwhile, the OGSTV framework achieves significantly higher values in key quantitative metrics, including SNR, PSNR, and SSIM. Although the TV model exhibits marginally shorter computational times in some blur scenarios, the OGSTV model maintains a better balance between computational efficiency and restoration precision.
4.3. Example 3
In this example, we examine the scenario where the parameter p is set to 3. We use the same AO retina image as in Example 1, with a portion of 256 × 256. In Example 3, the first PSF is constructed by (’mild’). The second PSF is a combination of (’mild’) and (’medium’), and the third PSF is a combination of (’mild’) and (‘mild’, ‘medium’, and ‘severe’) combination. The parameter is chosen as , and the parameter in the stopping criterion is set to . We add Gaussian noise with a noise level of 0.01. Let the real parameter = be a random quantity with We select a random image to serve as the initial value . The initial value of the parameter is choosen as where are constants in [0, 1] and satisfy
In
Figure 5, we show the corresponding resulting image when the BlurLevel of
is ‘severe’. A larger
p means that the AO image denoising is also more complex. In Example 3, the image restoration is significantly more computationally challenging than the previous two examples. By observing the images, unlike Example 1 and 2, we do not necessitate observing enlarged images to evaluate the image restoration. Just by observing
Figure 5c–d, it is obvious that the images recovered by the OGSTV model are closer to the real image than those recovered by the TV model. This demonstrates the superior denoising capability of the OGSTV model, particularly when processing images with intricate structures.
In
Figure 6, we plot the relative errors against the number of iterations for both models. As in Example 1 and 2, the OGSTV model outperforms the TV model in both the image restoration and parameter estimation tasks. In addition, the TV model terminates the operation at the fifth iteration, while the OGSTV model can perform more iterations and generate smaller relative error values in each iteration.
Table 3 presents the data comparison between the two models at three different Blurlevels. The results indicate a correlation between increased blur and higher computational complexity, as evidenced by the rise in processing time across both models. Although the TV model has some advantages in terms of computation time, its key indexes are significantly degraded under high blur conditions. In contrast, the OGSTV model showed significant advantages, with better performance in relative error, SNR, PSNR, and SSIM, which fully verified the validity and reliability of the proposed OGSTV model in AO retinal image restoration. In conclusion, our proposed OGSTV model outperforms the TV model in AO retinal image restoration in all test cases.
4.4. Performance Comparison
To evaluate the performance differences between the proposed OGSTV model and the TV model, we show the PSNR variations of both models during the iterative process across the above three examples. The corresponding curves of PSNR versus iteration numbers are shown in
Figure 7. The results demonstrate that our proposed OGSTV model exhibits significant convergence and achieves higher PSNR levels compared to the TV model.
Moreover, we would like to point out that the OGSTV model requires substantial computational resources for processing high-resolution images. Consequently, in the aforementioned simulations, OGSTV exhibited longer processing times compared to TV in some examples. Future research will prioritize the development of lightweight variants of OGSTV to reduce the computational cost while preserving the reconstruction accuracy.