Next Article in Journal
SVDDD: SAR Vehicle Target Detection Dataset Augmentation Based on Diffusion Model
Previous Article in Journal
Three-Dimensional Reconstruction of Space Targets Utilizing Joint Optical-and-ISAR Co-Location Observation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Deep Hyperspectral Inpainting with Plug-and-Play and Deep Image Prior Models

1
School of Engineering, University of Edinburgh, Edinburgh EH8 9YL, UK
2
Institute of Imaging, Data and Communications, University of Edinburgh, Edinburgh EH8 9YL, UK
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(2), 288; https://doi.org/10.3390/rs17020288
Submission received: 6 December 2024 / Revised: 9 January 2025 / Accepted: 10 January 2025 / Published: 15 January 2025
(This article belongs to the Section AI Remote Sensing)

Abstract

:
Hyperspectral images are typically composed of hundreds of narrow and contiguous spectral bands, each containing information regarding the material composition of the imaged scene. However, these images can be affected by various sources of noise, distortions, or data loss, which can significantly degrade their quality and usefulness. This paper introduces a convergent guaranteed algorithm, LRS-PnP-DIP(1-Lip), which successfully addresses the instability issue of DHP that has been reported before. The proposed algorithm extends the successful joint low-rank and sparse model to further exploit the underlying data structures beyond the conventional and sometimes restrictive unions of subspace models. A stability analysis guarantees the convergence of the proposed algorithm under mild assumptions, which is crucial for its application in real-world scenarios. Extensive experiments demonstrate that the proposed solution consistently delivers visually and quantitatively superior inpainting results, establishing state-of-the-art performance.

1. Introduction

Hyperspectral remote sensing has found widespread application in numerous fields such as astronomy, agriculture, environmental monitoring, and Earth observation. Hyperspectral images (HSIs) are often captured via satellite or airborne sensors, each presenting samples at different time slots, using push-broom strategies along the flying pathway [1]. The nature of the HSI acquisition system makes the HSI a high-resolution 3D data cube that covers hundreds or thousands of narrow spectral bands, conveying a wealth of spatio-spectral information. However, due to instrumental errors, imperfect navigation and atmospheric variations, practical HSIs can suffer from noise, and missing pixels or even entire lines of pixels (an example of this type of missing data can be found in the recently released Earth Surface Mineral Dust Source Investigation (EMIT) hyperspectral dataset, developed by NASA’s Jet Propulsion Laboratory and launched on 14 July 2022 [2]). These issues can severely impact the accuracy and reliability of subsequent analyses, making HSI inpainting a critical task in the field of remote sensing and Earth observation.

1.1. Hyperspectral Image Inpainting

Hyperspectral image inpainting refers to the restoration of missing or corrupted data in acquired HS images. Unlike RGB images, for which inpainting involves filling a single pixel value (red, green, and blue channels), HSI inpainting requires filling in a complex vector that contains extensive spectral information. This additional complexity makes the already challenging task of HSI inpainting even more difficult. The primary objective of HSI inpainting is to create visually convincing structures while ensuring that the texture of the missing regions is spectrally coherent. The inpainting of HS images typically involves two main steps: (a) estimating the missing or corrupted data by leveraging information from the neighboring spectral bands and spatially adjacent pixels and (b) refining the estimated data to ensure consistency with the overall spectral and spatial properties of the HS image. Traditional approaches, such as those proposed by [3], assume that the spectral vectors of the HSI always exist in some unknown low-dimensional subspaces, and the missing regions are estimated through projections. However, the performance of such techniques deteriorates significantly when large portions of the image are missing, and they may even fail when all spectral bands are absent. Another line of research involves the tensor completion process to restore incomplete observations [4,5,6]. These methods leverage the spatio-spectral correlations of the hyperspectral data cube, offering potential improvements over subspace-based techniques.
Over the past decade, the success of deep learning has brought new opportunities for solving HSI inpainting tasks. In [7,8], the authors proposed transformer-based frameworks to effectively capture both local and global contextual information in 3D images, demonstrating the potential of attention modules in exploring relationships between missing pixels and background pixels. In [9], the authors showed that missing pixels can be predicted using generative neural networks. However, this approach has two key limitations: (1) it requires training the networks on extensive datasets to achieve the desired performance, and (2) the inpainted areas often suffer from over-smoothing, resulting in the loss of critical information, such as abrupt changes in the surface materials. More recently, researchers discovered that the network’s structure itself may serve as a good inductive bias for regularizing image reconstruction tasks, such as image inpainting [10]. A deep image prior (DIP) enables the learning of priors directly from the image without the need for extensive training datasets, which was further developed and successfully applied to hyperspectral (HS) images [11]. Following [11], subsequent works [12,13,14,15,16,17,18] have shown that certain trained or untrained neural networks can be directly incorporated into iterative solvers to improve the reconstruction accuracy, i.e., as the regularizer. More recently, the authors of [19] introduced a self-supervised framework that leverages the powerful learning capabilities of diffusion models and sets new state-of-the-art benchmarks across several HSI reconstruction tasks. However, as it was reported in our preliminary report [20], as well as in a series of DIP-related works [13,21,22,23], the DIP may not always be the ideal solution, despite promising empirical results. On the practical side, this is because a DIP works by directly fitting a neural network to the corrupted HS image; in the absence of regularization or explicit denoising mechanisms, a DIP tends to overfit, which will eventually capture the noise in the image. On the theoretical side, a DIP lacks a strong theoretical foundations to explain its performance, and its convergence is highly sensitive to random initialization of the neural network. Different initializations could potentially lead to poor results. Although subsequent works [12,13,24] have attempted to mitigate this issue by introducing a plug-and-play (PnP) denoiser into the training of a DIP, they require either a suitable early stopping criterion or the manual control of the learning rate. As a result, there remains a need for a new HSI inpainting algorithm that combines the strengths of both PnP and DIP methods while offering convergence guarantees.

1.2. Sparsity and Low Rankness in HSIs

Despite the high dimensionality, HS images exhibit intrinsic sparse and low-rank structures due to the high correlation between spatial pixels and spectral channels. This property has made sparse representation—(SR) and low rank—(LR) based methods popular for the processing of HSIs [25,26,27]. SR relies on the key assumption that the spectral signatures of the pixels approximately lie in a low-dimensional subspace spanned by representative pixels from the same class. With a given dictionary, SR allows for sparse decomposition of HSIs into a linear combination of several atoms, thus exploiting the spatial similarity of HSIs (It is worth mentioning that, when the dictionary is not given, we can manually use the end-members of some pixels with pure spectra and build the dictionary, or learn the dictionary [28] in an unsupervised manner, using sparsity as a regularizer). This assumption about the underlying structures has been effectively applied in various HSI tasks, including HSI classification [29], denoising [30], and unmixing [31].
LR is another widely used regularizer in HSI processing, which assumes that pixels in the clean HSI image have high correlations in the spectral domain, thus capturing both the spectral similarity and global structure of HSIs. LR allows for the modeling of the HSI as a low-rank matrix; thus, the missing and corrupted information in the HSIs can be effectively reconstructed by leveraging the underlying low-rank structure. In many state-of-the-art HSI reconstruction algorithms, LR is often used in parallel with SR to achieve better reconstruction accuracy [30,32]. By combining the two priors together, algorithms can better handle missing data, noise, and corruption while maintaining the overall structure and spectral integrity of the image. For more detailed information on LR and SR of hyperspectral images, we refer readers to the review work in [33].
However, a notable limitation with such an LR or SR model is that the real HS data do not exactly follow the sparse and/or low-dimension assumptions.; e.g., there are long trails (numerous small but non-zero elements) in the descending sorted absolute sparse coefficients and singular values, where overly-restrictive assumptions on these coefficients could lead to under-fitting. In conclusion, both LR and SR rely on some predefined mathematical models that assume certain properties of the HS image, such as low-dimensional manifolds (low-rankness) or sparse representations in specific domains. While these models are useful, they struggle to capture more complex patterns in the real HS data, such as non-linear relationships and complex spatio-spectral dependencies. Moreover, when a significant percentage of the data are corrupted or entire spectral bands of pixels are missing, these methods often fail [3]. The question of how deep neural models can address this issue by exploiting more data-adaptive, low-dimensional models in a self-supervised learning framework is the main topic of this study.
In this paper, we start by giving a brief overview of the LRS-PnP-DIP algorithm [20]; we replaced the denoising/thresholding step with the PnP denoiser, and we used a DIP to replace the low-rank component of the formulation. Since there remains a lack of comprehensive theoretical analysis of the DIP and its extensions that can guarantee the convergence, we demonstrate both theoretically and empirically that small modifications can ensure the convergence of LRS-PnP and LRS-PnP-DIP, while the original algorithms in [20] may diverge. We propose a variant of the LRS-PnP-DIP algorithm dubbed LRS-PnP-DIP(1-Lip), which not only outperforms existing learning-based methods but also resolves the instability issue of a DIP in an iterative algorithm.

1.3. Contributions

This paper presents a novel HS inpainting algorithm that leverages the learning capability of deep networks while providing insights into the convergence behavior of such complex algorithms. Through rigorous mathematical analysis, we derive sufficient conditions for stability and fixed-point convergence and show that these conditions are satisfied by slightly modified algorithms. Our approach offers several key contributions, including the following:
  • This paper explores the LRS-PnP and LRS-PnP-DIP algorithm in more depth by showing the following: (1) The sparsity and low-rank constraints are both important to the success of the methods in [20]. (2) DIP can better explore the low-rank subspace compared to conventional subspace models such as singular value thresholding (SVT).
  • Under some mild assumptions, a fixed-point convergence proof is provided for the LRS-PnP-DIP algorithm (see Theorem 1). We introduce a variant to the LRS-PnP-DIP called LRS-PnP-DIP(1-Lip), which effectively resolves the instability issue of the algorithm by slightly modifying both the DIP and PnP denoiser.
  • To the best of our knowledge, this is the first time the theoretical convergence of PnP with DIP replaced the low-rank prior being analyzed under an iterative framework.
  • Extensive experiments were conducted on real-world data to validate the effectiveness of the enhanced LRS-PnP-DIP(1-Lip) algorithm with a convergence certificate. The results demonstrate the superiority of the proposed solution over existing learning-based methods in both stability and performance.
The remainder of this paper is organized as follows: Section 2 provides an introduction to our approach, and Section 3 describes our main convergence theory for the proposed LRS-PnP-DIP(1-Lip) algorithm. Section 4 provides the implementation details and discusses the experimental results. Section 5 concludes the paper with a brief discussion of potential future directions.

2. Proposed Methods

Mathematical Formulations

The task of HSI inpainting can be formulated as reconstructing the clean image, X, from a noisy and incomplete measurement, Y, where an additive noise, N, and a masking operator, M, are present:
Y = M { X } + N
The clean image X R q (where q = n r × n c × n b ), with n r and n c representing the spatial dimensions of the image, and n b represents the total number of spectral bands. The operator M R q × q is a binary mask, where zero represents the missing pixels, and one represents the observed and valid pixels. Thus, the masking operator M can be represented as a diagonal square matrix. N is the additive Gaussian noise of appropriate size. Usually, M is provided, and the formulation (1) describes a linear system, which can be expressed as follows:
y = M x + n
In this equation, x , y , and n represent the vectorized forms of X, Y, and N, respectively, and M is a diagonal matrix. We obtain the recovered, i.e., inpainted, image, x * , by applying sparse representation to each image patch, P i ( x ) . Here, the operator P i ( · ) extracts the i-th patch from image x , where each patch may consist of only valid pixels or a combination of valid and missing pixels, depending on the size of P i ( · ) . The inpainted image x * can be obtained by solving the following optimization problem:
( x * , α * ) = argmin x , α γ y M x 2 2 + w l r x * + i ( P i ( x ) Φ α i ) 2 2 + w s i α i 1
The objective in the expression above consists of three terms. The first term is the data fidelity term, weighted by the parameter γ . Since estimating x from y is inherently ill-posed, i.e., with more unknown variables than equations, the solution is non-unique. To address this, we introduce two additional “priors” to regularize the inpainting problem: low-rank and sparsity constraints. The second term penalizes solution x to be low-rank, which is typically employed as a surrogate for rank minimization. The third term restricts the missing pixels to be generated from the subspace approximated by the valid pixels. These terms are weighted by parameters w l r and w s , respectively. The sparse representation problem is solved using a given dictionary, Φ . Specifically, Φ can be constructed either by using the end-members of some pixels with pure spectra or by learning from the noisy pixels in the observations. It is here exclusively learned from noisy pixels in the observations using online dictionary learning [28]. By adopting the augmented Lagrangian and introducing the auxiliary variable u [34], problem (3) can be rewritten as follows:
( x * , α * ) = argmin x , α γ y M x 2 2 + w l r u * + w s i α i 1 + μ 1 2 i ( P i ( x ) Φ α i ) + λ 1 μ 1 2 2 s . t . x = u
The Lagrangian multiplier and penalty terms are denoted as λ 1 and μ 1 , respectively. Using the alternating direction method of multipliers (ADMM), we can sequentially update the three variables α , u , and x to solve problem (4):
(1) Fixing u and x , and updating α:
α k + 1 = argmin α μ 1 k 2 i ( P i ( x k ) + λ 1 k μ 1 k ) Φ α i 2 2 + w s i α i 1
The problem in Equation (5) is a patch-based, sparse coding problem that can be solved using iterative solvers. Let us denote the first term on the right-hand side of Equation (5) as f ( α ) and the second term as g ( α ) while dropping the subscript for simplicity. Problem (5) can be represented as a constrained optimization problem:
( α ^ , v ^ ) = argmin α , v f ( α ) + w s g ( v ) s . t . α = v
By introducing an auxiliary variable, u , and a multiplier, ρ , the augmented Lagrangian of (6) becomes the following:
L ( α ^ , v ^ , u ^ ) = f ( α ) + w s g ( v ) + u T ( α v ) + ρ 2 α v 2 2
Then, ADMM finds solutions by breaking the constrained optimization problem (6) into several sub-problems and updating them separately in an iterative fashion. The pseudocode is provided in Algorithm 1.
The v k + 1 sub-problem can be solved by the proximal operator [35]. The authors in [36] realized that α k + 1 + u k can be treated as a “noisy” version of v , and g ( v ) can be seen as a regularization term. Then, the whole process resembles a denoising operation applying to the intermediate results, which can be simply replaced with the PnP denoisers such as BM3D and the non-local mean (NLM) denoiser. In our approach, we employed PnP-ISTA [36] as an alternative solution for solving problem (5). We denote the gradient of f as f and I as an Identity matrix with an appropriate shape. The entire process can then be replaced with an off-the-shelf denoiser D , operating on f [36]. Each iteration takes the following form:
α k + 1 = D ( I f ) ( α k )
PnP-ISTA can significantly improve the reconstruction quality over traditional ISTA [37] due to the following reasons: (1) it enables the integration of state-of-the-art denoisers to handle different noise patterns and different types of images, which is particularly flexible, as one can apply any off-the-shelf denoiser without massively changing the existing framework; and (2) it allows the use of denoisers tailored to different problems (e.g., a denoiser that promotes sparsity or smoothness, or even a denoising neural network that encodes more complex priors).
Algorithm 1 Pseudocode of ADMM for solving problem (6).
Initialization  w s , ρ , k 0 , α k 0 , u k 0 , v k 0 .
while Not Converged do
     α k + 1 argmin α f ( x ) + ρ 2 α ( v k u k ) 2 2 .
     v k + 1 argmin v w s g ( v ) + ρ 2 v ( α k + 1 + u k ) 2 2 .
     u k + 1 1 ρ u k + ( α k + 1 v k + 1 ) .
end while
(2) We can update u in this setting by:
u k + 1 = argmin u w l r u * + μ 2 k 2 ( x k + λ 2 k μ 2 k ) u 2 2
where λ 2 and μ 2 are the corresponding Lagrangian multiplier and penalty term, respectively, which we use subscripts to distinguish between λ 1 and μ 1 in step (4). The problem can be solved using the singular value thresholding (SVT) algorithm [38] as follows:
u k + 1 = SVT ( x k + λ 2 k μ 2 k )
Specifically, denote w l r μ 2 k as τ , matrix x k + λ 2 k μ 2 k as A , and the soft thresholding operator S τ as follows:
S τ ( x ) = sign ( x ) · max ( | x τ | , 0 )
where s i g n ( · ) is an operator taking the positive part of its input. Applying (11) to the singular values of A promotes low-rankness:
u k + 1 = SVT ( A ) = U A S τ ( Σ A ) V A T
In the proposed LRS-PnP-DIP algorithm [20], this step is substituted with a deep image prior (DIP) f θ ( z ) , where θ denotes the network parameters that need to be updated, and the input z is set to x k + λ 2 k μ 2 k ; i.e., the latent image from the previous iterations:
u k + 1 = f θ ( x k + λ 2 k μ 2 k )
(3) Fixing α and u , and updating x :
x k + 1 = argmin x γ y M x 2 2 + i ( P i ( x ) + λ 1 k μ 1 k ) Φ α i k + 1 2 2 + μ 2 k 2 ( x + λ 2 k μ 2 k ) u k + 1 2 2
A closed-form solution for x exists as follows:
x k + 1 = ( γ M T M + μ 1 k i P i T P i + μ 2 k I ) 1 ( γ M T y + μ 1 k i P i Φ α i k + 1 + μ 2 k u k + 1 i P i λ 1 k λ 2 k I )
In the expression above, the patch selection operator P i ( · ) is rewritten in its matrix form P i . To calculate x k + 1 in practice, we can leverage the fact that, for the HSI inpainting task, both M T M and P i T P i are diagonal matrices. As a results, the matrix inversion of ( γ M T M + μ 1 k i P i T P i + μ 2 k I ) 1 can be simply implemented as an element-wise division. This trick significantly reduces the computational cost of the inversion, making the proposed algorithm more scalable for practical use.
(4) The Lagrangian and penalty terms are updated as follows:
λ 1 k + 1 = λ 1 k + μ 1 k ( x k + 1 Φ α k + 1 ) λ 2 k + 1 = λ 2 k + μ 2 k ( x k + 1 u k + 1 )
μ 1 k + 1 = ρ 1 μ 1 k μ 2 k + 1 = ρ 2 μ 2 k
The LRS-PnP-DIP algorithm proposed in [20] is summarized in Algorithm 2.
Algorithm 2 (LRS-PnP-DIP) algorithm.
Require: masking matrix: M , noisy and incomplete HSI: y , learned dictionary: Φ . denoiser: D . max iteration: I t m a x . DIP: f θ
Output: inpainted HSI image x .
Initialization DIP parameters θ , λ 1 , λ 2 , μ 1 , μ 2 , ρ 1 , ρ 2 .
while Not Converged do
     for i = 1 : I t m a x do:
    α k + 1 = D ( I f ) ( α k )
     update DIP parameters θ , with target: y and intermediate results: x k + λ 2 k μ 2 k as input.
     update x by (15).
     update Lagrangian parameters and penalty terms.
end while
  • Remarks. In contrast to its original form, it is here proposed to replace the conventional PnP denoiser in (8) with an averaged NLM denoiser, as introduced in [39]. Additionally, we modify the f θ ( z ) in (13) to obtain Lipschitz continuity with Lipschitz constant 1. The averaged NLM denoiser is a variant of NLM denoiser whose weight is designed to be a doubly symmetric matrix, ensuring that both the columns and rows sum to 1. With this doubly symmetric property, the spectral norm of the weight matrix is bounded by 1, meaning the averaged NLM denoiser is by design, non-expansive. (We refer the reader to [39] for detailed structures of the averaged NLM denoiser).
The enhanced algorithm is denoted as LRS-PnP-DIP(1-Lip), which is shown in Algorithm 3, and its flow chart is provided in Figure 1.
Algorithm 3 LRS-PnP-DIP(1-Lip) algorithm.
Require: masking matrix: M , noisy and incomplete HSI: y , learned dictionary: Φ . averaged NLM Denoiser: D . max iteration: I t m a x . 1-Lipschitz DIP: f θ
Output: inpainted HSI image x .
Initialization DIP parameters θ , λ 1 , λ 2 , μ 1 , μ 2 , ρ 1 , ρ 2 .
while Not Converged do
     for i = 1 : I t m a x do:
    α k + 1 = D ( I f ) ( α k )
     update DIP parameters θ , with target: y and intermediate results: x k + λ 2 k μ 2 k as input.
     update x by (15).
     update Lagrangian parameters and penalty terms.
end while

3. Convergence Analysis

In this section, we describe the fixed-point convergence of the proposed LRS-PnP-DIP(1-Lip) algorithm under mild assumptions. Fixed-point convergence refers to the type of convergence wherein the algorithm asymptotically enters a steady state. To establish our main theorem, we need the following results:
Definition 1
(Non-expansive operator). An operator, T : R n R n , is said to be non-expansive if, for any x , y R n , the following applies:
( T ( x ) T ( y ) ) 2 ( x y ) 2
Definition 2
( θ -averaged). An operator, T : R n R n , is said to be θ-averaged with some θ ( 0 , 1 ) if there exists a non-expansive operator, R, such that we can write T = ( 1 θ ) I + θ R .
Lemma 1.
Let T : R n R n be θ-averaged for some θ ( 0 , 1 ) . Then, for any x , y R n , the following applies:
( T ( x ) T ( y ) ) 2 ( x y ) 2 1 θ θ ( ( I T ) ( x ) ( I T ) ( y ) 2
Proof of this lemma can be found in [[40], Lemma 6.1].
Definition 3
(Fixed point). We say that x * R n is a fixed point of the operator T : R n R n if T ( x * ) = x * . We denote the set of fixed points as fix(T).
Definition 4
( β -smoothed). Let f : R n R n be differentiable; we say that f is β-smooth if there exists β > 0 , such that f ( x ) f ( y ) β x y for any x , y R n .
Definition 5
(Strong convexity). A differentiable function, f, is said to be strongly convex with modulus ρ > 0 if f ( x ) ρ 2 x 2 is convex.
Lemma 2
(Property of strong convexity). Let f be strongly convex with modulus ρ > 0 . Then, for any x , y R n , the following applies:
f ( x ) f ( y ) , x y ρ ( x y ) 2
Proof of this Lemma can be found in [41].
  • We made the following assumptions:
Assumption 1.
We assume that (1) the denoiser D used in the sparse coding step is linear and θ-averaged for some θ ( 0 , 1 ) , (2) the function f ( α ) = μ 1 k 2 ( x k + λ 1 k μ 1 k ) Φ α 2 2 is β-smoothed, and (3) fi x ( D ( I f ) ) .
  • Remarks. The θ -averaged property is a subset of the non-expansive operator that most of the existing PnP frameworks have worked with [39,42]. In fact, the non-expansive assumption is found to be easily violated for denoisers such as BM3D and NLM denoisers when used in practice. Nevertheless, the convergence of PnP methods using such denoisers can still be empirically verified. In the experiments, we adopted the modified NLM denoiser whose spectral norm is bounded by 1. Hence, it satisfies the θ -averaged property by design [39].
Lemma 3
(Fixed-point convergence of PnP sparse coding). If Assumption 1 holds, then, for any α, and ρ > β / 2 , the sequence ( α ) k 0 k generated via step (8) converges to some α * fi x ( D ( I f ) ) .
Proof of this Lemma can be found in [[40], Theorem 3.5].
Assumption 2.
We assume that the DIP function f θ is L-Lipschitz-bounded:
( f θ ( x ) f θ ( y ) ) 2 L ( x y ) 2
for any x, y, and L 1 .
  • Remarks. In the above assumption, the smallest constant L that makes the inequality hold is called the Lipschitz constant. The Lipschitz constant is expressed as the maximum ratio between the absolute change in the output and the input. It quantifies how much the function output changes with respect to the input perturbations or, roughly speaking, how robust the function f θ is. In deep learning, it. Assumption 2 guarantees that the trained DIP has a Lipschitz constant, L 1 , which is a desirable feature so that a network does not vary drastically in response to minuscule changes in its input. In the experiments, this can be achieved by constraining the spectral norm of each layer of the neural network during training.
We are now ready to state our main theorem.
Theorem 1
(Convergence of LRS-PnP-DIP(1-Lip) in the Lyapunov Sense). If both Assumptions 1 and 2 hold, with penalty μ, and with an L-Lipschitz constrained DIP ( L 1 ), then there exists a non-increasing function, H k = 2 x k x * 2 + 1 μ 2 λ 1 k λ 1 * 2 + 1 μ 2 λ 2 k λ 2 * 2 , such that all trajectories generated via LRS-PnP-DIP(1-Lip) are bounded, and that as k , x k x * 2 0 , α k α * 2 0 , and u k u * 2 0 ; i.e., even if the equilibrium states of x , α, and u are perturbed, they will finally converge to x * , α * , and u * , respectively. The proposed LRS-PnP-DIP(1-Lip) algorithm is, thus, asymptotically stable.

4. Experimental Results

4.1. Implementation Details

We evaluate the proposed inpainting model on two publicly available hyperspectral datasets:
  • The Chikusei airborne hyperspectral dataset [43] (the link to the Chikusei dataset can be found at https://naotoyokoya.com/Download.html, accessed on 4 January 2025); the test HS image consists of 192 channels, and it was cropped to 36 × 36 pixels in size.
  • The Indian Pines dataset from AVIRIS sensor [44]; the test HS image consisted of 200 spectral bands, and it was cropped to 36 × 36 pixels size.
For each dataset, we trained the dictionary Φ with a size of 1296 × 2000, using only noisy and incomplete HSI images (the size of the test HS image and the dictionary were chosen to match the practical use case where the proposed inpainting algorithms were to be mounted on each nano-satellite to process different small image tiles in parallel [45]), this is a necessary step if we do not have access to pure spectra. If a standard spectral dataset exists, learning can be performed in the form of a pre-training step. However, we used the input HS image for this task to demonstrate that the proposed algorithms operate in a self-supervised setting. Additionally, we introduced Gaussian noise with noise level σ = 0.12 to the cropped HS images. Since the main focus of this paper is image inpainting, instead of denoising, σ was kept fixed across all experiments. We applied mask M to all the spectral bands in the given region, which represents the most challenging case. For the choice of the PnP denoiser, we used both the BM3D and modified NLM denoisers [39]; the latter is an averaged denoiser, which enjoys a non-expansive property that is crucial to convergence analysis.
Our implementation of the deep image prior model follows the same structure as in the deep hyperspectral prior [11]. During the training of 1-Lipschitz DIP, we applied Lipschitz regularization to all layers, as in [46]. In each iteration, we fed the DIP with the latent image from the previous iteration, rather than the random signal, as suggested by [11]. We did not perturb the input to support our convergence analysis, which is different from the canonical DIP. To eliminate the need for manually selecting the total number of iterations for the DIP model, we adopted the early stopping criterion proposed in [47] with a fixed window size of 20 and a patience number of 100, which detects the near-peak MPSNR point using windowed moving variance (WMV). Note that, for the training of conventional DIP, the early-stopping strategy is often employed to prevent it from fitting to random noise after certain iterations. However, the proposed LRS-PnP-DIP(1-Lip) algorithm mitigates the over-fitting issue due to the imposed constraint. The early-stopping criterion is applied primarily to optimize the algorithm’s runtime so that one does not have to wait for the algorithm to reach the maximum number of iterations. In our formulation, the low-rank and sparsity constraints in the equation were weighted with parameters w l r and w s , respectively, while the data fidelity term was weighted with γ . Initially, we set w s / w l r to 1 and γ to 0.5. Notably, the choice of γ is highly dependent on the noise level of the observed image. If the noise level is low, the recovered image, X, should be similar to the noisy observation, Y, parameter γ should be large, and vice versa. The Adam optimizer was used, and the learning rate was set to 0.1. Finally, we used three widely used metrics, namely the mean signal-to-noise ratio (MPSNR), mean structural similarity (MSSIM), and mean spectral angle mapper (MSAM), to evaluate the performance of the model in all experiments.

4.2. Low-Rank vs. Sparsity

In this section, we examine the effectiveness of both sparsity and low-rank constraints in our proposed LRS-PnP algorithm. Figure 2 shows the inpainting performance of LRS-PnP over different τ for different masks (in percentages), where τ is defined as the ratio of the weight of the sparsity constraint and the low-rank constraint. For example, τ = 0 means that the sparsity constraint is disabled, and mask = 50% means that 50% of the pixels are missing. We can conclude from the results that (1) the LRS-PnP algorithm works better with both low-rank and sparsity constraints, as using either an overwhelming sparsity constraint or a low-rank constraint alone tends to reduce inpainting performance, and (2) when the percentage of missing pixels increases, more weight should be placed on the sparsity constraint in order to achieve high-quality reconstruction. On the one hand, approximation errors may be introduced in the sparse coding step; i.e., step (5) in Algorithms 2 and 3. These errors can potentially be suppressed by the constraint of low rank. On the other hand, the sparsity constraint exploits the spectral correlations; i.e., the performance will be enhanced if spatial information is considered.

4.3. Low Rankness Due to DIP

In the original LRS-PnP-DIP Algorithm [20], we propose solving the low-rank minimization problem using DIP instead of the traditional SVT projection. A natural question arises: does the DIP (e.g., the architecture proposed in Figure 1) effectively capture the redundancies in the spectral domain? This is empirically justified in Figure 3, where we trained DIP on the corrupted HS image with varying numbers of spectral bands as input and output. The results show that the inpainting performance of DIP improved substantially when more spectral bands were utilized during training; the peak MPSNR was obtained when the neural network saw the entire 128 bands of the input HS images, rather than each single band individually. This finding is counterintuitive since DIP with 2D convolution is generally expected to handle only the spatial correlation of input HSIs. Similar observations have been reported in [10,48], where the authors confirmed that DIP with 2D convolution implicitly explored the channel correlations, possibly due to the upsampling layers in the decoder part that manipulates the HSI channels in the spectral domain.
In Figure 4, we sort the singular values of the inpainted image of LRS-PnP (the pink line), LRS-PnP-DIP [20] (the blue line), and PnP sparse coding without a low-rank constraint (green line), in descending order. Among these algorithms, LRS-PnP performs singular value projection to promote low rankness, while it is replaced with DIP in LRS-PnP-DIP. It is observed that using a well-trained DIP alone was sufficient to capture these low-rank details. Thus, we suggest replacing the conventional SVT process with DIP. Although there is no theoretical evidence on how DIP mimics the SVT, we show through experiments that the DIP can well preserve some small yet important singular values that would otherwise be discarded via SVT projection. Hence, it has the ability to better exploit the low-rank subspace in a data-driven fashion, possibly owing to its highly non-linear network structures and inductive bias.

4.4. Convergence

The use of the PnP denoiser and DIP in the context of HSIs has been recently investigated in [12,13]. However, theoretical evidence of the stability of the algorithm is still lacking. To bridge this gap, we verify the convergence of LRS-PnP-DIP(1-Lip). In Figure 5, the top left, top right, and bottom left figures represent the successive differences of each state on a log scale, and the bottom right figure represents the inpainting performance. For LRS-PnP-DIP, we used the BM3D denoiser and DIP without constraints. For LRS-PnP-DIP(1-Lip), we used the modified NLM denoiser and the 1-Lipschitz constraint DIP. The latter automatically satisfies Assumptions 1 and 2; thus, the fixed-point convergence can be guaranteed. Interestingly, we noticed that LRS-PnP-DIP with BM3D and conventional DIP still works pretty well in practice even though there are fluctuations in state x , as can be seen in the top right and bottom left plots in Figure 5. We deduced that such instability mainly comes from the process of solving DIP sub-problems, rather than the denoising sub-problems, as the primal variable λ 1 smoothly converges with BM3D denoisers as well.
From the bottom right plot, we observed that the best results are obtained with LRS-PnP-DIP in terms of reconstruction quality, and the LRS-PnP-DIP(1-Lip) is only slightly lower than its unconstrained counterpart. Our observations on the reduced performance of Lipschitz DIP are in agreement with existing works [46,49], where imposing the Lipschitz constraint during training does hurt the capacity and expressivity of the neural network.

4.5. Comparison with State of the Art

The proposed algorithms were compared with learning-based algorithms: DHP [11], GLON [4], R-DLRHyIn [16], DeepRED [22], PnP-DIP [24], DeepHyIn [50], and DDS2M [19]. To ensure the fairness of comparison, we employed the same U-net as the backbone for DHP, DeepRED, R-DLRHyIn, PnP-DIP, and LRS-PnP-DIP(1-Lip) and the attention-based U-net with the same complexity for DeepHyIn and DDS2M. For GLON, we replaced its original FFDNet with the one trained on the hyperspectral dataset [51] and kept the others intact. (The reason for making this adaption is that the neural network proposed in GlON [4] was trained on RGB/grayscale images. It was found to be beneficial to use an FFDNet, e.g., in [51], trained on the HSI dataset.) For DeepHyIn, we used the end members extracted from the clean image patch. For DDS2M, we fine-tuned the diffusion-related parameters, as suggested in [19]. The statistical results of each test sample are averaged over 20 experiments to account for the effect of random seeds during training.
In Figure 6, we plot the recovered spectrum of the center of the missing area of LRS-PnP-DIP(1-Lip) for each method. It can be seen that LRS-PnP-DIP(1-Lip) produces a more consistent and realistic spectrum in the missing region compared to methods such as GLON, DHP, DeepRED, R-DLRHyIn, DDS2M, and DeepHyIn. We present the visual inpainting results for each method on the Chikusei dataset and the Indian Pines dataset, as depicted in Figure 7 and Figure 8, respectively. The quantitative statistical results for each dataset, evaluated under various mask shapes, are summarized in Table 1 for the Chikusei dataset and Table 2 for the Indian Pines dataset.
It is shown that LRS-PnP-DIP(1-Lip) can effectively suppress noise compared to methods such as R-DLRHyIn, DeepHyIn, and DDS2M. Moreover, it alleviates the over-smoothing phenomenon of the DeepRED and PnP-DIP algorithms by preserving more details in the non-missing region; this is clear for the last two test samples in Figure 7 and for the first four test samples in Figure 8. From Table 1 and Table 2, it can be seen that LRS-PnP-DIP(1-Lip) almost always demonstrates the highest inpainting performance on both the Chikusei and Indian Pines datasets compared to other methods. The second best is established via DDS2M, which is the closest method to ours in terms of inpainting quality. This is possibly due to its ability to leverage inductive bias from hierarchical diffusion structures. However, as is evident in Figure 7 and Figure 8, DDS2M and other DIP-based extensions occasionally exhibit limitations, such as producing noisy backgrounds or failing to effectively inpaint missing strips. It is important to note that these methods utilize conventional DIP as backbones; hence, they are susceptible to the instability issue of DIP where the algorithm may eventually fit into pure noise without proper constraints. In contrast, LRS-PnP-DIP(1-Lip) is superior, as it not only harvests the strong learning capability of DHP but also enjoys a stability guarantee. For the HSI inpainting task, we anticipate that some information, from the background/non-missing pixels to be carried to the deeper layers, there is a performance loss when removing the skip connections between the encoder and the decoder. Nevertheless, LRS-PnP-DIP(1-Lip) mutes these layers to trade performance for convergence, and it is empirically found that, even if LRS-PnP-DIP(1-Lip) is highly restricted due to layer-wise spectral normalization, it can preserve more textures and produce less noisy results than DDS2M. This observation leads to the conclusion that the proposed LRS-PnP-DIP(1-Lip) algorithm would be preferable for solving practical inpainting tasks compared to existing solutions. It has the ability to yield faithful reconstructions while ensuring stability and robustness.

5. Conclusions

The LRS-PnP algorithm, along with its variant LRS-PnP-DIP proposed in the previous work [20], are innovative hyperspectral inpainting methods designed to address the challenge of missing pixels in noisy and incomplete HS images. These techniques can better handle the challenging situations in practical HSI acquisition systems where the entire spectral band may be absent. Through experiments, we show that both sparsity and the low-rank priors can be effectively integrated as constraints during the training of DIP; the resulting LRS-PnP-DIP algorithm leverages the strong learning capability of the powerful generative DIP model and is able to exploit spectral and spatial redundancies inherent to HSIs. Because the proposed methods do not require any training data, they are particularly attractive for practical hyperspectral inpainting tasks, where only the observed corrupted image is available for reference. However, the LRS-PnP-DIP algorithm suffers from instability issues and may occasionally diverge due to the DIP. In this paper, this long-lasting issue was effectively addressed through the proposed LRS-PnP-DIP(1-Lip) algorithm. Experimental results on real datasets demonstrate that the LRS-PnP-DIP(1-Lip) algorithm not only resolves the instability issue but also delivers results competitive with the state-of-the-art HSI inpainting algorithms. Unlike previous works where the convergence of the algorithm is either missing or only empirically verified, such as [12,13,14,24,52], this work has provided theoretical evidence and established convergence guarantees under non-expansiveness and Lipschitz continuity assumptions on the PnP denoiser and the DIP neural network. However, the performance of DIP has been observed to be sensitive to the network structure [11] and to the training parameters, such as the number of iterations and the learning rate. When the proposed LRS-PnP-DIP(1-Lip) algorithm is given an insufficient number of iterations to learn, DIPs may not work as expected. In future work, we would like to explore the inductive bias of DHPs, as well as a broader variety of deep learning models, such as transformers (the transformer has recently been explored in the remote sensing community [53]) to exploit their potential in solving HSI inpainting problems. Additionally, the extension of the proposed methods to other HSI processing tasks, such as self-supervised HSI denoising, unmixing, and deblurring, is left to a future investigation.

Author Contributions

Conceptualization, S.L. and M.Y.; methodology, S.L. and M.Y.; software, S.L.; validation, S.L. and M.Y.; formal analysis, S.L. and M.Y.; investigation, S.L. and M.Y.; resources, S.L.; data curation, S.L.; writing—original draft preparation, S.L; writing—review and editing, S.L. and M.Y.; visualization, S.L. and M.Y.; supervision, M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The code for reproducing results is available at https://github.com/shuoli0708/LRS-PnP-DIP, accessed on 14 January 2025. The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Appendix A.1. On the Design of 1-Lipschitz DIP

In the proposed LRS-PnP-DIP(1-Lip) algorithm, we redesigned the DIP to have a Lipschitz constant of 1. To do so, we started by taking a deeper look at how the Lipschitz constant is calculated for each layer; then, we showed that the Lipschitz constraint can be enforced to construct a 1-Lipschitz DIP.

Appendix A.1.1. Convolutional Layer

The convolutional layer c o n v ( f ) works by performing the convolutional operation on the input to extract the features. The output of the convolution operation can be expressed as follows:
c o n v ( f ) = W · f
where W is a double-block circular matrix. Substituting (A1) into the definition of the Lipschitz constant, we have the following:
( W x 1 W x 2 ) 2 L ( x 1 x 2 ) 2 W ( x 1 x 2 ) 2 L ( x 1 x 2 ) 2 W x 2 L ( x ) 2 ( Let x = x 1 x 2 , x 1 x 2 )
It follows that
W ( x ) 2 x 2 L or equivalently , L smallest = sup x 0 W x 2 x 2 = W 2
The above expression reveals the fact that the spectral norm of W is bounded by the Lipschitz constant L. Hence, regularizing the convolutional layer c o n v ( f ) to have Lipschitz continuity boils down to constraining the matrix W ; e.g., in our case, W is expected to have all of its singular values be at most 1, meaning that c o n v ( f ) is 1-Lipschitz. To design such a matrix, W , we adopted the power iteration method, as suggested in [46]. The power iteration method works iteratively and approximates the largest singular value of W . Given a sufficient number of iterations, it converges to the true spectral norm of W with almost negligible errors. In practice, this can be achieved with only a few iterations, which is quite efficient compared to the SVD method.

Appendix A.1.2. Skip Connection Layer

The DIP has skip connections between the encoder and the decoder. These skip/residual connections ensure anuninterrupted gradient flow, which can effectively tackle the vanishing gradient problem. Moreover, it allows some previous features to be reused later in the network. In particular, for our HSI inpainting problem, we expect some information about the clean pixels from the background to be carried to the next layers.
  • The skip connections take the following form:
s k i p ( f ) = f + W ( f )
where W stands for the composition of possible different functions or layers (e.g., fully connected layers and activation functions). Since the skip connections are additive operators, the Lipschitz constant is determined by both of its components. Suppose that functions f 1 and f 2 , are Lipschitz continuous with the Lipschitz constants L 1 and L 2 , respectively. Then, the composition f = f 1 + f 2 satisfies the following:
f ( x 1 x 2 ) = ( f 1 ( x 1 ) + f 2 ( x 1 ) ) ( f 1 ( x 2 ) + f 2 ( x 2 ) ) 2 = ( f 1 ( x 1 ) f 1 ( x 2 ) ) + ( f 2 ( x 1 f 2 ( x 2 ) ) 2
Applying the triangular inequality to the L 2 norm, we have the following:
( f 1 ( x 1 ) f 1 ( x 2 ) ) + ( f 2 ( x 1 f 2 ( x 2 ) ) 2 f 1 ( x 1 ) f 1 ( x 2 ) 2 + f 2 ( x 1 ) f 2 ( x 2 ) 2
Furthermore, the right-hand side of inequality (A6) can be bounded by the Lipschitz constant of functions f 1 and f 2 as follows:
f 1 ( x 1 ) f 1 ( x 2 ) 2 + f 2 ( x 1 ) f 2 ( x 2 ) 2 L 1 x 1 x 2 2 + L 2 x 1 x 2 2 = ( L 1 + L 2 ) x 1 x 2 2
We can conclude that the Lipschitz constant of f = f 1 + f 2 is bounded by the sum of the Lipschitz constants of f 1 and f 2 . Therefore, for the skip connections defined in (A4), the Lipschitz constant is bounded by 1 + L W , with a non-negative L W . Because the skip connections would have a Lipschitz constant exceeding 1, we removed them from the implementation of the 1-Lipschitz DIP.

Appendix A.1.3. Pooling Layer

The pooling layer p o o l ( f ) works by dividing the input image into a set of sub-image patches and converting them to a single value. We used max-pooling in the encoder to extract feature maps from the input HSIs. Max pooling takes the maximum values of the neurons and discards the rest:
p o o l ( f ) = m a x ( f 1 , f 2 , f 3 f n )
where n denotes the number of image patches. The expression above can be seen as an affine transformation with a fixed weight; i.e.,
p o o l ( f ) = W · f W i = 1 if f i = m a x ( f 1 , f 2 , f 3 f n ) 0 otherwise
The gradient of W is 1 for the neurons with the maximum value and 0 otherwise; the latter does not participate in the propagation. Hence, by definition, the max pooling layer is Lipschitz-continuous with a Lipschitz constant of 1.

Appendix A.1.4. Activation Layer

The Lipschitz constant of the activation layer a c t ( f ) is trivial, as most activation functions such as ReLu, Leaky-ReLu, Tanh, and Sigmoid have Lipschitz constants of 1.

Appendix A.1.5. Batch Normalization Layer

The batch normalization layer B N ( f ) works by scaling and normalizing the inputs so as to mitigate issues related to internal covariance shifts. It is defined as follows:
B N ( f ) = γ · f E ( f ) Var ( f ) + b
In the above formulation, E ( f ) is the mini-batch mean, and Var ( f ) is the mini-batch variance. f is normalized, scaled by γ , and shifted with respect to the bias b. γ and b are two learned parameters during training, which forces the network to also learn the identity transformation and the ability to make a decision on how to balance them. From (A10), it is evident that B N ( f ) is an affine transform with the transformation matrix W being as follows:
W = diag γ Var ( f )
Therefore, similar to the analysis of the convolutional layers, the Lipschitz constant of the batch normalization layer is the spectral norm of W . However, B N ( f ) typically normalizes its input by the variance and scales it by the factor γ ; this would inevitably destroy the Lipschitz continuity. Hence, we propose a variant to the batch normalization layer as follows:
B N ( f ) = L · ( f E ( f ) ) + b
The modified BN operator B N centers its input by subtracting the mean, which is quite different from its original form, (A10). In the experiments, we observed that such an adaptation did not affect the performance of the DIP.

Appendix A.1.6. Lipschitz Constant of the Full Network

Consider a neural network, f n e t , which is a composition of n independent sub-networks, f 1 , f 2 , f n , each of which has its own Lipschitz constant, L 1 , L 2 , L n :
f n e t ( x ) = ( f 1 f 2 f n ) ( x )
Then, the Lipschitz constant of f n e t can be upper-bounded by the following:
L n e t L 1 · L 2 · · L n = i = 1 n L i
Hence, we can tackle the Lipschitz constant of each layer independently and combine them to construct the upper bound of the entire network. We note that the bound certified in (A14) is not tight and that there are spaces for establishing a tighter bound by considering the network as a whole, which we will leave for future work.

Appendix A.1.7. Enforcing Lipschitz Constraint

Inspired by the work in [46], we introduce an extra projection step after each weight update as follows:
P ( W , L ) = 1 max ( 1 , W 2 / L ) W
where L stands for the desired Lipschitz constant, particularly L = 1 for 1-Lipschitz DIP. This step aims to project the weight matrix, W, back to the feasible set for those layers that violate the Lipschitz constraint.

Appendix A.2. Proof of Theorem 1

The proof relies on the Lyapunov stability theory, which is the heart of the dynamic system analysis [54]. Lyapunov stability theory can be categorized into (a) the indirect method, which analyzes the convergence through the system state equation, and (b) the direct method, which explicitly describes the behavior of the system’s trajectories and its convergence by making use of the Lyapunov function. We refer the reader to [55] for a more detailed definition of the Lyapunov function (specifically, Theorem 1.2 and Theorem 3.3). In some contexts, it is also known as an energy or dissipative function [56]. Compared to the former, the direct method is more appealing, as the convergence can be established by only showing the existence of such a function. In this proof, we begin by defining a function for our proposed LRS-PnP-DIP(1-Lip) algorithm, and we prove its validity. That is, we construct an energy function, H k , that describes the behaviors of each states in Algorithm 3, and to prove that H k is a non-increasing function as iteration proceeds; hence, the total energy of the dynamic system established via Algorithm 3 is conservative and stable.
  • Let H k = 2 x k x * 2 + 1 μ 2 λ 1 k λ 1 * 2 + 1 μ 2 λ 2 k λ 2 * 2 , for a non-zero μ .
  • Remarks. This design follows similar structures as in the original convergence proof of the ADMM algorithm [34] and in a recent work [57]. Here, H k is a function of the system’s state change, which is, by design, non-negative. The first two assumptions in Theorem 1.2 [55] automatically hold. Thus, we only need to show that the proposed candidate H k is a non-increasing function in order to be a valid Lyapunov function. More specifically, we will show that H k is a decreasing function, which satisfies the following:
H k H k + 1 C
where C is a positive constant.
Firstly, recall that the LRS-PnP-DIP(1-Lip) algorithm takes the following update steps:
α k + 1 = T ( x k + λ 1 k μ )
u k + 1 = f θ ( x k + λ 2 k μ )
x k + 1 = argmin x y M x 2 2 + μ 2 ( x + λ 1 k μ ) Φ α k + 1 2 2 + μ 2 ( x + λ 2 k μ ) u k + 1 2 2
λ 1 k + 1 = λ 1 k + μ ( x k + 1 Φ α k + 1 ) λ 2 k + 1 = λ 2 k + μ ( x k + 1 u k + 1 )
We define x e k = x k x * , u e k = u k u * , α e k = α k α * , λ 1 e k = λ 1 k λ 1 * , and λ 2 e k = λ 2 k λ 2 * in the subsequent proof for simplicity. This results in the following:
x e k + 1 = x k + 1 x * u e k + 1 = u k + 1 u * α e k + 1 = α k + 1 α * λ 1 e k + 1 = λ 1 k + 1 λ 1 * λ 2 e k + 1 = λ 2 k + 1 λ 2 *
We take Equation (A19) as our starting point and denote the first term, y M x 2 2 , as f ( x ) . The first-order optimality of equation (A19) implies the following:
f ( x ) + μ ( x + λ 1 k μ Φ α k + 1 ) + μ ( x + λ 2 k μ u k + 1 ) = 0
Since the minimizer x k + 1 satisfies (A22), we plug it into (A22) to obtain the following:
f ( x ) k + 1 + μ ( x k + 1 Φ α k + 1 + λ 1 k μ ) + μ ( x k + 1 u k + 1 + λ 2 k μ ) = ( A20 ) f ( x ) k + 1 + μ ( λ 1 k + 1 λ 1 k μ + λ 1 k μ ) + μ ( λ 2 k + 1 λ 1 k μ + λ 2 k μ ) = f ( x ) k + 1 + λ 1 k + 1 + λ 2 k + 1 = 0
The critical point satisfies, i.e., k :
f ( x ) * + λ 1 * + λ 2 * = 0
Due to the strong convexity of f ( x ) , using Lemma 2 with x = x k + 1 and y = x * yields the following:
f ( x ) k + 1 f ( x ) * , x k + 1 x * ρ ( x k + 1 x * ) 2
Combining the above inequality with Equations (A23) and (A24), we have the following:
λ 1 e k + 1 λ 2 e k + 1 , x e k + 1 ρ x e k + 1 2
Secondly, using Assumption 2, the DHP f θ ( z ) is L-Lipschitz with L 1 and x = x k + λ 2 k μ , y = x * + λ 2 * μ ; we get the following:
u k + 1 u * 2 ( x k x * ) + λ 2 k λ 2 * μ 2 o r   u e k + 1 2 x e k + λ 2 e k μ 2
Thirdly, using Assumption 1 and the resulting Lemma 1, suggesting that the operator T used in the α updating step (A17) is θ -averaged, and x = x k + λ 1 k μ , y = x * + λ 1 * μ , we have the following:
α k + 1 α * 2 ( x k x * ) + λ 1 k λ 1 * μ 2 o r α e k + 1 2 x e k + λ 1 e k μ 2
Now, we multiply (A26) on both sides by 2 μ , and we gather the resulting inequality with (A27) and (A28):
2 μ λ 1 e k + 1 , x e k + 1 2 μ λ 2 e k + 1 , x e k + 1 2 μ ρ x e k + 1 2 x e k + λ 2 e k μ 2 u e k + 1 2 x e k + λ 1 e k μ 2 α e k + 1 2
We put μ inside the left-hand side of the first inequality, and we add them to give the following:
2 λ 1 e k + 1 μ , x e k + 1 2 λ 2 e k + 1 μ , x e k + 1 + x e k + λ 2 e k μ 2 + x e k + λ 1 e k μ 2 2 μ ρ x e k + 1 2 + u e k + 1 2 + α e k + 1 2
which can be written as:
( 1 μ 2 λ 1 e k + 1 2 + x e k + 1 2 λ 1 e k + 1 μ x e k + 1 2 ) ( 1 μ 2 λ 2 e k + 1 2 + x e k + 1 2 λ 2 e k + 1 μ x e k + 1 2 ) + 1 μ 2 λ 2 e k 2 + x e k 2 + 2 λ 2 e k μ , x e k + 1 μ 2 λ 1 e k 2 + x e k 2 + 2 λ 1 e k μ , x e k 2 μ ρ x e k + 1 2 + u e k + 1 2 + α e k + 1 2
After rearrangement, we get the following:
2 ( x e k 2 x e k + 1 2 ) + 1 μ 2 ( λ 1 e k 2 λ 1 e k + 1 2 ) + 1 μ 2 ( λ 2 e k 2 λ 2 e k + 1 2 ) 2 μ ρ x e k + 1 2 + u e k + 1 2 + α e k + 1 2 + λ 1 e k + 1 μ x e k + 1 2 + λ 2 e k + 1 μ x e k + 1 2 2 λ 2 e k μ , x e k 2 λ 1 e k μ , x e k
Recall that
H k H k + 1 = 2 x k x * 2 + 1 μ 2 λ 1 k λ 1 * 2 + 1 μ 2 λ 2 k λ 2 * 2 2 x k + 1 x * 2 1 μ 2 λ 1 k + 1 λ 1 * 2 1 μ 2 λ 2 k + 1 λ 2 * 2 = ( ) 2 ( x e k 2 x e k + 1 2 ) + 1 μ 2 ( λ 1 e k 2 λ 1 e k + 1 2 ) + 1 μ 2 ( λ 2 e k 2 λ 2 e k + 1 2 )
It can be seen that the left-hand side of inequality (A32) recovers exactly H k H k + 1 . To make H k a non-increasing function, we require the entire right-hand side to be non-negative; it is thus sufficient to show that the last two terms, 2 λ 2 e k μ , x e k 2 λ 1 e k μ , x e k , are non-negative. This is straightforward to show if we plug k = k 1 into the first line of (A29) to get the following:
H k H k + 1 2 μ ρ x e k + 1 2 + u e k + 1 2 + α e k + 1 2 + λ 1 e k + 1 μ x e k + 1 2 + λ 2 e k + 1 μ x e k + 1 2 + 2 μ ρ x e k 2 0
Now, if we add both sides of (A33), from k = 0 to k = , it follows that
k = 0 2 μ ρ x e k + 1 2 + k = 0 u e k + 1 2 + k = 0 α e k + 1 2 H 0 H <
We can conclude that sequences x e k + 1 , u e k + 1 and α e k + 1 are all bounded sequences due to the Lyapunov theorem. That is, as k :
lim k x k + 1 x * 2 0 lim k u k + 1 u * 2 0 lim k α k + 1 α * 2 0
Thus, the iterations generated via the LRS-PnP-DIP(1-Lip) algorithm converge to the critical points ( x * , u * , α * ) with a sufficiently large k, and all the trajectories are bounded. LRS-PnP-DIP(1-Lip) is also asymptotically stable according to Theorem 3.3 [55].

Appendix A.3. Parameters Tuning and Ablation Tests

Appendix A.3.1. Sensitivity Analysis and Hyperparameters Selections

In this section, we provide the detailed selection of hyperparameters for the proposed LRS-PnP-DIP(1-Lip) algorithm in the experiments. In Table A2, we provide the inpainting performance of LRS-PnP-DIP(1-Lip) with the weights of low-rank constraint w l r and sparsity constraints w s varying within the range of 0.1 to 1. It can be seen that the optimal performance is obtained when both low-rank and sparsity constraints are present, as long as there is no overwhelming of either w l r or w s . This indicates that LRS-PnP-DIP(1-Lip) is insensitive to the precise value of the low-rank constraint w l r or sparsity constraints w s , which is different from the LRS-PnP algorithm [20] discussed in Section 4.2. Therefore, we set both w l r and w s to 1 across all experiments.
Table A1. The selection of hyperparameters for the proposed LRS-PnP-DIP(1-Lip) algorithm.
Table A1. The selection of hyperparameters for the proposed LRS-PnP-DIP(1-Lip) algorithm.
Parameters Number of Iterations γ w l r w s μ 1 μ 2 λ 1 λ 2 σ y
2001/2111/21/2000.12
Parameters DIP Learning Rate DIP perturbed Noise Level Lipschitz Constant : L PnP-ISTA Iterations λ I S T A                             
       0.101500.1                            
Table A2. Inpainting MPSNR of LRS-PnP-DIP(1-Lip) algorithm with varying w l r and w s . Using either overwhelming w l r or w s could deteriorate the inpainting results.
Table A2. Inpainting MPSNR of LRS-PnP-DIP(1-Lip) algorithm with varying w l r and w s . Using either overwhelming w l r or w s could deteriorate the inpainting results.
w l r 00.10.250.50.751
w s
030.69237.479(±0.63)38.121(±0.46)38.475(±0.48)38.686(±0.52)38.884(±0.31)
0.137.00440.849(±0.31)39.107(±0.35)39.321(±0.37)39.443(±0.30)39.812(±0.34)
0.2537.21140.355(±0.29)40.798(±0.36)40.266(±0.29)39.770(±0.23)39.906(±0.31)
0.537.59239.766(±0.27)40.807(±0.24)40.831(±0.26)40.532(±0.25)40.245(±0.32)
0.7537.66339.897(±0.37)40.373(±0.33)40.565(±0.27)40.807(±0.30)40.780(±0.26)
137.95738.591(±0.46)39.248(±0.31)39.848(±0.25)39.902(±0.25)40.835(±0.28)

Appendix A.3.2. Computational Efficiency

Table A3 provides per-iteration running time and the total running time upon converge for each method discussed in Section 4.5. To ensure a fair comparison, all methods, except for GLON [4], which uses a pre-trained FFDNet, were implemented with the same DIP backbone. All experiments were conducted in Python with PyTorch 1.13 using the NVIDIA GeForce RTX 3090 GPU. It can be seen that GLON suggests the fastest inference time, but at the cost of sub-optimal inpainting performance. DHP [11], R-DLRHyIn [16], and DeepHyIn [50] share similar computational complexity; these are followed by DDS2M [19], which typically requires a large number of diffusion steps in order to yield faithful reconstruction. On the other hand, methods such as DeepRED [22], PnP-DIP [24] and LRS-PnP-DIP(1-Lip) are computationally more expensive, which is mainly due to the use of PnP denoisers as regularizers at each iteration. Although LRS-PnP-DIP(1-Lip) introduces extra computation complexity to solve the sparsity-constrained problem (5), its overall running time is competitive with that of the DeepRED and PnP-DIP methods, This is mainly because LRS-PnP-DIP(1-Lip) enjoys a stability guarantee, which makes it converge faster with fewer iterations.
Table A3. Different algorithms and their running times on both the Chikusei and Indian Pines datasets.
Table A3. Different algorithms and their running times on both the Chikusei and Indian Pines datasets.
DatasetCostGLON [4]DHP [11]R-DLRHyIn [16]DeepHyIn [50]DDS2M [19]DeepRED [22]PnP-DIP [24]LRS-PnP-DIP(1-Lip)
Chikuseiper-iteration0.1380.0830.0890.1020.1142.4122.5423.622
all iterations69.30289.596106.85595.351382.027482.985473.064394.601
Indian Pinesper-iteration0.1040.1090.1150.1360.1492.7322.9673.905
all iterations74.74197.215115.541103.344401.654494.088474.156425.601

Appendix A.3.3. Effect of the Lipschitz Constraint DIP

As an extension to the LRS-PnP-DIP algorithm, LRS-PnP-DIP(1-Lip) enforces the Lipschitz continuity (A14) of DIP by applying the Lipschitz constraint to each layer of the neural network independently with Lipschitz constant L = 1 . In the proposed LRS-PnP-DIP(1-Lip) algorithm, we remove the residual connections from the DIP (as was discussed in Appendix A.1.2). However, changing from Skip-Net to U-Net may potentially lead to reduced performance, which is part of the trade-off with the convergence. In Figure A1, we provide an ablation test comparing the performance of LRS-PnP-DIP with/without 1-Lipschitz constraint. It can be seen that 1-Lipschitz constraint promotes the stability of the reconstruction process, ensuring the convergence of the algorithm. This improvement comes at the cost of only a slight reduction in inpainting performance.
Figure A1. Inpainting performance of LRS-PnP-DIP algorithm with/without 1-Lipschitz constraint.
Figure A1. Inpainting performance of LRS-PnP-DIP algorithm with/without 1-Lipschitz constraint.
Remotesensing 17 00288 g0a1

Appendix A.3.4. Effect of the Averaged Denoiser

Compared to the conventional NLM denoiser, the averaged property of the modified NLM denoiser imposes a significantly more restrictive condition. Such a modification also raises the question as to whether it will damage the performance of the original one. In Table A4, we record the inpainting MPSNR and MSSIM of the LRS-PnP-DIP algorithm [20] with different PnP denoisers; the LRS-PnP-DIP with averaged NLM denoiser achieves almost the same performance as with the BM3D or the conventional NLM denoiser.
Table A4. Inpainting performance of LRS-PnP-DIP algorithm with different PnP denoisers, under noise level σ y = 0.1 , 0.15 , 0.2 , respectively. The mean and variance over 20 samples are shown here. The best results are highlighted in bold.
Table A4. Inpainting performance of LRS-PnP-DIP algorithm with different PnP denoisers, under noise level σ y = 0.1 , 0.15 , 0.2 , respectively. The mean and variance over 20 samples are shown here. The best results are highlighted in bold.
MethodMetricInputBM3D DenoiserNLM DenoiserAveraged NLM Denoiser
σ y = 0.1 MPSNR ↑31.7641.23(±0.25)41.25(±0.30)41.22(±0.13)
              MSSIM ↑0.3040.920(±0.01)0.918(±0.01)0.920(±0.01)
σ y = 0.15 MPSNR ↑30.3839.55(±0.29)39.60(±0.22)39.55(±0.220)
MSSIM ↑0.2470.913(±0.01)0.915(±0.01)0.914(±0.01)
σ y = 0.2 MPSNR ↑28.7537.68(±0.32)37.65(±0.35)37.67(±0.24)
MSSIM ↑0.2290.904(±0.01)0.903(±0.01)0.903(±0.01)

Appendix A.3.5. Effect of the DIP Network Architectures

As was reported in a series of DIP-related works [10,11,24,50], the performance of DIP is sensitive to the structure of the neural networks. Table A5 provides a comparative analysis of several network architectures, including ResNet 2D/3D, U-Net 2D/3D, and Skip-Net 2D/3D. Empirical results indicate that Skip-Net 2D (i.e., a U-Net architecture with skip connections and 2D convolution layers) achieves the best performance for the HSI inpainting task. Hence, Skip-Net 2D was selected and implemented as the backbone architecture for the proposed LRS-PnP-DIP, LRS-PnP-DIP(1-Lip), and all other competing methods in the comparative analysis. Furthermore, it was found that ResNet architectures yield inferior results within the proposed framework, highlighting the importance of network design choices for achieving optimal inpainting performance. We left the exploration and identification of the most suitable DIP network structure for HSI inpainting as a future research direction.
Table A5. On the choice of different DIP network architectures in the proposed LRS-PnP-DIP algorithm. For the metrics MPSNR and MSSIM, higher values indicate better performance. The best results are highlighted in bold.
Table A5. On the choice of different DIP network architectures in the proposed LRS-PnP-DIP algorithm. For the metrics MPSNR and MSSIM, higher values indicate better performance. The best results are highlighted in bold.
MethodsMPSNR↑MSSIM↑
Input22.5820.178
ResNet 2D30.975(±0.62)0.610(±0.02)
ResNet 3D29.661(±1.20)0.589(±0.03)
UNet 2D35.963(±0.42)0.882(±0.02)
UNet 3D35.438(±0.56)0.868(±0.01)
Skip-Net 2D 37.971 ( ± 0.30 ) 0.899 ( ± 0.01 )
Skip-Net 3D37.050(±0.32)0.890(±0.01)

References

  1. Ortega, S.; Guerra, R.; Diaz, M.; Fabelo, H.; López, S.; Callico, G.M.; Sarmiento, R. Hyperspectral push-broom microscope development and characterization. IEEE Access 2019, 7, 122473–122491. [Google Scholar] [CrossRef]
  2. EMIT L1B At-Sensor Calibrated Radiance and Geolocation Data 60 m V001. Available online: https://search.earthdata.nasa.gov/search/granules?p=C2408009906-LPCLOUD&pg[0][v]=f&pg[0][gsk]=-start_date&q=%22EMIT%22&tl=1711560236!3!! (accessed on 27 August 2024).
  3. Zhuang, L.; Bioucas-Dias, J.M. Fast hyperspectral image denoising and inpainting based on low-rank and sparse representations. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 730–742. [Google Scholar] [CrossRef]
  4. Zhao, X.L.; Yang, J.H.; Ma, T.H.; Jiang, T.X.; Ng, M.K.; Huang, T.Z. Tensor completion via complementary global, local, and nonlocal priors. IEEE Trans. Image Process. 2021, 31, 984–999. [Google Scholar] [CrossRef]
  5. Li, B.Z.; Zhao, X.L.; Wang, J.L.; Chen, Y.; Jiang, T.X.; Liu, J. Tensor completion via collaborative sparse and low-rank transforms. IEEE Trans. Comput. Imaging 2021, 7, 1289–1303. [Google Scholar] [CrossRef]
  6. Luo, Y.S.; Zhao, X.L.; Jiang, T.X.; Chang, Y.; Ng, M.K.; Li, C. Self-supervised nonlinear transform-based tensor nuclear norm for multi-dimensional image recovery. IEEE Trans. Image Process. 2022, 31, 3793–3808. [Google Scholar] [CrossRef]
  7. Wu, Y.; Liu, J.; Gong, M.; Liu, Z.; Miao, Q.; Ma, W. MPCT: Multiscale point cloud transformer with a residual network. IEEE Trans. Multimed. 2023, 26, 3505–3516. [Google Scholar] [CrossRef]
  8. Yuan, Y.; Wu, Y.; Fan, X.; Gong, M.; Ma, W.; Miao, Q. EGST: Enhanced geometric structure transformer for point cloud registration. IEEE Trans. Vis. Comput. Graph. 2023, 30, 6222–6234. [Google Scholar] [CrossRef] [PubMed]
  9. Wong, R.; Zhang, Z.; Wang, Y.; Chen, F.; Zeng, D. HSI-IPNet: Hyperspectral imagery inpainting by deep learning with adaptive spectral extraction. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4369–4380. [Google Scholar] [CrossRef]
  10. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Deep image prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9446–9454. [Google Scholar]
  11. Sidorov, O.; Yngve Hardeberg, J. Deep hyperspectral prior: Single-image denoising, inpainting, super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  12. Meng, Z.; Yu, Z.; Xu, K.; Yuan, X. Self-supervised neural networks for spectral snapshot compressive imaging. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2622–2631. [Google Scholar]
  13. Lai, Z.; Wei, K.; Fu, Y. Deep Plug-and-Play Prior for Hyperspectral Image Restoration. Neurocomputing 2022, 481, 281–293. [Google Scholar] [CrossRef]
  14. Wu, Z.; Yang, C.; Su, X.; Yuan, X. Adaptive Deep PnP Algorithm for Video Snapshot Compressive Imaging. arXiv 2022, arXiv:2201.05483. [Google Scholar] [CrossRef]
  15. Gan, W.; Eldeniz, C.; Liu, J.; Chen, S.; An, H.; Kamilov, U.S. Image reconstruction for mri using deep cnn priors trained without groundtruth. In Proceedings of the 2020 54th Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 1–4 November 2020; pp. 475–479. [Google Scholar]
  16. Niresi, K.F.; Chi, C.Y. Robust Hyperspectral Inpainting via Low-Rank Regularized Untrained Convolutional Neural Network. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  17. Hong, D.; Yao, J.; Li, C.; Meng, D.; Yokoya, N.; Chanussot, J. Decoupled-and-coupled networks: Self-supervised hyperspectral image super-resolution with subpixel fusion. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5527812. [Google Scholar] [CrossRef]
  18. Yao, J.; Hong, D.; Chanussot, J.; Meng, D.; Zhu, X.; Xu, Z. Cross-attention in coupled unmixing nets for unsupervised hyperspectral super-resolution. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Part XXIX 16. pp. 208–224. [Google Scholar]
  19. Miao, Y.; Zhang, L.; Zhang, L.; Tao, D. Dds2m: Self-supervised denoising diffusion spatio-spectral model for hyperspectral image restoration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12086–12096. [Google Scholar]
  20. Li, S.; Yaghoobi, M. Self-Supervised Hyperspectral Inpainting with the Optimisation inspired Deep Neural Network Prior. In Proceedings of the 2023 31st European Signal Processing Conference (EUSIPCO), Helsinki, Finland, 4–8 September 2023; pp. 471–475. [Google Scholar]
  21. Heckel, R.; Hand, P. Deep decoder: Concise image representations from untrained non-convolutional networks. arXiv 2018, arXiv:1810.03982. [Google Scholar]
  22. Mataev, G.; Milanfar, P.; Elad, M. DeepRED: Deep image prior powered by RED. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  23. Liu, J.; Sun, Y.; Xu, X.; Kamilov, U.S. Image restoration using total variation regularized deep image prior. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 7715–7719. [Google Scholar]
  24. Sun, Z.; Latorre, F.; Sanchez, T.; Cevher, V. A plug-and-play deep image prior. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, Canada, 6–11 June 2021; pp. 8103–8107. [Google Scholar]
  25. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  26. Zhang, H.; He, W.; Zhang, L.; Shen, H.; Yuan, Q. Hyperspectral image restoration using low-rank matrix recovery. IEEE Trans. Geosci. Remote Sens. 2013, 52, 4729–4743. [Google Scholar] [CrossRef]
  27. Li, C.; Zhang, B.; Hong, D.; Yao, J.; Chanussot, J. LRR-Net: An Interpretable Deep Unfolding Network for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–12. [Google Scholar] [CrossRef]
  28. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G. Online dictionary learning for sparse coding. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; pp. 689–696. [Google Scholar]
  29. Peng, J.; Sun, W.; Du, Q. Self-paced joint sparse representation for the classification of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2018, 57, 1183–1194. [Google Scholar] [CrossRef]
  30. Zhao, Y.Q.; Yang, J. Hyperspectral image denoising via sparse representation and low-rank constraint. IEEE Trans. Geosci. Remote Sens. 2014, 53, 296–308. [Google Scholar] [CrossRef]
  31. Iordache, M.D.; Bioucas-Dias, J.M.; Plaza, A. Sparse unmixing of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2014–2039. [Google Scholar] [CrossRef]
  32. Xue, J.; Zhao, Y.; Liao, W.; Kong, S.G. Joint spatial and spectral low-rank regularization for hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2017, 56, 1940–1958. [Google Scholar] [CrossRef]
  33. Ma, G.; Huang, T.Z.; Huang, J.; Zheng, C.C. Local low-rank and sparse representation for hyperspectral image denoising. IEEE Access 2019, 7, 79850–79865. [Google Scholar] [CrossRef]
  34. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar]
  35. Parikh, N.; Boyd, S. Proximal algorithms. Found. Trends® Optim. 2014, 1, 127–239. [Google Scholar] [CrossRef]
  36. Venkatakrishnan, S.V.; Bouman, C.A.; Wohlberg, B. Plug-and-play priors for model based reconstruction. In Proceedings of the 2013 IEEE Global Conference on Signal and Information Processing, Austin, TX, USA, 3–5 December 2013; pp. 945–948. [Google Scholar]
  37. Daubechies, I.; Defrise, M.; De Mol, C. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Commun. Pure Appl. Math. J. Issued Courant Inst. Math. Sci. 2004, 57, 1413–1457. [Google Scholar] [CrossRef]
  38. Cai, J.F.; Candès, E.J.; Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  39. Sreehari, S.; Venkatakrishnan, S.V.; Wohlberg, B.; Buzzard, G.T.; Drummy, L.F.; Simmons, J.P.; Bouman, C.A. Plug-and-play priors for bright field electron tomography and sparse interpolation. IEEE Trans. Comput. Imaging 2016, 2, 408–423. [Google Scholar] [CrossRef]
  40. Nair, P.; Gavaskar, R.G.; Chaudhury, K.N. Fixed-point and objective convergence of plug-and-play algorithms. IEEE Trans. Comput. Imaging 2021, 7, 337–348. [Google Scholar] [CrossRef]
  41. Ekeland, I.; Temam, R. Convex Analysis and Variational Problems; SIAM: Philadelphia, PA, USA, 1999. [Google Scholar]
  42. Ryu, E.; Liu, J.; Wang, S.; Chen, X.; Wang, Z.; Yin, W. Plug-and-play methods provably converge with properly trained denoisers. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 5546–5557. [Google Scholar]
  43. Yokoya, N.; Iwasaki, A. Airborne Hyperspectral Data over Chikusei; SAL-2016-05-27; University of Tokyo: Tokyo, Japan, 2016. [Google Scholar]
  44. Baumgardner, M.F.; Biehl, L.L.; Landgrebe, D.A. 220 band AVIRIS hyperspectral image data set: June 12, 1992 Indian Pine test site 3. Purdue Univ. Res. Repos. 2015, 10, 991. [Google Scholar]
  45. Denby, B.; Lucia, B. Orbital edge computing: Nanosatellite constellations as a new class of computer system. In Proceedings of the Twenty-Fifth International Conference on Architectural Support for Programming Languages and Operating Systems, Lausanne, Switzerland, 16–20 March 2020; pp. 939–954. [Google Scholar]
  46. Gouk, H.; Frank, E.; Pfahringer, B.; Cree, M.J. Regularisation of neural networks by enforcing lipschitz continuity. Mach. Learn. 2021, 110, 393–416. [Google Scholar] [CrossRef]
  47. Wang, H.; Li, T.; Zhuang, Z.; Chen, T.; Liang, H.; Sun, J. Early Stopping for Deep Image Prior. arXiv 2021, arXiv:2112.06074. [Google Scholar]
  48. Nguyen, H.V.; Ulfarsson, M.O.; Sigurdsson, J.; Sveinsson, J.R. Deep sparse and low-rank prior for hyperspectral image denoising. In Proceedings of the IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 1217–1220. [Google Scholar]
  49. Miyato, T.; Kataoka, T.; Koyama, M.; Yoshida, Y. Spectral normalization for generative adversarial networks. arXiv 2018, arXiv:1802.05957. [Google Scholar]
  50. Rasti, B.; Ghamisi, P.; Gloaguen, R. Unsupervised deep hyperspectral inpainting using a new mixing model. In Proceedings of the IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 1221–1224. [Google Scholar]
  51. Maffei, A.; Haut, J.M.; Paoletti, M.E.; Plaza, J.; Bruzzone, L.; Plaza, A. A single model CNN for hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2516–2529. [Google Scholar] [CrossRef]
  52. Cascarano, P.; Sebastiani, A.; Comes, M.C. ADMM DIP-TV: Combining Total Variation and Deep Image Prior for image restoration. arXiv 2020, arXiv:2009.11380. [Google Scholar]
  53. Jiao, L.; Zhang, X.; Liu, X.; Liu, F.; Yang, S.; Ma, W.; Li, L.; Chen, P.; Feng, Z.; Guo, Y.; et al. Transformer meets remote sensing video detection and tracking: A comprehensive survey. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 1–45. [Google Scholar] [CrossRef]
  54. Shevitz, D.; Paden, B. Lyapunov stability theory of nonsmooth systems. IEEE Trans. Autom. Control 1994, 39, 1910–1914. [Google Scholar] [CrossRef]
  55. Bof, N.; Carli, R.; Schenato, L. Lyapunov theory for discrete time systems. arXiv 2018, arXiv:1809.05289. [Google Scholar]
  56. Hill, D.; Moylan, P. The stability of nonlinear dissipative systems. IEEE Trans. Autom. Control 1976, 21, 708–711. [Google Scholar] [CrossRef]
  57. Zhang, T.; Shen, Z. A fundamental proof of convergence of alternating direction method of multipliers for weakly convex optimization. J. Inequal. Appl. 2019, 2019, 128. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flow chart of the proposed LRS-PnP-DIP(1-Lip) algorithm. At each iteration, α k + 1 u k + 1 and x k + 1 are sequentially updated. The 1-Lipschitz DIP is implemented by imposing Lipschitz constraints on all layers. We use red color to highlight the differences between this work and the LRS-PnP-DIP algorithm [20]. The detailed design of the 1-Lipschitz DIP is placed in Appendix A.1.
Figure 1. Flow chart of the proposed LRS-PnP-DIP(1-Lip) algorithm. At each iteration, α k + 1 u k + 1 and x k + 1 are sequentially updated. The 1-Lipschitz DIP is implemented by imposing Lipschitz constraints on all layers. We use red color to highlight the differences between this work and the LRS-PnP-DIP algorithm [20]. The detailed design of the 1-Lipschitz DIP is placed in Appendix A.1.
Remotesensing 17 00288 g001
Figure 2. Comparison of MPSNR value of LRS-PnP among different τ ( τ = w s / w l r ) under different masks.
Figure 2. Comparison of MPSNR value of LRS-PnP among different τ ( τ = w s / w l r ) under different masks.
Remotesensing 17 00288 g002
Figure 3. Learning capability of DIP vs. number of input and output channels. Training is conducted with a single-band HS image, meaning that the input HSIs are processed; there is no correlation in the spectral domain. There is a significant performance gain when there are more input bands, indicating that the DIP with 2D convolution has the ability to exploit the correlation between channels.
Figure 3. Learning capability of DIP vs. number of input and output channels. Training is conducted with a single-band HS image, meaning that the input HSIs are processed; there is no correlation in the spectral domain. There is a significant performance gain when there are more input bands, indicating that the DIP with 2D convolution has the ability to exploit the correlation between channels.
Remotesensing 17 00288 g003
Figure 4. The amplitude of the singular value of the reconstructed image upon converge. The important singular values are captured and preserved via the 2D-convolution DIP, which is even more accurate than the traditional SVT projection.
Figure 4. The amplitude of the singular value of the reconstructed image upon converge. The important singular values are captured and preserved via the 2D-convolution DIP, which is even more accurate than the traditional SVT projection.
Remotesensing 17 00288 g004
Figure 5. Empirical converge of LRS-PnP-DIP(1-Lip) with modified NLM denoiser and non-expansive/1-Lipschitz DIP. Top left, top right, and bottom left: successive difference of x , λ 1 , and λ 2 in the log scale, respectively. Bottom right: the inpainting MPSNR vs. the number of iterations.
Figure 5. Empirical converge of LRS-PnP-DIP(1-Lip) with modified NLM denoiser and non-expansive/1-Lipschitz DIP. Top left, top right, and bottom left: successive difference of x , λ 1 , and λ 2 in the log scale, respectively. Bottom right: the inpainting MPSNR vs. the number of iterations.
Remotesensing 17 00288 g005
Figure 6. Different algorithms and their recovered spectrum of the center pixel with the assumption that the whole spectrum bands are missing. MPSNR increases from the thin line to the thick line.
Figure 6. Different algorithms and their recovered spectrum of the center pixel with the assumption that the whole spectrum bands are missing. MPSNR increases from the thin line to the thick line.
Remotesensing 17 00288 g006
Figure 7. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Chikusei dataset. From top to bottom: (1) clean image, (2) input image, (3) GLON, (4) DHP, (5) R-DLRHyIn, (6) DeepRED, (7) DeepHyIn, (8) PnP-DIP, (9) DDS2M, and (10) LRS-PnP-DIP(1-Lip). All images are visualized at band 80.
Figure 7. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Chikusei dataset. From top to bottom: (1) clean image, (2) input image, (3) GLON, (4) DHP, (5) R-DLRHyIn, (6) DeepRED, (7) DeepHyIn, (8) PnP-DIP, (9) DDS2M, and (10) LRS-PnP-DIP(1-Lip). All images are visualized at band 80.
Remotesensing 17 00288 g007
Figure 8. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Indian Pines dataset. From top to bottom: (1) clean image, (2) input image, (3) GLON, (4) DHP, (5) R-DLRHyIn, (6) DeepRED, (7) DeepHyIn, (8) PnP-DIP, (9) DDS2M, and (10) LRS-PnP-DIP(1-Lip). All images are visualized at band 150.
Figure 8. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Indian Pines dataset. From top to bottom: (1) clean image, (2) input image, (3) GLON, (4) DHP, (5) R-DLRHyIn, (6) DeepRED, (7) DeepHyIn, (8) PnP-DIP, (9) DDS2M, and (10) LRS-PnP-DIP(1-Lip). All images are visualized at band 150.
Remotesensing 17 00288 g008
Table 1. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Chikusei dataset: the mean and variance over 20 samples are shown here. During the experiments, four different mask shapes were used, which are depicted in the first four columns of Figure 7. The best results are underlined for each type of mask. For the metrics MPSNR and MSSIM, higher values indicate better performance. Conversely, for the MSAM metric, the lower the better.
Table 1. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Chikusei dataset: the mean and variance over 20 samples are shown here. During the experiments, four different mask shapes were used, which are depicted in the first four columns of Figure 7. The best results are underlined for each type of mask. For the metrics MPSNR and MSSIM, higher values indicate better performance. Conversely, for the MSAM metric, the lower the better.
Mask TypeMethodsInputGLON [4]DHP [11]R-DLRHyIn [16]DeepRED [22]DeepHyIn [50]PnP-DIP [24]DDS2M [19]LRS-PnP-DIP(1-Lip)
Mask Type 1MPSNR ↑33.0740 41.0324 41.3956(±0.52)41.5714(±0.31)41.5765(±0.28)41.6902(±0.57)41.7595(±0.25) 41.7728 ( ± 0.35 ) 41.8023 ( ± 0.16 ) =
MSSIM ↑0.2441 0.9078 0.9102(±0.03)0.9135(±0.01)0.9121(±0.02)0.9180(±0.02)0.9250(±0.02) 0.9268 ( ± 0.02 ) 0.9275 ( ± 0.01 ) =
MSAM ↓0.73410.13420.1296(±0.01)0.1250(±0.01)0.1237(±0.01)0.1221(±0.01) 0.1203 ( ± 0.01 ) 0.1183 ( ± 0.01 ) = 0.1185 ( ± 0.01 )
Mask Type 2MPSNR ↑31.715138.410339.7510(±0.65)40.1586(±0.84)40.4119(±0.33)40.6260(±0.40)40.6661(±0.52)40.7904(±0.74) 41.4910 ( ± 0.23 ) =
MSSIM ↑0.23380.87680.8961(±0.01)0.9041(±0.01)0.9049(±0.01)0.9100(±0.01)0.9121(±0.01)0.9130(±0.01) 0.9189 ( ± 0.01 ) =
MSAM ↓0.82960.15180.1436(±0.02)0.1373(±0.01)0.1320(±0.01)0.1315(±0.01)0.1298(±0.01) 0.1264 ( ± 0.01 ) 0.1203 ( ± 0.01 ) =
Mask Type 3MPSNR ↑30.325835.692738.1440(±0.90)38.9580(±0.79)39.6952(±0.99)39.8608(±0.68)39.8150(±0.40)40.0779(±0.82) 40.8400 ( ± 0.49 ) =
MSSIM ↑0.20180.85580.8805(±0.01)0.8871(±0.01)0.8902(±0.01)08950(±0.01)0.8958(±0.01)0.9003(±0.01) 0.9088 ( ± 0.01 ) =
MSAM ↓0.84410.16700.1563(±0.01)0.1507(±0.01)0.1495(±0.01)0.1480(±0.01)0.1476(±0.01) 0.1455 ( ± 0.01 ) 0.1392 ( ± 0.01 ) =
Mask Type 4MPSNR ↑27.980234.501136.6018(±1.15)37.6504(±0.89)37.9080(±0.59)38.1961(±0.77)38.3442(±0.73)38.7529(±0.88) 39.4068 ( ± 0.54 ) =
              MSSIM ↑0.17000.82670.8693(±0.01)0.8740(±0.01)0.8772(±0.01)0.8769(±0.01)0.8820(±0.01)0.8849(±0.01) 0.8927 ( ± 0.01 ) =
              MSAM ↓0.88470.18030.1714(±0.03)0.1636(±0.02)0.1602(±0.01)0.1580(±0.01)0.1557(±0.01) 0.1532 ( ± 0.01 ) 0.1449 ( ± 0.01 ) =
Table 2. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Indian Pines dataset: The mean and variance over 20 samples are shown here. During the experiments, four different mask shapes were used, which are depicted in the first four columns of Figure 7. The best results are underlined for each type of mask.
Table 2. Comparison between the proposed algorithm and other learning-based inpainting algorithms on the Indian Pines dataset: The mean and variance over 20 samples are shown here. During the experiments, four different mask shapes were used, which are depicted in the first four columns of Figure 7. The best results are underlined for each type of mask.
Mask TypeMethodsInputGLON [4]DHP [11]R-DLRHyIn [16]DeepRED [22]DeepHyIn [50]PnP-DIP [24]DDS2M [19]LRS-PnP-DIP(1-Lip)
Mask Type 1MPSNR ↑31.6903 39.7676 40.1917(±0.33)40.5490(±0.41)40.7982(±0.32)40.9309(±0.60)40.8897(±0.34) 41.1084 ( ± 0.29 ) 41.3423 ( ± 0.24 ) =
MSSIM ↑0.2268 0.8884 0.8917(±0.02)0.8930(±0.02)0.8932(±0.01)0.9028(±0.01)0.9050(±0.01) 0.9092 ( ± 0.01 ) 0.9139 ( ± 0.01 ) =
MSAM ↓0.77230.15940.1580(±0.01)0.1369(±0.01)0.1368(±0.01)0.1320(±0.01) 0.1255 ( ± 0.01 ) 0.1240 ( ± 0.01 ) 0.1256 ( ± 0.01 ) =
Mask Type 2MPSNR ↑29.681536.909238.1331(±0.22)38.7072(±0.25)38.5630(±0.19)38.7976(±0.40)39.0021(±0.32)39.9861(±0.40) 40.3010 ( ± 0.23 ) =
MSSIM ↑0.2098 0.8457 0.8849(±0.01)0.8902(±0.01)0.8864(±0.01)0.8919(±0.01)0.8972(±0.01) 0.9011 ( ± 0.01 ) 0.9048 ( ± 0.01 ) =
MSAM ↓0.87230.15140.1509(±0.01)0.1469(±0.01)0.1450(±0.01)0.1444(±0.01) 0.1420 ( ± 0.01 ) 0.1380 ( ± 0.01 ) 0.1310 ( ± 0.01 ) =
Mask Type 3MPSNR ↑28.619536.744737.0714(±0.28)37.4790(±0.20)37.3976(±0.39)37.4555(±0.39)37.7821(±0.22)37.8070(±0.52) 38.0599 ( ± 0.33 ) =
MSSIM ↑0.1864 0.8395 0.8787(±0.01)0.8870(±0.01)0.8869(±0.01)0.8893(±0.01)0.8921(±0.01) 0.8945 ( ± 0.01 ) 0.8982 ( ± 0.01 ) =
MSAM ↓0.88230.15840.1579(±0.01)0.1539(±0.01)0.1521(±0.01)0.1490(±0.01) 0.1488 ( ± 0.01 ) 0.1470 ( ± 0.01 ) 0.1458 ( ± 0.01 ) =
Mask Type 4MPSNR ↑24.699834.563035.8009(±0.30)35.8292(±0.22)35.8145(±0.32)35.8905(±0.23)36.1022(±0.18)36.4938(±0.34) 36.5104 ( ± 0.23 ) =
MSSIM ↑0.1672 0.8342 0.8592(±0.01)0.8651(±0.01)0.8678(±0.01)0.8698(±0.01)0.8733(±0.01) 0.8764 ( ± 0.01 ) 0.8779 ( ± 0.01 ) =
              MSAM ↓0.89770.16440.1610(±0.02)0.1611(±0.01)0.1602(±0.01)0.1588(±0.01) 0.1580 ( ± 0.01 ) 0.1552 ( ± 0.01 ) = 0.1554 ( ± 0.01 )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.; Yaghoobi, M. Self-Supervised Deep Hyperspectral Inpainting with Plug-and-Play and Deep Image Prior Models. Remote Sens. 2025, 17, 288. https://doi.org/10.3390/rs17020288

AMA Style

Li S, Yaghoobi M. Self-Supervised Deep Hyperspectral Inpainting with Plug-and-Play and Deep Image Prior Models. Remote Sensing. 2025; 17(2):288. https://doi.org/10.3390/rs17020288

Chicago/Turabian Style

Li, Shuo, and Mehrdad Yaghoobi. 2025. "Self-Supervised Deep Hyperspectral Inpainting with Plug-and-Play and Deep Image Prior Models" Remote Sensing 17, no. 2: 288. https://doi.org/10.3390/rs17020288

APA Style

Li, S., & Yaghoobi, M. (2025). Self-Supervised Deep Hyperspectral Inpainting with Plug-and-Play and Deep Image Prior Models. Remote Sensing, 17(2), 288. https://doi.org/10.3390/rs17020288

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop