Next Article in Journal
Methodology for Assessing the Digital Image of an Enterprise with Its Industry Specifics
Previous Article in Journal
A Modified Iterative Algorithm for Numerical Investigation of HIV Infection Dynamics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Color Channels Based Group Sparse Model for Image Restoration

1
Department of Information Engineering, Weihai Ocean Vocational College, Rongcheng 264300, China
2
School of Cyber Science and Engineering, Qufu Normal University, Qufu 273165, China
*
Authors to whom correspondence should be addressed.
Algorithms 2022, 15(6), 176; https://doi.org/10.3390/a15060176
Submission received: 24 April 2022 / Revised: 13 May 2022 / Accepted: 18 May 2022 / Published: 24 May 2022

Abstract

:
The group sparse representation (GSR) model combines local sparsity and nonlocal similarity in image processing, and achieves excellent results. However, the traditional GSR model and all subsequent improved GSR models convert the RGB space of the image to YCbCr space, and only extract the Y (luminance) channel of YCbCr space to change the color image to a gray image for processing. As a result, the image processing process cannot be loyal to each color channel, so the repair effect is not ideal. A new group sparse representation model based on multi-color channels is proposed in this paper. The model processes R, G and B color channels simultaneously when processing color images rather than processing a single color channel and then combining the results of different channels. The proposed multi-color-channels-based GSR model is compared with state-of-the-art methods. The experimental contrast results show that the proposed model is an effective method and can obtain good results in terms of objective quantitative metrics and subjective visual effects.

1. Introduction

Image restoration is one of the fundamental tasks in image processing and computer vision, and it aims to recover the original image x from the observed degraded image y . Image restoration is an ill-posed inverse problem and can be chartered by the following Formula (1):
y = H x + n
Here, x and y represent the original image and the degraded image respectively, H denotes the degenerate operator, and n denotes additive Gaussian noise. H is modeled according to different image processing tasks. When H is a unit matrix, Formula (1) represents an image denoising task [1,2]. When H is a blurring operator, Formula (1) represents an image deblurring task [3]. When H is a mask, Formula (1) represents image restoration [4,5]. When H is a random projection matrix, Formula (1) represents image compressive sensing applications [6,7].
The image restoration task is studied in this paper. In order to overcome the ill-posed characteristic and recover a satisfied copy of the original image, image priors are used to regularize the minimal problem,
arg min x 1 2 H x y 2 2 + λ Ψ ( x )
where the first term 1 2 H x y 2 2 represent the data fidelity term, Ψ ( x ) represents the regularization term of the image a priori, and λ is a regularization parameter. Image priors play an important role in image restoration, and many relevant algorithms are proposed, such as total-variation-based methods [8,9,10], sparse-representation-based methods [11,12,13,14], nonlocal self-similarity-based methods [1,2,15] and deep-convolutional-neural-network-based methods [16,17,18].
Sparse representation is an effective technique in image processing. Sparse-representation-based algorithms can be divided into two categories, i.e., patch sparse representation (PSR) [11,19] and group sparse representation (GSR) [12,13,15]. In the PSR model, the basic sparse representation unit is the image block. PSR assumes that each block of the image to be repaired can be perfectly modeled by a sparse linear combination of some fixed or learnable basic elements. These basic elements are called atoms and constitute a set called a dictionary. The PSR model not only does well in image denoising [20], but also can be applied to various other image processing and computer vision tasks [21]. However, the PSR model with an over-complete dictionary usually produces visual artifacts which are not expected in image restoration [22]. In addition, the PSR model assumes that the coded image blocks using sparse representation are independent and ignores the correlation among similar image blocks [13,23], which results in image degradation. To solve the above problems, the GSR model uses an image group as its basic processing unit to integrate the local sparsity with the nonlocal self-similarity of images. GSR shows great potential in various image processing tasks, and many literatures are provided. Zhang et al. [13] proposed the traditional GSR model for image restoration, which is essentially equal to the low rank minimization model [24]. Dong et al. [25] proposed an effective Gaussian scale mixing-based image restoration model under the GSR framework. Zha et al. [26] proposed a group-based adaptive dictionary, which connects the GSR model with the low rank minimization model. Zha et al. [14] combined the PSR and GSR models to reduce their shortcomings, respectively. Zha et al. [27] introduced the group sparse residual constraints to the GSR model to further define and simplify the problem of image restoration.
The traditional GSR model and the improved GSR models usually transform the RGB space of the image into YCbCr space and extract the Y (luminance) component for image processing. The Y channel in YCbCr space contains much valuable information about edge, shadow, object, texture and others. Moreover, human vision is sensitive to the change of the Y channel. These good characteristics make the GSR model and its improved models very successful in various image processing tasks. However, these methods only extract the Y channel, and cannot be loyal to each color channel of the color image. In this paper, the color channels refer to red (R), green (G), and blue (B) channels in RGB space. In order to exploit all the three channels, we propose a new GSR model based on multi-color channels. Compared with the traditional GSR model, which only extracts the Y channel in YCbCr space, the proposed model preserves as much image information as possible.
The main contribution of the paper is to propose the multi-color-channels-based GSR model for image restoration. In this proposed model, three channels are exploited simultaneously, and hence as much image color information as possible is retained. The paper is organized as follows. Section 1 introduces briefly the image restoration models and the development of the sparse-representation-based methods. Section 2 describes the traditional GSR model. The proposed multi-color-channels-based GSR model is provided in Section 3. Section 4 provides the visual and quantitative experimental results. Finally, conclusions are given in Section 5.

2. GSP Model

Compared with the patch sparse representation (PSR) model, the traditional group sparse representation (GSR) model uses the image group instead of a single image block as the basic unit of sparse representation, which produces satisfactory results in image processing tasks [13,28,29]. The model proposed in this paper is based on the traditional GSR model, which is briefly introduced here.

2.1. Image Group Construction

The construction of the image group in GSR model is illustrated in Figure 1. The image x with the size of N × N is divided into n  overlapped patches with the size of b × b . The overlapped patches are denoted by x i where i = 1 ,   2 ,   ,   n . For each x i , its m similar matched patches are constructed using the Euclidean distance as the similarity metric in a training window with the size of W × W . The similar matched patches form a set denoted by S i . All the similar patches in S i form a matrix with the size of b × m , denoted by X i R b × m . Each element in S i is a column of the matrix X i , i.e., X i = { x i , 1 , x i , 2 , , x i , m } . The column x i , j denotes the jth similar patch of the ith image group corresponding to x i . The matrix X i includes all the similar patches as columns, and hence is referred to as a group.

2.2. Sparse Representation Model

After constructing the image group, a sparse representation model is established in the image group. In order to obtain the sparse representation model of the image group, a dictionary, denoted by D i is given, which is usually learned from the image group, such as the graph-based dictionary [30] and principal component analysis (PCA) dictionary [31]. The aim of the sparse representation is to present the image group using as few atoms of the dictionary as possible. That is, the task is to code the dictionary sparsely using a sufficiently sparse vector. Each sparse representation, denoted by β ^ i , corresponding to the image group X i , is obtained via minimizing the L 0 norm using the following Formula (3).
β ^ i = arg min β i ( 1 2 X i D i β i 2 2 + λ β i 0 ) ,       i
Here, β i denotes the group sparse coefficient corresponding to the ith image group X i . 1 2 X i D i β i 2 2 is the fitting term used to measure the similarity between the image to be repaired via the sparse representation model and the original image. λ is a regularization parameter, and 0 denotes the L 0 norm used to calculate the no-zero terms of each column in β i . Since L 0 norm minimization is NP-hard, L 1 norm minimization is used instead of L 0 norm minimization via the following Formula (4).
β ^ i = arg min β i ( 1 2 X i D i β i 2 2 + λ β i 1 ) ,       i
The original image x is not available in the image restoration task, and the image groups are from the degraded image y R N . The image y is segmented into n overlapped patches with the same size of b × b , i.e., a vector with the length of b . The m similar matched patches are found to generate the image groups corresponding to the n overlapped patches. Each image group denoted by Y i R b × m is a matrix with the size of b × m , which also can be written as Y i = { y i , 1 ,   y i , 2 ,   ,   y i , m } . Now, the image restoration task is to restore X i from Y i using the GSR model via the following Formula (5).
β ^ i = arg min β i ( 1 2 Y i D i β i 2 2 + λ β i 1 ) ,       i  
After obtaining β ^ i , i = 1 ,   2 ,   ,   n , the restored whole image x ^ can be represented by a set of sparse codes { β ^ i } i = 1 n .

2.3. Learning a Dictionary

Sparse representation aims to obtain the sparsest representation of the original signal on the learned dictionary, and the signal is represented as a linear combination of a small number of dictionary atoms [32]. Therefore, in the framework of sparse representation, the construction of the dictionary is the key step of image restoration. However, in image restoration tasks, the original image x is unknown, and the original image group X i cannot be obtained directly. The dictionary is only constructed according to the degraded image y . The image y is first initialized using the bilinear interpolation algorithm. The following operations are similar to the procedures depicted in Section 2.1. Specifically, the estimated image group r i corresponding to the original image group X i is extracted from the initialized image y . Then the singular value decomposition is performed on the estimation group r i via the following Equation (6).
r i = U i i V i T = j = 1 n 1 γ i , j ( u i , j v i , j T )
Here, γ i denotes [ γ i , 1 ,   γ i , 2 ,   ,   γ i , n 1 ] , i denotes a diagonal matrix, γ i , j denotes the main diagonal elements of i , d represents each atom of the dictionary D i , and n 1 is defined as the minimum of d and m . u i , j and v i , j T denotes the column vectors of the matrixes U i and V i T , respectively. Each atom d i j R b × m of the adaptive dictionary D i is defined as follows,
d i j = u i , j v i , j T ,   j = 1 , 2 , 3 , , n 1
In this way, the adaptive dictionary D i , corresponding to each image group can be represented as follows:
D i = [ d i 1 , d i 2 , d i n 1 ]
The singular value decomposition is only needed to be performed on each group once to learn the adaptive dictionary corresponding to the image group. After the dictionary is learned, the sparse representation model is updated and then the L 1 norm minimization is solved. Finally, the split Bregman iteration (SBI) algorithm [13] is used for iterative optimization to obtain the final restoration result.

3. Proposed Multi-Color Channels Based GSR Model

The image restoration algorithm based on the GSR model has achieved good results in applications. However, most image restoration algorithms based on the GSR model transform the image into YCbCr space when processing color images, and only process the Y channel. Finally, the processed Y channel image is combined with other unprocessed channels to obtain the final restoration image. Although the Y channel of the image is formed by the integration of the three channels of a RGB color image in a certain proportion, it cannot preserve the maximum fidelity of the original R, G and B channels. This paper proposes a new GSR model based on multi-color channels, which processes three color channels of a RGB color image at the same time.

3.1. Construction of the Proposed Model

3.1.1. Construction of Image Groups

For an input degraded image y , the image group Y i C h , C h { R ,   G ,   B } , is constructed based on the multi-color channel image, where R , G and B denote the corresponding image channels. The construction of each channel image group Y i C h is the same as that of Y i in Section 2.2. Given a RGB image, in each channel, m similar patches of the repairing patch with the size of b × b are found in the training window with the size of W × W . Y i R , Y i G , and Y i B denote the image groups corresponding to the R , G and B channels, respectively. The three image groups form the total image group Y i C as Formula (9).
Y i C = [ Y i , 1 R     Y i , 2 R         Y i , m R Y i , 1 G     Y i , 2 G         Y i , m G Y i , 1 B     Y i , 2 B         Y i , m B ]     R 3 b × m
Y i = [ Y i , 1 Y     Y i , 2 Y         Y i , m Y ] R b × m
The above Formula (10) denotes the image group constructed using the Y channel in a traditional GSR model. The traditional GSR model and its improved models transform the RGB space into YCbCr space and extract the Y component for processing. The advantage of traditional GSR model is that human vision is sensitive to the change of the Y component in YCbCr space, and the Y component requires fewer storage units than the three color channels in RGB space, which is conducive to transmission and processing. However, the Y component cannot be loyal to each color channel in image processing. In this paper, as shown in Equation (9), stacking the image groups of three different color channels R, G and B in a matrix make it possible to tackle the three channels simultaneously. The total image group Y i C standing for the three color channels can be directly decomposed by singular value decomposition to obtain an adaptive dictionary and the sparse representation model. In the proposed model, the total image group is processed at the same time to avoid the loss of information caused by the transformation from RGB space to YCbCr space. Moreover, the rich information between different color channels can make up for each other [33]. On the other hand, there is no need to set and adjust the parameters of each color channel respectively when processing three color channels simultaneously, compared with separately processing the image group of each color channel to fuse different results to form the final combination results. Therefore, the proposed construction of image groups is more accurate that of the traditional GSR model. The advantages of the proposed multi-color-channels-based GSR model are low computational cost, low time cost and high efficiency.

3.1.2. Proposed Sparse Representation Model

After constructing a multi-color channel image group Y i C , a sparse representation model can be established based on the image group. By taking the traditional model Formula (5) and Y i C in Formula (9) in the regularization framework of Formula (2), the group sparse representation model based on multi-color channels of the proposed model in this paper is obtained.
β ^ = arg min β 1 2 y D C β C + λ i = 1 n β i C 1
Here, D C denotes all the adaptive dictionaries obtained via performing singular value decomposition on { Y i C } i = 1 n , i.e., D C = [ D 1 C , D 2 C , , D n C ] . β C represents the sparse coefficient of all multi-color channel image groups { Y i C } i = 1 n , i.e., β C = [ β 1 C , β 2 C , , β n C ] . It can be seen from Formula (11) that the proposed model simultaneously processes the image groups of multi-color channels.

3.1.3. Adaptive Dictionaries

An important problem of the proposed multi-color-channels-based GSR model is the selection of the dictionary. The dictionary is usually learned from natural images. Compared with the dictionaries of traditional analysis and design [33], such as wavelet, curvelet curve and discrete cosine transform, the dictionary directly learned from images can better adapt to the local structure of images, enhance the sparsity, and hence obtain higher performance. For example, the well-known K-SVD over-complete dictionary not only preserves the local structure of images, but also provides excellent denoising results [20]. However, the sparse representation on the over-complete dictionary is very computationally expensive and easily leads to artifacts in the image restoration [34]. In order to better adapt to the local structure of images, an adaptive dictionary D i C is learned for each group Y i C . The learning procedures of the adaptive dictionary D i C are similar to those depicted in Section 2.3. Firstly, the bilinear interpolation algorithm is used to initialize the degraded image y , and the estimation group r i C as in Formula (12) is obtained from the initialized image. Then singular value decomposition is performed on the estimation group, and the adaptive dictionary D i C of each group is learned through the estimation group r i C .
r i C = [ r i , 1 R   r i , 2 R     r i , m R r i , 1 G   r i , 2 G     r i , m G r i , 1 B   r i , 2 B     r i , m B ] R 3 b × m

3.2. Implementation of the Proposed Method

3.2.1. Solution of the Coefficients

In this paper, an alternating minimization algorithm [15] is used to solve the minimization problem of Equation (11). Solving the GSR model based on multi-color channels is to solve the β i minimization problem in Equation (11). This subsection provides implementation details of solving the β i minimization problem. Before solving the model proposed in this paper, the following theorem [13] is introduced.
Theorem 1. 
Let x ,   y R N , and X i ,   Y i R b × m . e R N is the error vector, i.e., e = x y . e j is the jth element of the error vector. Suppose that e j is independent and comes from a Gaussian distribution with zero mean and variance σ s 2 . Then for any ϖ > 0 , we have the following property to describe the relation between 1 N x y 2 2 and 1 S i = 1 n X i Y i 2 2 , that is
lim N S P ( | 1 N x y 2 2 1 S i = 1 n X i Y i 2 2 | < ϖ ) = 1  
where P ( ) represents the probability and S = b × m × n . (The proof is referred to in [13]).
Based on Theorem 1, Formula (11) is transformed into the following formul:
{ β ^ i C } i = 1 n = arg min { β i C } i = 1 n i = 1 n ( 1 2 Y i C D i C β i C 2 2 + τ β i C 1 )
where τ = λ S N , and S = b × m × n . Formula (14) may be tackled as the solution to the traditional GSR model, where the problem is divided into n sub-problems corresponding to each Y i C . In order to adapt to the local structure of the images, each Y i C corresponds to a learning adaptive dictionary D i C . Formula (14) is transformed into
{ β ^ i C } = arg min { β i C } i = 1 n i = 1 n ( 1 2 Y i C D i C β i C 2 2 + τ β i C 1 ) = arg min { β i C } i = 1 n i = 1 n ( 1 2 r i C D i C β i C 2 2 + τ β i C 1 ) = arg min { β i C } i = 1 n i = 1 n ( 1 2 γ i C β i C 2 2 + τ β i C 1 )
where Y i C = H r i C and r i C = D i C γ i C . The minimization problem of the Formula (14) is simplified to the solution to the problem of Formula (15). Different from the traditional GSR model, the proposed method obtains the coefficients via a soft threshold [35] as Formula (16).
β ^ i C = s o f t ( γ i C , 2 τ ) ,       i

3.2.2. Adaptive Parameters

In general, λ is set to a fixed value empirically. However, a fixed value cannot guarantee the stability of the whole algorithm. In order to overcome this challenge, an adaptive parameter-setting scheme is used to make the proposed multi-color-channels-based GSR model more stable and practical. Using spatially adaptive Laplacian priors, the regularization parameters adaptively update in each iteration [36]:
λ i = 2 2 σ w 2 σ i + ε
Here σ i is the ith local estimated variance, and ε is a sufficiently small constant to guarantee a no-zero denominator. The likelihood estimation [36] of σ i can be represented as
σ ^ i = max ( γ ^ i 2 / m σ m 2 , 0 )
where γ ^ i denotes the singular value of the estimation r i C , and m is the number of similar patches. The proposed method adopts an iterative regularization strategy [8] to alternatively update the estimation of the noise variance σ w and the signal variance σ i .
σ ^ w k = κ σ w 2 y y k + 1 2 2
σ ^ i k + 1 = max ( ( γ ^ i k ) 2 / m ( σ m k ) 2 , 0 )
Here κ is the scale factor controlling the noise variance, and k is the number of iterations. The complete description of the proposed GSR model based on multi-color channels is shown in the following Algorithm 1.
Algorithm 1 The proposed multi-color-channels-based GSR model
Procedures of multi-color-channels-based GSR model
Initialization: set x ^ ( 0 ) = y , λ = λ 0 ;
For k = 1, …, Iter do
Construct image group Y i C .
For Each group Y i C do
Construct D i C via SVD
Update λ via (17)
Update β ^ i C via (16)
Update X ^ C i = D i C β ^ i C
end for
end for
Output: the final restored image x ^

4. Experimental Results

This section provides lots of experimental results to verify the effectiveness of the proposed GSR model based on multi-color channels. In the restoration experiments, four types of corrupted images corresponding to different pixel missing rates are used as examples. The selected four pixel missing rates are 80%, 70%, 60%, and 50%. The pixels are lost randomly under different missing rates. The size of image block, b × b , is set to 5 × 5 , the overlap width between adjacent blocks is set to 1 pixel, the size of search window, W × W , is set to 20 × 20 , the number of similar blocks, m , is set to 75 , and the constant, ε , is set to 0.2 .
The quality of the restored images is evaluated according to objective quantitative metrics and subjective visual effects. Peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) [37] are used as objective metrics. The PSNR is calculated as shown in Formulas (21) and (22),
M S E = 1 H × W i = 0 H 1 j = 0 W 1 X ( i , j ) Y ( i , j ) 2
P S N R = 10 log 10 ( ( 2 n 1 ) 2 M S E )
Here, X and Y denote the original image and the repaired image, respectively. H and W denote the height and the width of the of the image, respectively. Formula (21) is used to calculate the mean square variance (MSE) of the original image and the repaired image. Formula (22) is used for calculating PSNR, where n is the number of bits per pixel. The larger the value of PSNR, the smaller the distortion. The calculation of SSIM is shown in Formulas (23)–(27),
l ( X , Y ) = 2 u X u Y + C 1 u X 2 + u Y 2 + C 1 ,   c ( X , Y ) = 2 σ X σ Y + C 2 σ X 2 + σ Y 2 + C 2 ,   s ( X , Y ) = 2 σ X Y + C 3 σ X σ Y + C 3
u X = 1 H × W i = 1 H j = 1 W X ( i , j )
σ X 2 = 1 H × W 1 i = 1 H j = 1 W ( X ( i , j ) u X ) 2
σ X Y = 1 H × W 1 i = 1 H j = 1 W ( ( X ( i , j ) u X ) ( Y ( i , j ) u Y ) )
S S I M ( X , Y ) = l ( X , Y ) c ( X , Y ) s ( X , Y )
SSIM is measured according to brightness, l , contrast, c , and image structure, s , in Formula (23). H and W denote the height and the width of the original image X and the repaired image Y . u X and u Y denote the mean value of X and Y . σ X and σ Y denote the variance of X and Y , respectively, and σ X Y denotes the covariance of X and Y . C 1 , C 2 and C 3 are constants to be introduced to avoid the denominator being zero. The range of SSIM is [ 0 ,   1 ] . The SSIM metric is closer to a human’s subjective feelings. The larger SSIM value indicates that the two images are more similar and the restoration effect is better.
In this paper, 10 images as shown in Figure 2 below are selected as test images. The quantitative experimental results and visual restoration effects given in the subsection are obtained on these images. The names of the images are Bahoon, Butterfly, Cowboy, Flower, Girl, Lake, Lena, Mickey, Mural, and Nanna, respectively.

4.1. Objective Assessment

The proposed model in this paper is compared with 11 state-of-the-art models. These models used for comparison include SALSA [38], BPFA [4], GSR [13], NGS [39], TSLRA [40], JPG-SR [14], GSRC-NLP [27], MAR-LOW [41], HSSE [42], IR-CNN [17], and IDBP [43]. Among the comparison models, GSR [13], JPG-SR [14], GSRC-NLP [27], and HSSE [42] are based on traditional GSR or improved GSR models, which are the same type as the proposed multi-color-channels-based GSR model in this paper. In order to comprehensively evaluate the performance of the proposed model for image restoration, the proposed model is also compared with non-GSR-based image restoration models. The BPFA [4] model uses the nonparametric Bayesian dictionary learning method for image sparse representation. The SALSA [38] model proposes an algorithm belonging to the extended Lagrangian method family to deal with constrained problems. The NGS [39] model proposes a new gradient regularization framework using nonlocal similarity. The TSLRA [40] and MAR-LOW [41] are the type of low rank minimization methods. The MAR-LOW [41] combines spatial autoregression with low rank minimization methods and achieves excellent results. The IR-CNN [17] and IDBP [43] are deep-learning-based methods.
Table 1 and Table 2 show the PSNR values of the restoration results on different images under the four pixels missing rates of 80%, 70%, 60% and 50% of the proposed model and the comparison models. In the tables, the best results are highlighted in bold. Table 3 shows the SSIM values of the results on different images under the four pixel missing rates of the proposed model and the comparison models, and the best results are highlighted in bold. Table 1 and Table 2 show the comparison results between the proposed model and the traditional models, and Table 3 shows the comparison results between the proposed model and the deep-learning-based models. In order to reduce the burden of experiments, the result data of NGS [39], TSLRA [40], and HSSE [42], provided in HSSE [42], are used here directly to avoid repetitive implementations. The experimental data of the deep learning methods IR-CNN [17] and IDBP [43] are directly from JPG-SR [14] and GSRC-NLP [27].
In addition to PSNR and SSIM values of each test image, the average PSNR and SSIM values of all test images under different pixel missing rates corresponding to different models are calculated for overall comparison. It can be seen from Table 1, Table 2 and Table 3 that the proposed multi-color-channels-based GSR model in this paper obtains better results than other competitive models. Firstly, the proposed model is compared with the non-GSR-based image restoration models. In terms of the average metric values, the proposed model is 7.20 dB higher than the SALSA model, 7.77 dB higher than the BPFA model, 6.00 dB higher than the NGS model, 4.94 dB higher than the TSLRA model, and 0.06 dB higher than the MAR-LOW model. Among the GSR based models, the JPG-SR method combines the PSR model with the GSR model, and it not only preserves the advantages of the two models, but also reduces their disadvantages, and therefore the sparse representation model is unified. The GSRC-NLP method introduces group sparse residual constraints to the traditional GSR model to approximate the group sparsity of the original group and improve the quality of image restoration. The HSSR model combines the external GSR model with the internal GSR model using the NSS prior knowledge of degraded images and external data sets to optimize the image. However, the above three GSR-based-models and many other improved models not compared in this paper transform the color image into YCbCr space, and only extract the Y channel for processing. In this way, image processing cannot be loyal to each color channel of the original color image, which makes the effect of image restoration poor. The proposed method utilizes three color channels in image restoration to obtain better results, and Table 1 and Table 2 prove this conclusion. Compared with the aforementioned three GSR-based models in terms of average PSNR metric values, the proposed model is 4.15 dB higher than the traditional GSR model, 3.99 dB higher than the JPG-SR model, 3.77 dB higher than the GSRC-NLP model, and 3.70 dB higher than the HSSE model. It can be concluded from Table 1 and Table 2 that the PSNR of the proposed model in this paper on the images with 80%, 70%, 60%, and 50% pixel missing rates are higher than the nine state-of-the-art comparison methods.
Table 3 provides the comparison results of the proposed model with deep-learning-based IR-CNN and ISBP models. In terms of average metric values, the PSNR and SSIM values of the proposed model are 4.1 dB and 0.049 higher than those of the IR-CNN model, and 4.72 dB and 0.067 higher than those of the IDBP model, respectively. It can be seen from the PSNR and SSIM values in Table 3 that the proposed model based on multi-color channels is better than the IR-CNN and IDBP models. The experimental results in Table 1, Table 2 and Table 3 show that the model proposed in this paper is effective and can obtain good restoration results compared with the state-of-the-art models.

4.2. Subjective Assessment

In order to further demonstrate the effectiveness of the proposed multi-color-channels-based GSR model, Figure 3 and Figure 4 respectively show the visual butterfly and girl restoration results of the proposed model and some state-of-the-art models with an 80% pixel missing rate. As can be seen from Figure 3 and Figure 4, the SALSA model cannot restore sharp edges and fine details. The BPFA model can have better visual effects than the salsa model, but it still has undesirable visual artifacts, such as ringing artifacts in Figure 4d. The GSR, JPG-SR and GSRC-NLP models produce over-smoothing effects. Compared with the model proposed in this paper, the MAR-LOW has a similar visual effect, which is not easy to distinguish with the naked eye. However, according to the experimental results in Table 1, Table 2 and Table 3, the proposed model has better objective evaluation results. In general, the visual effects of Figure 3 and Figure 4 show that the GSR model based on multi-color channels proposed in this paper performs best in image restoration. It not only preserves fine details, but also effectively removes visual artifacts.

5. Conclusions

In order to improve the performance of the traditional GSR model in image restoration, the paper extends the GSR model to a new multi-color-channels-based case, and process R, G, and B color channels of an image simultaneously. The problem of image restoration is transformed into how to use multi-color channels for sparse representation. The alternating minimization method with an adaptive parameter setting strategy is used to solve the proposed multi-color-channels-based model. The proposed method is compared with state-of-the-art methods. The experimental results show that the proposed model can obtain good results in terms of visual effects and quantitative metrics.

Author Contributions

Conceptualization, Y.K., C.Z. (Caiyue Zhou) and C.Z. (Chongbo Zhou); methodology, C.Z. (Chongbo Zhou); software, C.Z. (Caiyue Zhou); validation, Y.K. and C.Z. (Chongbo Zhou); formal analysis, L.S.; investigation, C.Z. (Chuanyong Zhang) and C.Z. (Chongbo Zhou); resources, C.Z. (Chongbo Zhou); data curation, L.S.; writing—original draft preparation, C.Z. (Caiyue Zhou); writing—review and editing, C.Z. (Chongbo Zhou); visualization, L.S.; supervision, C.Z. (Chongbo Zhou); project administration, C.Z. (Chongbo Zhou); funding acquisition, L.S. All authors have read and agreed to the published version of the manuscript.

Funding

The research of this paper was Funded by Ship Control Engineering and Intelligent Systems Engineering Technology Research Center of Shandong Province (Grant Number: SSCC20220003).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Buades, A.; Coll, B.; Morel, J.M. A Non-Local Algorithm for Image Denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; pp. 60–65. [Google Scholar]
  2. Dabov, K.; Foi, A.; Egiazarian, K. Video Denoising by Sparse 3D Transform-Domain Collaborative Filtering. In Proceedings of the 2007 15th European Signal Processing Conference, Poznan, Poland, 3–7 September 2007; pp. 145–149. [Google Scholar]
  3. Pan, J.; Sun, D.; Pfister, H.; Yang, M.H. Blind Image Deblurring Using Dark Channel Prior. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1628–1636. [Google Scholar]
  4. Zhou, M.; Chen, H.; Paisley, J.; Ren, L.; Li, L.; Xing, Z.; Dunson, D.; Sapiro, G.; Carin, L. Nonparametric Bayesian dictionary learning for analysis of noisy and incomplete images. IEEE Trans. Image Process. 2012, 21, 130–144. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Jin, K.H.; Ye, J.C. Annihilating filter-based low-rank hankel matrix approach for image inpainting. IEEE Trans. Image Process. A Publ. IEEE Signal Process. Soc. 2015, 24, 3498–3511. [Google Scholar]
  6. Zhang, J.; Zhao, D.; Zhao, C.; Xiong, R.; Ma, S.; Gao, W. Compressed Sensing Recovery via Collaborative Sparsity. In Proceedings of the 2012 Data Compression Conference, Snowbird, UT, USA, 10–12 April 2012; pp. 287–296. [Google Scholar]
  7. Yuan, X.; Huang, G.; Jiang, H.; Wilford, P.A. Block-Wise Lensless Compressive Camera. In Proceedings of the 2017 IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 31–35. [Google Scholar]
  8. Osher, S.; Burger, M.; Goldfarb, D.; Xu, J.; Yin, W. An iterative regularization method for total variation-based image restoration. Siam J. Multiscale Model. Simul. 2005, 4, 460–489. [Google Scholar] [CrossRef]
  9. Yuan, X. Generalized Alternating Projection Based Total Variation Minimization for Compressive Sensing. In Proceedings of the IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 2539–2543. [Google Scholar]
  10. Rudin, L. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  11. Li, H.; Liu, F. Image Denoising via Sparse and Redundant Representations over Learned Dictionaries in Wavelet Domain. In Proceedings of the 2009 Fifth International Conference on Image and Graphics, Xi’an, China, 20–23 September 2009; pp. 754–758. [Google Scholar]
  12. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-Local Sparse Models for Image Restoration. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  13. Zhang, J.; Zhao, D.; Gao, W. Group-based sparse representation for image restoration. IEEE Trans. Image Process. A Publ. IEEE Signal Process. Soc. 2014, 23, 3336. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Zha, Z.; Yuan, X.; Wen, B.; Zhang, J.; Zhou, J.; Zhu, C. Image Restoration Using Joint Patch-Group Based Sparse Representation. IEEE Trans. Image Process. 2020, 29, 7735–7750. [Google Scholar] [CrossRef]
  15. Liu, Y.; Yuan, X.; Suo, J.; Brady, D.J.; Dai, Q. Rank Minimization for Snapshot Compressive Imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2990–3006. [Google Scholar] [CrossRef] [Green Version]
  16. Zhang, J.; Ghanem, B. ISTA-Net: Interpretable Optimization-Inspired Deep Network for Image Compressive Sensing. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1828–1837. [Google Scholar]
  17. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning Deep CNN Denoiser Prior for Image Restoration. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2808–2817. [Google Scholar]
  18. Dong, W.; Wang, P.; Yin, W.; Shi, G.; Wu, F.; Lu, X. Denoising prior driven deep neural network for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2305–2318. [Google Scholar] [CrossRef] [Green Version]
  19. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef]
  20. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  21. Lun, D.P.K. Robust fringe projection profilometry via sparse representation. IEEE Trans. Image Process. 2016, 25, 1726–1739. [Google Scholar]
  22. Dong, W.; Zhang, L.; Shi, G.; Wu, X. Image deblurring and super-resolution by adaptive sparse domain selection and adaptive regularization. IEEE Trans. Image Process. 2011, 20, 1838–1857. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Wang, Q.; Zhang, X.; Wu, Y.; Tang, L.; Zha, Z. Nonconvex weighted lp minimization based group sparse representation framework for image denoising. IEEE Signal Process. Lett. 2017, 24, 1686–1690. [Google Scholar] [CrossRef]
  24. Dong, W.; Shi, G.; Li, X. Nonlocal image restoration with bilateral variance estimation: A low-rank approach. IEEE Trans. Image Process. A Publ. IEEE Signal Process. Soc. 2013, 22, 700–711. [Google Scholar] [CrossRef] [PubMed]
  25. Dong, W.; Shi, G.; Ma, Y.; Li, X. Image restoration via simultaneous sparse coding: Where structured sparsity meets gaussian scale mixture. Int. J. Comput. Vis. 2015, 114, 217–232. [Google Scholar] [CrossRef]
  26. Zha, Z.; Yuan, X.; Wen, B.; Zhou, J.; Zhang, J.; Zhu, C. A Benchmark for Sparse Coding: When Group Sparsity Meets Rank Minimization. IEEE Trans. Image Process. 2020, 29, 5094–5109. [Google Scholar] [CrossRef] [Green Version]
  27. Zha, Z.; Yuan, X.; Wen, B.; Zhou, J.; Zhu, C. Group sparsity residual constraint with non-local priors for image restoration. IEEE Trans. Image Process. 2020, 29, 8960–8975. [Google Scholar] [CrossRef]
  28. Xu, J.; Zhang, L.; Zuo, W.; Zhang, D.; Feng, X. Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 244–252. [Google Scholar]
  29. Zha, Z.; Liu, X.; Huang, X.; Shi, H.; Xu, Y.; Wang, Q.; Tang, L.; Zhang, X. Analyzing the Group Sparsity Based on the Rank Minimization Methods. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo, Hong Kong, China, 10–14 July 2017; pp. 883–888. [Google Scholar]
  30. Hu, W.; Li, X.; Cheung, G.; Au, O. Depth Map Denoising Using Graph-Based Transform and Group Sparsity. In Proceedings of the 2013 IEEE 15th International Workshop on Multimedia Signal Processing, Pula, Italy, 30 September–2 October 2013; pp. 1–6. [Google Scholar]
  31. Routray, S.; Ray, A.K.; Mishra, C. An efficient image denoising method based on principal component analysis with learned patch groups. Signal Image Video Process. 2019, 13, 1405–1412. [Google Scholar] [CrossRef]
  32. Mo, J.; Zhou, Y. The research of image inpainting algorithm using self-adaptive group structure and sparse representation. Clust. Comput. 2018, 22, 7593–7601. [Google Scholar] [CrossRef]
  33. Wen, B.; Ravishankar, S.; Bresler, Y. Structured overcomplete sparsifying transform learning with convergence guarantees and applications. Int. J. Comput. Vis. 2015, 114, 137–167. [Google Scholar] [CrossRef]
  34. Elad, M.; Yavneh, I. A plurality of sparse representations is better than the sparsest one alone. IEEE Trans. Inf. Theory 2009, 55, 4701–4714. [Google Scholar] [CrossRef]
  35. Daubechies, I.; Defrise, M.; Mol, C.D. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Commun. Pure Appl. Math. 2010, 57, 1413–1457. [Google Scholar] [CrossRef] [Green Version]
  36. Chang, S.G.; Yu, B. Adaptive wavelet thresholding for image denoising and compression. IEEE Trans. Image Process. A Publ. IEEE Signal Process. Soc. 2000, 9, 1532. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Afonso, M.V.; Bioucas-Dias, J.M.; Figueiredo, M.A. An augmented Lagrangian approach to the constrained optimization formulation of imaging inverse problems. IEEE Trans. Image Process. 2011, 20, 681–695. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Liu, H.; Xiong, R.; Zhang, X.; Zhang, Y.; Ma, S.; Gao, W. Nonlocal gradient sparsity regularization for image restoration. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1909–1921. [Google Scholar] [CrossRef]
  40. Guo, Q.; Gao, S.; Zhang, X.; Yin, Y.; Zhang, C. Patch-based image inpainting via two-stage low rank approximation. IEEE Trans. Vis. Comput. Graph. 2017, 24, 2023–2036. [Google Scholar] [CrossRef]
  41. Li, M.; Liu, J.; Xiong, Z.; Sun, X.; Guo, Z. MARLow: A Joint Multiplanar Autoregressive and Low-Rank Approach for Image Completion. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 819–834. [Google Scholar]
  42. Zha, Z.; Wen, B.; Yuan, X.; Zhou, J.; Zhu, C.; Kot, A.C. A hybrid structural sparsification error model for image restoration. IEEE Trans. Neural Netw. Learn. Syst. 2021. [Google Scholar] [CrossRef]
  43. Tirer, T.; Giryes, R. Image restoration by iterative denoising and backward projections. IEEE Trans. Image Process. 2019, 28, 1220–1234. [Google Scholar] [CrossRef]
Figure 1. The construction diagram of the image group.
Figure 1. The construction diagram of the image group.
Algorithms 15 00176 g001
Figure 2. Test images.
Figure 2. Test images.
Algorithms 15 00176 g002
Figure 3. The restoration results of butterfly with pixel missing rate of 80%. (a) Butterfly, (b) butterfly with pixels missing rate 80%, (c) SALSA [38] (PSNR = 22.85), (d) BPFA [4] (PSNR = 21.06), (e) GSR [13] (PSNR = 26.03), (f) MAR-LOW [41] (PSNR = 30.86), (g) JPG-SR [14] (PSNR = 26.58), (h) GSRC-NLP [27] (PSNR = 26.78), (i) our model (PSNR = 30.90).
Figure 3. The restoration results of butterfly with pixel missing rate of 80%. (a) Butterfly, (b) butterfly with pixels missing rate 80%, (c) SALSA [38] (PSNR = 22.85), (d) BPFA [4] (PSNR = 21.06), (e) GSR [13] (PSNR = 26.03), (f) MAR-LOW [41] (PSNR = 30.86), (g) JPG-SR [14] (PSNR = 26.58), (h) GSRC-NLP [27] (PSNR = 26.78), (i) our model (PSNR = 30.90).
Algorithms 15 00176 g003aAlgorithms 15 00176 g003b
Figure 4. The restoration result of girl with pixel missing rate of 80%. (a) Girl, (b) girl with pixels missing rate 80%, (c) SALSA [38] (PSNR = 23.79), (d) BPFA [4] (PSNR = 22.47), (e) GSR [13] (PSNR = 25.50), (f) MAR-LOW [41] (PSNR = 29.21), (g) JPG-SR [14] (PSNR = 25.55), (h) GSRC-NLP [27] (PSNR = 26.02), (i) our model (PSNR = 29.28).
Figure 4. The restoration result of girl with pixel missing rate of 80%. (a) Girl, (b) girl with pixels missing rate 80%, (c) SALSA [38] (PSNR = 23.79), (d) BPFA [4] (PSNR = 22.47), (e) GSR [13] (PSNR = 25.50), (f) MAR-LOW [41] (PSNR = 29.21), (g) JPG-SR [14] (PSNR = 25.55), (h) GSRC-NLP [27] (PSNR = 26.02), (i) our model (PSNR = 29.28).
Algorithms 15 00176 g004
Table 1. PSNR values of the proposed method and some state-of-the-art methods with pixel missing rates 80% and 70%.
Table 1. PSNR values of the proposed method and some state-of-the-art methods with pixel missing rates 80% and 70%.
Pixel Missing RateMethodsBahoonButterflyGirlLenaMuralNannaAverage
80%SALSA24.4121.9123.0327.3222.4523.2023.72
BPFA23.2521.0622.4727.5621.1022.3822.97
GSR24.5726.0325.5031.4126.0125.2426.46
NGS24.8923.8524.2728.8723.7824.5825.04
TSLRA25.4925.3225.1530.5825.2125.5526.22
JPG-SR25.4026.5825.5531.2526.2925.9226.83
GSRC-NLP25.5526.7826.0230.5126.5626.1726.93
MAR-LOW26.4030.8629.2133.8329.6629.6829.94
HSSE25.5726.6126.1231.8326.2326.2727.11
Proposed26.4330.9029.2833.8529.6729.7029.97
70%SALSA25.7124.8524.9929.6224.6925.3425.87
BPFA24.5723.9524.7130.3723.3424.4725.24
GSR26.1728.9227.8633.5428.4627.8928.81
NGS26.0826.3626.1830.7726.2926.3527.01
TSLRA26.7227.7627.0932.6427.2227.3228.13
JPG-SR26.8029.2427.9633.4028.5028.2429.02
GSRC-NLP26.9829.4728.2033.8528.7128.5129.29
MAR-LOW28.7033.8232.1436.2132.2632.7532.65
HSSE26.8729.2928.2533.8628.5728.4929.22
Proposed28.7233.9032.2636.2332.3032.7732.70
Table 2. PSNR values of the proposed method and some state-of-the-art methods with pixel missing rates 60% and 50%.
Table 2. PSNR values of the proposed method and some state-of-the-art methods with pixel missing rates 60% and 50%.
Pixel Missing RateMethodsBahoonButterflyGirlLenaMuralNannaAverage
60%SALSA26.7827.0226.7931.4226.5026.9727.58
BPFA25.8226.0626.6832.3825.1726.1427.04
GSR27.7431.0929.5435.8029.9830.1330.71
NGS27.2928.3727.8332.8127.9928.0628.73
TSLRA27.9229.4228.9134.2629.0729.1729.79
JPG-SR28.1431.1529.8735.4430.1530.3730.85
GSRC-NLP28.3131.4630.1735.9530.3330.5931.14
MAR-LOW30.7836.3334.5838.3034.6435.4435.01
HSSE28.2531.5430.2035.9330.3130.5331.13
Proposed30.8036.4134.7038.3434.7035.5535.08
50%SALSA27.9829.0328.3233.2628.1728.5729.22
BPFA27.1328.1628.4634.1527.2028.1728.88
GSR29.4232.7831.9337.6431.7332.1632.61
NGS28.4930.2829.6034.5629.8829.7130.42
TSLRA29.1531.0130.4835.5230.6230.8731.28
JPG-SR29.6132.8331.7737.1831.7232.2132.55
GSRC-NLP29.7533.0231.9537.6431.8832.3632.77
MAR-LOW32.8838.7037.0140.2436.9838.2237.34
HSSE29.7933.2932.1937.8431.9532.5032.93
Proposed32.9138.8037.1540.3037.1038.3737.44
Table 3. PSNR and SSIM values of the proposed method and two deep-learning-based methods with four different pixel missing rates.
Table 3. PSNR and SSIM values of the proposed method and two deep-learning-based methods with four different pixel missing rates.
ImagesPixel Missing Rate 80%Pixel Missing Rate 70%Pixel Missing Rate 60%Pixel Missing Rate 50%
IR-CNNIDBPProposedIR-CNNIDBPProposedIR-CNNIDBPProposedIR-CNNIDBPProposed
Cowboy25.4724.4329.3827.5427.1532.3129.9528.9735.3631.9631.4038.12
0.86220.84000.93430.91630.89550.96210.94790.92730.97700.96600.95220.9854
Girl25.3625.0329.2827.7827.4532.2630.1029.3234.7032.1031.0937.15
0.81350.79990.92120.89100.86980.95490.93270.91020.97200.95640.93710.9823
Flower28.4128.1232.1430.7730.4935.1132.9732.1637.5935.2834.1540.00
0.85860.84130.93810.91400.89430.96520.94600.92640.97860.96610.95080.9864
Lake25.2425.3928.8627.5927.8731.3329.7529.5333.7631.8031.4735.84
0.82780.82030.91150.89140.88050.94330.93020.91550.96230.95350.94210.9743
Mickey26.4525.4029.5729.6628.6932.2431.8231.1834.5134.2533.1436.62
0.86710.84360.92600.92200.90230.95320.94740.93190.96870.96370.95170.9787
Mural25.7525.2629.6728.7327.6832.3030.5829.7934.7032.2031.3537.10
0.79230.76940.89600.87210.83880.93930.90910.88240.96330.93610.91480.9780
Average26.1125.6129.8228.6828.2232.5930.8630.1635.1032.9332.1037.47
0.83690.81910.92110.90110.88020.95300.93560.91560.97030.95700.94150.9808
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kong, Y.; Zhou, C.; Zhang, C.; Sun, L.; Zhou, C. Multi-Color Channels Based Group Sparse Model for Image Restoration. Algorithms 2022, 15, 176. https://doi.org/10.3390/a15060176

AMA Style

Kong Y, Zhou C, Zhang C, Sun L, Zhou C. Multi-Color Channels Based Group Sparse Model for Image Restoration. Algorithms. 2022; 15(6):176. https://doi.org/10.3390/a15060176

Chicago/Turabian Style

Kong, Yanfen, Caiyue Zhou, Chuanyong Zhang, Lin Sun, and Chongbo Zhou. 2022. "Multi-Color Channels Based Group Sparse Model for Image Restoration" Algorithms 15, no. 6: 176. https://doi.org/10.3390/a15060176

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop