Next Article in Journal
Nontrivial Solutions of Systems of Perturbed Hammerstein Integral Equations with Functional Terms
Previous Article in Journal
Delay Equivalences in Tuning PID Control for the Double Integrator Plus Dead-Time
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Method of Measurement Matrix Optimization for Compressed Sensing Based on Alternating Minimization

1
Institute of Electronic Countermeasure, National University of Defense Technology, Hefei 230000, China
2
Huayin Ordnance Test Center, Weinan 714000, China
*
Author to whom correspondence should be addressed.
Mathematics 2021, 9(4), 329; https://doi.org/10.3390/math9040329
Submission received: 5 January 2021 / Revised: 25 January 2021 / Accepted: 26 January 2021 / Published: 7 February 2021

Abstract

:
In this paper, a new method of measurement matrix optimization for compressed sensing based on alternating minimization is introduced. The optimal measurement matrix is formulated in terms of minimizing the Frobenius norm of the difference between the Gram matrix of sensing matrix and the target one. The method considers the simultaneous minimization of the mutual coherence indexes including maximum mutual coherence μ m a x , t-averaged mutual coherence μ a v e and global mutual coherence μ a l l , and solves the problem that minimizing a single index usually results in the deterioration of the others. Firstly, the threshold of the shrinkage function is raised to be higher than the Welch bound and the relaxed Equiangular Tight Frame obtained by applying the new function to the Gram matrix is taken as the initial target Gram matrix, which reduces μ a v e and solves the problem that μ m a x would be larger caused by the lower threshold in the known shrinkage function. Then a new target Gram matrix is obtained by sequentially applying rank reduction and eigenvalue averaging to the initial one, leading to lower. The analytical solutions of measurement matrix are derived by SVD and an alternating scheme is adopted in the method. Simulation results show that the proposed method simultaneously reduces the above three indexes and outperforms the known algorithms in terms of reconstruction performance.

1. Introduction

Compressed sensing (CS) [1] can sample the sparse or compressible signals at a sub-Nyquist rate, which brings great convenience for data storage, transmission, and processing. By adopting the reconstruction algorithms, the signal can be exactly reconstructed from the sampled data. As a marvel way of signal processing, CS is applied in different fields such as image encryption [2], wideband spectrum sensing [3], wireless sensor network data processing [4], etc.
The original signal x R N × 1 is assumed to have a sparse representation in a known domain Ψ R N N × L ( N L ) as x = Ψ s where Ψ is the dictionary matrix and s is a K -sparse signal. The incomplete measurement y R M × 1 is obtained through the linear model
y = Φ x = D s
where Φ R M × N ( M < N ) is called the measurement matrix and D = Φ Ψ is the sensing matrix.
Some specific properties of Φ have great impacts on the reconstruction performance. In [5,6], spark and restricted isometric property (RIP) are respectively proposed as the sufficient conditions on Φ to recovery guarantee. However, computing the spark of a matrix has combinatorial complexity and certifying RIP for a matrix requires combinatorial search, that is to say, these tasks are NP-hard and difficult to accomplish. To a large extent, the coherence between Φ and Ψ reflects the performance of meeting the above conditions. Actually, the coherence is equivalent to the mutual coherence of D . Since the mutual coherence can be easily manipulated to provide recovery guarantees, it is commonly used to measure the performance of Φ . The frequently-used mutual coherence indexes include the maximum mutual coherence μ m a x [7], t-averaged mutual coherence μ a v e [8] and global mutual coherence μ a l l [9], which respectively represent the maximum, the average and the sum of squares of the correlation between any distinct pair of columns in D .
The first attempt to consider the optimal design of Φ is given in [8]. The simulation results carried out in [8] show that the optimized Φ leads to smaller μ a v e and a substantially better CS reconstruction performance is obtained. Then the optimization of Φ becomes an important issue in CS. The recent works try to find the optimal Φ which has excellent performance in reducing the mutual correlation by minimizing the Frobenius norm of the difference between the Gram matrix G = D D and the target Gram matrix G t . The main work focuses on designing G t and finding the best Φ .
In [8], G t is obtained by shrinking the off-diagonal entries in G . The shrinkage technique reduces μ a v e but it is time-consuming. Furthermore, μ m a x is still large, which ruins the worst-case guarantees of the reconstruction algorithms. A suitable point between the current solution and the one obtained using a new shrinkage function is chosen to design the G t in [10]. It is of very strong competitiveness in μ a v e and μ m a x . However, the optimal point is hard to determine and the unsuitable point may seriously degrade the algorithm’s performance. In [9], G t is obtained by averaging the eigenvalues of G . Simulation results show that μ a l l is reduced effectively but reducing μ a v e and μ m a x is hard to be guaranteed, which means μ a v e and μ m a x may maintain large values. Duarte-Carajalino and Sapiro [11] set G t as an identity matrix. Since D is overcomplete and G cannot be an identity matrix, simply minimizing the difference between G and G t does not imply low μ m a x [12]. In [13,14,15,16,17], G t is chosen from a set of relaxed Equiangular Tight Frames (ETF) [18]. The set can be formulated as S = G t R L × L : G t = G t , d i a g G t = 1 , max i j G t i , j μ w e l c h where μ w e l c h denotes the Welch bound [19] and G t i , j denotes the i , j th entry of G t . However, the maximum absolute value of off-diagonal entries in G is almost always greater than μ w e l c h . In this case, the optimization usually implies a solution D with low μ a v e but high μ m a x . In summary, the target Gram matrices mentioned above only focus on a certain mutual coherence index, and fail to take into account μ a v e , μ m a x and μ a l l simultaneously. When a certain index is targeted, the other indexes may not decrease significantly or even increase. Therefore, Φ is not ‘good’ enough and the reconstruction performance is well below par.
After designing G t , the next step is to find the ‘best’ Φ by approaching G to G t . In [8,9,10,11,17], G is obtained by applying SVD to G t primarily and then the square root D ^ of G is built as D ^ D ^ = G . At last, Φ is obtained by Φ = D ^ Ψ where denotes the Moore Penrose pseudoinverse. This kind of method is intuitive, but the generalized pseudoinverse poses problems of calculation accuracy and robustness [15]. In [14,15], gradient algorithm and quasi-Newtonian algorithm are respectively utilized to attain Φ . Firstly, the cost function F Φ with Φ as the variable is constructed. Then the search direction is determined by the derivative of F Φ . Finally, Φ is obtained with a fixed step size. However, choosing a suitable step size which has a great influence on the accuracy of the solution requires a lot of comparison work. Moreover, the gradient algorithm and quasi-Newtonian algorithm cannot converge until a certain number of iterations is accomplished, resulting in high computational cost. In [11,16], the method for designing Φ shares the same concept as K-SVD [20], that is to update a matrix row by row. Eigenvalue decomposition is required to find the square root of the maximum eigenvalue for each row, which results in a significant increase in the calculation. For solving this problem, Hong et al. [16] utilize the power method instead of eigenvalue decomposition. However, the eigenvalue obtained by power method is the one with the largest absolute value. When the eigenvalue is negative, eigenvalue decomposition is still necessary.
The primary contributions of this paper are threefold:
  • The new target Gram matrix G t targets μ a v e , μ m a x , and μ a l l of D simultaneously is designed. Firstly, a new shrinkage function whose threshold exceeds μ w e l c h is utilized to determine the initial target Gram matrix. Then G t is obtained by sequentially applying rank reduction and eigenvalue averaging to the initial matrix.
  • Analytical solutions of the measurement matrix Φ to minimize the difference between G = Ψ Φ Φ Ψ and G t are derived by SVD.
  • Based on alternating minimization, an iterative method is proposed to optimize the measurement matrix. The simulation results confirm the effectiveness of the proposed method in decreasing the mutual coherence indexes and improving reconstruction performance.
The remainder of this paper is organized as follows. Some basic definitions related to mutual coherence indexes and frames are described in the next section. The main results are presented in Section 3, where the solutions to the G t design are characterized and a class of the solutions to the optimal Φ is derived in detail. The procedure of our method and the discussion can be also found in Section 3. In Section 4, simulations are carried out to confirm the effectiveness of the proposed method. In the end, the conclusion is drawn.

2. Mutual Coherence Indexes and ETFs

2.1. Mutual Coherence Indexes

Rewrite D = d 1 , d 2 d L R M × L where d i R M × 1 and d i 2 = 1 . Denote g i j = d i d j the entry at the position of row i and column j in G , where i , j = 1 , 2 L . Here, we quote the definitions of the mutual coherence indexes as that presented by Donoho [7], Elad [8], and Zhao [9].
Definition 1.
For a matrix D , the maximum mutual coherence μ m a x is defined as the largest absolute and normalized inner product between all columns in D that can be described as
μ m a x = max i j d i d j = max i j g i j
Definition 2.
For a matrix D , the t-averaged mutual coherence μ a v e is defined as the average of all absolute and normalized inner products between different columns in D that are above t and can be described as
μ a v e = i j g i j t g i j i j g i j t
Definition 3.
For a matrix D , the global mutual coherence μ a l l is defined as the sum of squares of normalized inner products between all columns in D that can be described as
μ a l l = i j g i j 2
As shown in [5], the original signal can be exactly reconstructed as long as K < 1 + 1 / μ m a x / 2 . The conclusion is true from a worst-case standpoint which means that μ m a x does not do justice to the actual behavior of sparse representations. Therefore, Elad considers that an “average” measure of mutual coherence, namely μ a v e , is more likely to describe its true behavior. Different from the previous two indexes, μ a l l reflects the overall property of D .
In fact, the purpose of reducing the mutual coherence indexes of D is to attain G that meets the following requirements: (1) The maximum absolute value of off-diagonal entries in G is sufficiently small; (2) The number of off-diagonal entries with large absolute value is minimized; (3) The average of off-diagonal entries with large absolute value is as small as possible. However, when a certain mutual coherence index is targeted solely, we cannot guarantee that the obtained G will fully meet the requirements. Therefore, the decrease of a certain index does not always mean better Φ and improved reconstruction performance. When the three indexes are reduced simultaneously, the requirements are better satisfied and better performance is obtained.

2.2. ETFs

It is shown in [19] that μ m a x of D R M M × L is lower bounded by
μ m a x μ w e l c h = L M M L 1
The bound is achievable for ETF. Here, we recall the definition of ETF [18].
Definition 4.
Let F be a M × L matrix whose columns are f 1 , f 2 f L . The matrix F is called an equiangular tight frame if it satisfies three conditions
(1) 
Each column has a unit norm: f i 2 = 1 for i = 1 , 2 L .
(2) 
The columns are equiangular. For some nonnegative θ , we have f i f j = θ when i , j = 1 , 2 L and i j .
(3) 
The columns form a tight frame. That is, F F = L / M I M where I M is an M × M identity matrix.
Sustik et al. [18] show that a real M × L ( 1 < M < L 1 ) ETF exists on if L min M M + 1 / 2 , L M L M + 1 / 2 holds. Furthermore, M L 1 / L M and L M L 1 / M must be odd integers when L 2 M , M , and 2 M 1 must be an odd number and the sum of two squares respectively when L = 2 M . Fickus et al. [21] surveys some known construction of ETFs and tabulates existence for sufficiently small dimensions. The above studies show that M and L must meet some exacting requirements when an ETF is available for D . However, it is really difficult to meet the requirements in practice, which means the maximum absolute value of the off-diagonal entries in G is usually significantly larger than μ w e l c h .

3. The Proposed Method

The off-diagonal entries in G actually are the inner products between different columns in D . Reducing those entries is likely to lead to lower mutual coherence indexes and better performance. The most straightforward approach is to replace large off-diagonal values with small ones. However, it is impossible to solve Φ from a certain G because of the inequality of rank between Ψ Φ Φ Ψ and G when the approach is adopted. Therefore, a feasible approach is to minimize the difference between G and G t that can be formulated as
min G t G F 2
where G = Ψ Φ Φ Ψ . This problem can be solved by alternating minimization strategy [14,16], which iteratively minimizes (5) to find the desired Φ . The idea is to update G t and Φ alternatively and repeat this proceeding until a stop criterion is reached. In this section, we design G t firstly and then derive the analytical solutions of Φ . Finally, an iterative method is proposed to optimize the measurement matrix based on alternating minimization.

3.1. The Design of G t

It can be seen from (5) that G t plays an important role in measurement matrix optimization. In recent works, G t is frequently set as the relaxed ETF matrix, which is obtained by applying the following shrinkage function
G t i , j = g i j , g i j ς s i g n g i j ς , o t h e r w i s e
where ς = μ w e l c h for i , j = 1 , 2 L and i j . Such a scheme in designing G t guarantees that the off-diagonal entries with large value of G will be intensively constrained, which means lower μ a v e and μ m a x .
Recall from Section 2.2 that the Welch bound is not achievable for G in most cases. As shown in [14,16], different ς yields different results and μ w e l c h is not the optimal value. Li [22] et al. found that a smaller μ m a x is available when ς is slightly larger than μ w e l c h . Inspired by [22], we propose an improved shrinkage function which divides the entries in G into three segments through two thresholds. One of the thresholds is μ w e l c h and the other is larger than μ w e l c h . The function is as follows
G t i , j = s i g n g i j T h r , g i j T h r s i g n g i j μ w e l c h , μ w e l c h g i j < T h r g i j , g i j < μ w e l c h
where T h r = μ w e l c h + c and 0 < c < μ w e l c h . As can be seen from Equation (7), the maximum absolute value of off-diagonal entries in G t is raised from μ w e l c h to T h r . According to the previous analysis, the new function is likely to lead to a further reduction in μ m a x while maintaining the advantage of Equation (6) with respect to μ a v e .
After shrinkage, G t becomes full rank generally [8], that is R a n k G t = L . However, the rank of G is identically equal to M . Thus, we consider mending this by forcing a rank M . A new target Gram matrix, denoted as G t _ M , is obtained by solving
min G t G t _ M F 2 ,   s . t .   R a n k G t _ M = M
The solutions to this problem are given by Theorem 1 below.
Theorem 1.
Let G t R L L × L be the matrix obtained by applying the shrinkage operation shown as Equation (7) to G and G t = P Λ P be the eigendecomposition of G t . P is orthonormal with dimension L and Λ = d i a g λ 1 , λ 2 λ L with λ 1 λ 2 λ L . The solutions of the minimization problem defined by (8) are characterized by
G t _ M = P A Λ P
where A = I M 0 0 0 R L × L .
Proof. 
Denote G t _ M = X X where X R M M × L . Let X = U X Σ X 0 V X be an SVD of X , where U X R M × M and V X R L × L are unitary. Then G t _ M can be rewritten as
G t _ M = V X Σ X 2 0 0 0 V X
Denote f = G t G t _ M F 2 . By substituting G t _ M with G t _ M = X X , can be rewritten as a function of matrix X . Let f / X be the derivative of f with respect to X . The optimal X should satisfy f / X = 0 . Equivalently, we have
X X X = X G t
It then follows from X = U X Σ X 0 V X that
Σ X 2 0 0 0 = A V X G t V X
Substituting Equation (12) into Equation (10), we obtain
G t _ M = V X A V X G t
It turns out from the unitary invariance with Equation (13) that
f = I L V X A V X G t F 2
With a few manipulations, we conclude that the solution of (8) is equivalent to solving
max t r G t V X A V X G t
where t r denotes the matrix trace operation. Noting that A = A A and G t = P Λ P , the problem in (15) is equivalent to
max A V X P Λ F 2
Denote B = V X P and rewrite B as B = b 1 , b 2 b L where b i R L × 1 for i = 1 , 2 L . Rewrite A as A = e 1 , e 2 e M , 0 , 0 0 , where e j R L × 1 denotes a unit vector with the ith entry is equal to 1 for j = 1 , 2 M . Then, it is easy to obtain that
A V X P Λ = λ 1 e 1 b 1 λ 2 e 1 b 2 λ L e 1 b L λ 1 e 2 b 1 λ 2 e 2 b 2 λ L e 2 b L λ 1 e M b 1 λ 2 e M b 2 λ L e M b L 0 0 0 0 0 0
Let b j i = e j b i be the jth entry of b i . It can be shown with some manipulations that the problem in (16) is equivalent to
max i = 1 L j = 1 M λ i 2 b j i 2
With λ 1 λ 2 λ L and i = 1 L j = 1 M b j i 2 = M , it is straightforward that i = 1 L j = 1 M λ i 2 b j i 2 reaches the maximum value λ 1 2 + λ 2 2 + + λ M 2 only if j = 1 M b j i 2 = 1 for i = 1 , 2 M . In this case, j = 1 M b j i 2 = 0 and j = 1 M b i j 2 = 0 hold for i = M + 1 , M + 2 L . Rewrite B as B = B 1 B 2 B 3 B 4 where B 1 R M × M . Accordingly, all of the entries in both B 2 and B 3 are zero. Noting that B is unitary, it can be shown that B 1 and B 4 are unitary. Hence, we have
G t _ M = V X A V X P Λ P = P B A B Λ P = P A Λ P
As can be seen, the rank of G t _ M is equal to M . The proof is then completed. □
After the operation of rank reduction, the rank of G t _ M is equal to that of G . Additionally, G t _ M is most similar to G t in terms of Frobenius norm. Inspired by [9], we reduce the sum of squares of all off-diagonal values of G t _ M , namely μ ^ a l l , by eigenvalue averaging. When minimizing the difference between G and G t _ M , a smaller μ ^ a l l is more likely to lead to a smaller μ a l l . μ ^ a l l can be formulated as
μ ^ a l l = i , j = 1 L g ^ i j 2 i = 1 L g ^ i i 2
where g ^ i j denotes the i , j th entry of G t _ M and g ^ i i = 1 holds for i = 1 , 2 L .
Noting that i , j = 1 L g i j 2 = G ^ t F 2 and G t _ M = P Λ ^ P where Λ ^ = d i a g λ 1 , λ 2 λ M , 0 0 , μ ^ a l l can be rewritten as
μ ^ a l l = i = 1 M λ i 2 i = 1 L g ^ i i 2
Assuming that i = 1 M λ i is invariable, it then follows from the Cauchy BuniakowskySchwarz Inequality that i = 1 M λ i 2 takes the minimum value only if λ i = 1 M i = 1 M λ i for i = 1 , 2 M . Let λ ^ = 1 M i = 1 M λ i and Λ ^ = d i a g λ ^ , λ ^ λ ^ M , 0 0 , a new target Gram matrix denoted as G t _ o p t with M equal all non-zero eigenvalues is given by
G t _ o p t = P Λ ^ P
Recall that G t _ M is most similar to G t in terms of Frobenius norm, it means that G t _ M is of good competitiveness in μ a v e and μ m a x . Furthermore, as a variant of G t _ M , G t _ o p t reduces the sum of squares of all off-diagonal values of G t _ M , leading to a better performance in minimizing μ a l l . Therefore, G t _ o p t is more likely to be an ideal solution of target Gram matrix which leads to better μ a v e , μ m a x , and μ a l l simultaneously.

3.2. The Analytical Solutions of Φ

After obtaining the target Gram matrix G t _ o p t , the next step of the optimization is to find the best Φ . To handle the problem, we try to find the optimal solution by minimizing the difference between Ψ Φ Φ Ψ and G t _ o p t as
min G t _ o p t Ψ Φ Φ Ψ F 2
Let D = U D Σ D 0 V D be an SVD of D , where U D R M × M and V D R L × L are unitary. Similarly, Let Ψ = U Ψ Σ Ψ 0 V Ψ be an SVD of Ψ , where U Ψ R N × N and V Ψ R L × L are unitary. The solutions to this problem are given by Theorem 2 below.
Theorem 2.
Let G t _ o p t be the matrix shown as Equation (21) and Λ M R M × M be the M th principal submatrix of Λ ^ . Then the solutions of the minimization problem defined by (22) are characterized by
Φ o p t = U Z Λ M 1 2 0 P V Ψ Σ Ψ 1 0 U Ψ
where U Z R M × M is an arbitrary unitary matrix.
Proof. 
Assume that the off-diagonal values of Σ D and Σ Ψ are non-zero, Φ can be written as
Φ = U D Σ D 0 V D V Ψ Σ Ψ 1 0 U Ψ
By substituting Φ in (22) with Equation (24), it can be shown with some manipulations that the solutions of the problem in (22) are equivalent to the solutions of
min V D P Λ ^ P V D Σ D 2 0 0 0 F 2
Let Z = V D P Λ ^ P V D and z i be the ith diagonal entry of Z . Denote Λ Z = d i a g z 1 , z 2 z M . With further manipulations, we simplify (25) to
min G t _ o p t F 2 + Λ Z Σ D 2 F 2 Λ Z F 2
Obviously, the minima are achievable only if Λ Z = Σ D 2 holds and Λ Z F 2 takes the maximum value. Let U = V D P and u i j be the i , j th entry of U where i , j = 1 , 2 L . Noting that the top M diagonal entries of Λ ^ are all equal to λ ^ , we have
z i = λ ^ u i 1 2 + u i 2 2 + + u i M 2
It is worth noting that Λ Z F 2 = i = 1 M z i 2 and i = 1 L z i = M λ ^ where 0 z i λ ^ . Hence, it is clear that the maxima of i = 1 M z i 2 are reached when z i takes the maximum value λ ^ for i = 1 , 2 M . Rewrite U as U = U 1 U 2 U 3 U 4 where U 1 R M × M . Since z i = λ ^ holds for i = 1 , 2 M , we have u i 1 2 + u i 2 2 + + u i M 2 = 1 and U 2 = 0 , U 3 = 0 accordingly. As U Z is a unitary matrix, it is easy to verify that U 1 and U 4 are both unitary matrices. Then it follows that
V D = U 1 0 0 U 4 P
Substituting Equation (28) into Equation (24), the optimal solution is obtained by
Φ o p t = U Z Λ M 1 2 0 P V Ψ Σ Ψ 1 0 U Ψ
The proof is then completed. □

3.3. Comments

According to Section 3.1 and Section 3.2, the procedure for measurement matrix optimization has been summarized in Algorithm 1.
Algorithm 1. The proposed optimization method.
Input: Dictionary matrix Ψ R N N × L which has an SVD form of Ψ = U Ψ Σ Ψ 0 V Ψ , number of iterations Iter, constant c , Welch bound μ w e l c h .
Output: Measurement matrix Φ o p t .
Initialization: Initialize Φ 0 R M × N to a random matrix, initial U Z R M × M to a unitary matrix.
For l = 1 to I t e r do
  • Compute the sensing matrix D = Φ l Ψ and normalize the columns in D .
  • Compute Gram matrix G = D D .
  • Shrink G and obtain G t by G t i , j = s i g n g i j μ w e l c h + c , g i j μ w e l c h + c s i g n g i j μ w e l c h , μ w e l c h g i j < μ w e l c h + c g i j , g i j < μ w e l c h
  • Apply eigenvalue decomposition to G t and obtain G t = P Λ P .
  • Compute the average of the top M diagonal entries in Λ , denoted as λ ^ .
  • Construct Λ M R M × M as Λ M = d i a g λ ^ , λ ^ λ ^ .
  • Update Φ by Φ l = U Z Λ M 1 2 0 P V Ψ Σ Ψ 1 0 U Ψ .
end
return Φ I t e r
Noting that G t plays an important role in measurement matrix optimization, Algorithm 1 takes μ a v e , μ m a x and μ a l l into consideration simultaneously when designing the G t . By minimizing (5), the Gram matrix is most similar to G t in terms of Frobenius norm, leading to maintain the advantage of G t in reducing the mutual coherence indexes. Therefore, Algorithm 1 is effective in reducing μ a v e , μ m a x , and μ a l l .
In the shrinkage function, a different threshold yields different results. Inspired by [22], we propose a shrinkage function shown as (7) which has a new threshold μ w e l c h + c . We have not derived the optimal value of c in theory, but setting c to a proper value can also lead to a moderate result.
After averaging the eigenvalues of G t _ M , the first term on the right part of (20) is minimized. However, the diagonal entries of G t _ M change accordingly. Hence, we can’t assure that μ ^ a l l reaches the minima, that is to say, G t _ o p t may not be the optimal solution in terms of μ a l l . Fortunately, we find that the change of i = 1 M λ i 2 is much greater than that of i = 1 L g ^ i i 2 in (20), which means our approach is effective in reducing μ a l l .
The proposed algorithm is an iterative one. The main complexity of Algorithm 1 for each iteration is located at steps 1, 2, 4, and 7. For those steps, the flops required are O M N L , O M L 2 , O L 3 , and O L 3 respectively. Hence, the complexity of Algorithm 1 is approximate to be O I t e r L 3 . Since the complexity for similar algorithms in [8,16,17] which apply eigenvalue decomposition or SVD is no less than O I t e r L 3 , the proposed algorithm has not increased the complexity significantly.

4. Simulation Results and Discussion

In this section, we conduct simulations to predetermine a suitable c firstly. Then, we examine the mutual coherence indexes and reconstruction performance of the proposed method and compare them with the well-established similar algorithms given in [8,16,17] by presenting the empirical results. Last, we verify the effectiveness of our method with various measurement matrices and dictionary matrices. The iteration number Iter is set to 100 and t is set to μ w e l c h . For a given dictionary matrix Ψ R 80 × 120 , x R 120 × 1 has a sparse representation as x = Ψ s where s is K -sparse and each non-zero entry is randomly positioned with a Gaussian distribution of i.i.d. zero-mean and unit variance. Orthogonal Matching Pursuit (OMP) [23] algorithm is employed in signal reconstruction. Denote ε = x e x 2 / x 2 the reconstruction error where x e is the reconstructed signal. The reconstruction is identified as a success, called exact reconstruction, provided ε 10 6 . Denote P s u c the percentage of successful reconstruction. In Section 4.1, Section 4.2 and Section 4.3, Φ 0 and Ψ are both Gaussian random matrices.

4.1. The Choice of c

Since the analytical solution of c is extremely difficult, here, we conduct a serious of simulations to find a suitable c . Figure 1 illustrates the change tendency of mutual coherence indexes and P s u c with argument c . We fix the row number of M to 28, the sparsity to 8, and varies c from 0 to 0.16. The experiment is performed for 1000 random sparse ensembles and the results are recorded.
When c = 0 , the shrinkage function shown as (7) is the same as (6). As can be seen from the graphs, when c increases, μ a v e increases, μ m a x and μ a l l decrease first and then increase, P s u c increases firstly and then decreases. μ m a x and μ a l l reach the minima when c = 0.02 and c = 0.03 respectively. It is worth noting that appropriate increase of c leads to decrease of μ m a x and μ a l l but increase of μ a v e . When c = 0.01 , better μ m a x and μ a l l are obtained and the loss in μ a v e is tolerable. Moreover, P s u c reaches the maxima. Therefore, 0.01 may be a moderate value for c and c is set to 0.01 in the simulations in Section 4.2, Section 4.3 and Section 4.4.

4.2. Comparing the Mutual Coherence Indexes

This section presents a series of simulations to compare our method with algorithms given in [8,16,17] on the three mutual coherence indexes of D obtained by D = Φ o p t Ψ where Φ o p t is the optimized measurement matrix. For convenience, each method is denoted as Propose, Elad, Hong, and Entezari. The down-scaling factor for Elad is set to 0.95. The inner iteration number for Hong is set to 2, which means K-SVD is applied twice in every updating of Φ . The point is set to 0.5 to update the G t in Entezari.
Figure 2 illustrates the change tendency of mutual coherence indexes with iteration number for M = 28 . As can be seen from the figure, the indexes corresponding to different algorithms all change monotonously with the iteration number. When μ m a x and μ a v e converge, the number of iterations required by our method is almost equal to that of Hong and significantly less than that of Elad. When μ a l l converges, the number of iterations required by our method is equivalent to that of Entezari and significantly less than that of Hong and Elad.
Figure 3 presents the histogram of the absolute off-diagonal values of Φ o p t Ψ Φ o p t Ψ for M = 28 . It is seen from the figure that Elad and Entezari have long tails, showing that the number of off-diagonal values that exceed 0.34 is relatively large. The tail of Hong is shorter than that of Elad and Entezari, and reaches the maximum of 0.34. Compared with Hong, our method has a shorter tail which reaches the maximum of 0.32 and has more off-diagonal values below the μ w e l c h (0.1662).
Table 1, Table 2 and Table 3 present μ m a x , μ a v e , and μ a l l by Elad, Hong, Entezari, and our method versus measurement dimension M respectively. From Table 1, it can be observed that the μ m a x of our method is significantly less than that of Elad and Entezari, and less than Hong, which means our method is effective in reducing the maximum mutual coherence. In Table 2, we see that the μ a v e of our method is significantly less than that of Elad and Entezari, with an advantage of more than 0.03. It is worth noting that the μ a v e of our method is slightly larger than Hong. In Table 3, the μ a l l of our method is significantly less than that of Elad and Hong, with an advantage of more than 70 and 20 respectively. On the other hand, the μ a l l of our method is almost the same as Entezari.
In conclusion, while effectively reducing μ m a x and μ a l l , our method can maintain a small μ a v e at the same time. Additionally, the number of iterations required for the convergence of each index of our method is significantly less than that of Elad. Therefore, from the view of mutual coherence indexes, the measurement matrix obtained by our method has better properties than the other three methods. This coincides with the theoretical results obtained in the Section 3.

4.3. Comparing the Reconstruction Performance

Case 1.
Comparison of the P s u c in the noiseless case.
In this case, we conduct two separate CS experiments, first by fixing K = 8 and varying M from 12 to 44 and second by fixing M = 28 and varying K from 4 to 20. Each experiment is performed for 1000 random sparse ensembles and the number of exact reconstruction is recorded.
Figure 4 and Figure 5 reveal that the P s u c of our method is the highest, which indicates its superiority over the other three methods.
Case 2.
Comparison of the ε in the noisy case.
To show the robustness of the proposed method in noisy cases we consider the noisy model y = Φ x + v where v is the vector of additive Gaussian noise with zero means. We conduct the experiment by fixing M = 28 , K = 8 , and varying SNR from 10 to 50 dB. The experiment is performed for 1000 random sparse ensembles and the average reconstruction error is recorded. From Figure 6, we can see that the reconstruction errors decrease with the increase of SNR, and the error of the proposed method is smaller than that of the others.
Table 3 presents that μ a l l of the Entezari is slightly larger than that of our method. It is interesting to note that the number of off-diagonal entries with smaller absolute values in the Entezari is significantly larger than that of our method from Figure 3. Moreover, it can be seen from Table 2 that μ a v e of Hong is slightly lower than that of our method. However, the simulation results show that our method outperforms the others in terms of reconstruction performance. It is also worthy noting that our method reduces μ a v e , μ m a x , and μ a l l simultaneously, leading to better reconstruction performance in CS. This implies that a single mutual coherence index cannot accurately reflect the actual performance of the methods, and verifies the necessity of using multiple indexes simultaneously in measurement matrix optimization.

4.4. Different Kinds of Φ and Ψ Optimized by the Proposed Methods

To analyze the performance of our method with various measurement matrices and dictionary matrices, a serious of simulations are carried out in this section. We choose the measurement matrix as a Gaussian random matrix and a Bernoulli random matrix, and choose the dictionary matrix as a Gaussian random matrix and the DCT matrix, respectively. We compare the mutual coherence indexes and the reconstruction performance before and after optimization. When Ψ is the Gaussian random matrix, Φ belongs to R M × 80 and Ψ belongs to R 80 × 120 . When Ψ is the DCT matrix, Φ belongs to R M × 120 and Ψ belongs to R 120 × 120 . Each experiment is performed for 1000 random sparse ensembles.
The mutual coherence indexes of different measurement matrices Φ with different dictionary matrices Ψ are shown in Figure 7. As seen from the simulations, all the optimized measurement matrices produce smaller μ m a x , μ a v e , and μ a l l than the random ones.
Figure 8 and Figure 9 present the reconstruction performance of OMP with the optimized measurement matrices and the random ones. It is seen from the graphs in these figures that all the optimized matrices outperform the random ones in terms of the percentage of exact reconstruction.

5. Conclusions

This paper focused on the optimization of measurement matrix for compressed sensing. To decrease μ m a x , μ a v e , and μ a l l simultaneously, we designed a new target Gram matrix which was obtained by applying a new shrinkage function to the Gram matrix and updated by performing rank reduction and eigenvalue averaging. Then, we characterized the analytical solutions of the measurement matrix by SVD. Based on alternating minimization, we proposed an iterative method to optimize the measurement matrix. The simulation results show that the proposed method reduces μ m a x , μ a v e , and μ a l l simultaneously and outperforms the existing algorithms in terms of reconstruction performance. In addition, the proposed method is computationally less expensive than some existing algorithms in the literature.
As detailed, we gave the optimal value of c under a fixed matrix scale through simulation. When the scale changes, the value of c in Section 4.1 may no longer be applicable. Therefore, it is meaningful to find the theoretical ‘optimal value’ of c . Furthermore, noting that lower mutual coherence indexes mean potentially higher reconstruction performance, further efforts are needed to decrease the indexes simultaneously.

Author Contributions

Conceptualization: R.Y.; Methodology: C.C.; Software: R.Y.; Validation: B.W. and Y.G.; Formal analysis: C.C.; Data curation: R.Y.; Writing—original draft preparation: R.Y. and B.W.; Writing—review and editing: C.C. and Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

No funding was received for conducting this study.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  2. Xie, Y.; Yu, J.; Guo, S.; Ding, Q.; Wang, E. Image Encryption Scheme with Compressed Sensing Based on New Three-Dimensional Chaotic System. Entropy 2019, 21, 819. [Google Scholar] [CrossRef] [Green Version]
  3. Fang, Y.; Li, L.; Li, Y.; Peng, H.; Yang, Y. Low Energy Consumption Compressed Spectrum Sensing Based on Channel Energy Reconstruction in Cognitive Radio Network. Sensors 2020, 20, 1264. [Google Scholar] [CrossRef] [Green Version]
  4. Martinez, J.A.; Ruiz, P.M.; Skarmeta, A.F. Evaluation of the Use of Compressed Sensing in Data Harvesting for Vehicular Sensor Networks. Sensors 2020, 20, 1434. [Google Scholar] [CrossRef] [Green Version]
  5. Donoho, D.L.; Elad, M. Optimally sparse representation in general (nonorthogonal) dictionaries via 1 minimization. Proc. Natl. Acad. Sci. USA 2003, 100, 2197–2202. [Google Scholar] [CrossRef] [Green Version]
  6. Candes, E.J.; Tao, T. Decoding by linear programming. IEEE Trans. Inf. Theory 2005, 51, 4203–4215. [Google Scholar] [CrossRef] [Green Version]
  7. Stark, D.P.B. Uncertainty Principles and Signal Recovery. SIAM J. Appl. Math. 1989, 49, 906–931. [Google Scholar]
  8. Elad, M. Optimized Projections for Compressed Sensing. IEEE Trans. Signal Process. 2007, 55, 5695–5702. [Google Scholar] [CrossRef]
  9. Shaohai, H.U. An Optimization Method for Measurement Matrix Based on Eigenvalue Decomposition. Signal Process. 2012. [Google Scholar] [CrossRef]
  10. Yan, W.; Wang, Q.; Shen, Y. Shrinkage-Based Alternating Projection Algorithm for Efficient Measurement Matrix Construction in Compressive Sensing. IEEE Trans. Instrum. Meas. 2014, 63, 1073–1084. [Google Scholar] [CrossRef]
  11. Duartecarvajalino, J.M.; Sapiro, G. Learning to Sense Sparse Signals: Simultaneous Sensing Matrix and Sparsifying Dictionary Optimization. IEEE Trans. Image Process. 2009, 18, 1395–1408. [Google Scholar] [CrossRef] [Green Version]
  12. Lu, C.; Li, H.; Lin, Z. Optimized Projections for Compressed Sensing via Direct Mutual Coherence Minimization. Signal Process. 2018, 151, 45–55. [Google Scholar] [CrossRef] [Green Version]
  13. Xu, J.; Pi, Y.; Cao, Z. Optimized projection matrix for compressive sensing. EURASIP J. Adv. Signal Process. 2010, 2010, 560349. [Google Scholar] [CrossRef] [Green Version]
  14. Abolghasemi, V.; Ferdowsi, S.; Sanei, S. A gradient-based alternating minimization approach for optimization of the measurement matrix in compressive sensing. Signal Process. 2012, 92, 999–1009. [Google Scholar] [CrossRef] [Green Version]
  15. Zheng, H.; Li, Z.; Huang, Y. An Optimization Method for CS Projection Matrix Based on Quasi-Newton Method. Acta Electron. Sin. 2014, 42, 1977–1982. [Google Scholar]
  16. Hong, T.; Bai, H.; Li, S.; Zhu, Z. An efficient algorithm for designing projection matrix in compressive sensing based on alternating optimization. Signal Process. 2016, 125, 9–20. [Google Scholar] [CrossRef]
  17. Entezari, R.; Rashidi, A. Measurement matrix optimization based on incoherent unit norm tight frame. AEU Int. J. Electron. Commun. 2017, 82, 321–326. [Google Scholar] [CrossRef]
  18. Sustik, M.A.; Tropp, J.A.; Dhillon, I.S.; Heath, R.W. On the existence of equiangular tight frames. Linear Algebra Its Appl. 2007, 426, 619–635. [Google Scholar] [CrossRef] [Green Version]
  19. Welch, L. Lower bounds on the maximum cross correlation of signals (Corresp.). IEEE Trans. Inf. Theory 1974, 20, 397–399. [Google Scholar] [CrossRef]
  20. Aharon, M.; Elad, M.; Bruckstein, A.M. K-SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse Representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  21. Fickus, M.; Mixon, D.G. Tables of the existence of equiangular tight frames. arXiv 2015, arXiv:1504.00253. [Google Scholar]
  22. Li, G.; Zhu, Z.; Yang, D.; Chang, L.; Bai, H. On Projection Matrix Optimization for Compressive Sensing Systems. IEEE Trans. Signal Process. 2013, 61, 2887–2898. [Google Scholar] [CrossRef]
  23. Tropp, J.A.; Gilbert, A.C. Signal Recovery from Random Measurements Via Orthogonal Matching Pursuit. IEEE Trans. Inf. Theory 2007, 53, 4655–4666. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a) μ m a x and P s u c versus c ; (b) μ a v e and P s u c versus c ; (c) μ a l l and P s u c versus c , both with M = 28 and K = 8 .
Figure 1. (a) μ m a x and P s u c versus c ; (b) μ a v e and P s u c versus c ; (c) μ a l l and P s u c versus c , both with M = 28 and K = 8 .
Mathematics 09 00329 g001aMathematics 09 00329 g001b
Figure 2. The convergence results: (a) evolution of μ m a x , (b) μ a v e and (c) the evolution of μ a l l , all versus iteration number, where M = 28 .
Figure 2. The convergence results: (a) evolution of μ m a x , (b) μ a v e and (c) the evolution of μ a l l , all versus iteration number, where M = 28 .
Mathematics 09 00329 g002aMathematics 09 00329 g002b
Figure 3. Histogram of the absolute off-diagonal values of Φ o p t Ψ Φ o p t Ψ for M = 28 .
Figure 3. Histogram of the absolute off-diagonal values of Φ o p t Ψ Φ o p t Ψ for M = 28 .
Mathematics 09 00329 g003
Figure 4. The change tendency of P s u c with M while K = 8 in the noiseless case.
Figure 4. The change tendency of P s u c with M while K = 8 in the noiseless case.
Mathematics 09 00329 g004
Figure 5. The change tendency of P s u c with K while M = 28 in the noiseless case.
Figure 5. The change tendency of P s u c with K while M = 28 in the noiseless case.
Mathematics 09 00329 g005
Figure 6. The change tendency of ε with SNR while M = 40 and K = 8 .
Figure 6. The change tendency of ε with SNR while M = 40 and K = 8 .
Mathematics 09 00329 g006
Figure 7. The evolution of (a) μ m a x , (b) μ a v e and (c) μ a l l , all versus measurements number with different D .
Figure 7. The evolution of (a) μ m a x , (b) μ a v e and (c) μ a l l , all versus measurements number with different D .
Mathematics 09 00329 g007
Figure 8. The change tendency of P s u c with M while K = 8 in the noiseless case.
Figure 8. The change tendency of P s u c with M while K = 8 in the noiseless case.
Mathematics 09 00329 g008
Figure 9. The change tendency of P s u c with K while M = 28 in the noiseless case.
Figure 9. The change tendency of P s u c with K while M = 28 in the noiseless case.
Mathematics 09 00329 g009
Table 1. μ m a x by Elad, Hong, Entezari, and proposed method versus measurement dimension M .
Table 1. μ m a x by Elad, Hong, Entezari, and proposed method versus measurement dimension M .
MEladHongEntezariPropose μ w e l c h
120.89710.88840.84650.80720.2750
160.81570.64880.75170.52110.2337
200.76100.46250.67100.42180.2050
240.69720.39200.58740.36180.1833
280.61860.34030.53510.32230.1662
320.54600.30690.49110.29190.1520
360.48790.28040.44720.26590.1400
400.44540.25910.41770.24870.1296
440.38020.24040.39270.23020.1205
Table 2. μ a v e by Elad, Hong, Entezari, and proposed method versus measurement dimension M .
Table 2. μ a v e by Elad, Hong, Entezari, and proposed method versus measurement dimension M .
MEladHongEntezariPropose μ w e l c h
120.41680.45280.41410.36080.2750
160.35070.29640.35320.29790.2337
200.30160.25660.31110.25810.2050
240.24310.22720.27720.22880.1833
280.22120.20420.25120.20600.1662
320.20560.18540.23020.18790.1520
360.19370.17000.21180.17270.1400
400.18410.15670.19610.16000.1296
440.17600.14540.18240.14910.1205
Table 3. μ a l l by Elad, Hong, Entezari, and proposed method versus measurement dimension M .
Table 3. μ a l l by Elad, Hong, Entezari, and proposed method versus measurement dimension M .
MEladHongEntezariPropose
121220.202638.471081.181080.19
16898.80802.77780.92780.16
20721.41630.91605.72600.13
24571.34512.24480.67480.12
28474.76425.25394.86394.39
32404.35358.49330.55330.11
36353.96307.11280.55280.10
40315.03265.40240.55240.10
44285.51231.43207.79207.38
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yi, R.; Cui, C.; Wu, B.; Gong, Y. A New Method of Measurement Matrix Optimization for Compressed Sensing Based on Alternating Minimization. Mathematics 2021, 9, 329. https://doi.org/10.3390/math9040329

AMA Style

Yi R, Cui C, Wu B, Gong Y. A New Method of Measurement Matrix Optimization for Compressed Sensing Based on Alternating Minimization. Mathematics. 2021; 9(4):329. https://doi.org/10.3390/math9040329

Chicago/Turabian Style

Yi, Renjie, Chen Cui, Biao Wu, and Yang Gong. 2021. "A New Method of Measurement Matrix Optimization for Compressed Sensing Based on Alternating Minimization" Mathematics 9, no. 4: 329. https://doi.org/10.3390/math9040329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop