Next Article in Journal
A Collaborative Optimization Scheme for Beamforming and Power Control in MIMO-Based Internet of Vehicles
Previous Article in Journal
Chaos Prediction and Nonlinear Dynamic Analysis of a Dimple-Equipped Electrostatically Excited Microbeam
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computing One-Bit Compressive Sensing via Alternating Proximal Algorithm

1
College of Computer Science and Information Engineering, Harbin Normal University, Harbin 150025, China
2
College of Mathematics and Computer Science, Guangdong Ocean University, Zhanjiang 524088, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(18), 2926; https://doi.org/10.3390/math13182926
Submission received: 30 June 2025 / Revised: 17 August 2025 / Accepted: 8 September 2025 / Published: 10 September 2025

Abstract

It is challenging to recover a real sparse signal using one-bit compressive sensing. Existing methods work well when there is no noise (sign flips) in the measurements or the noise level or a priori information about signal sparsity is known. However, the noise level and a priori information about signal sparsity are not always known in practice. In this paper, we propose a robust model with a non-smooth and non-convex objective function. In this model, the noise factor is considered without knowing the noise level or a priori information about the signal sparsity. We develop an alternating proximal algorithm and prove that the sequence generated from the algorithm converges to a local minimizer of the model. Our algrithm possesses high time efficiency and recovery accuracy. It performs better than other algorithms tested in our experiments when the the noise level and the sparsity of the signal is known.

1. Introduction

One-bit compressive sensing (CS) was originally introduced in [1]. It has been applied extensively, such as in [2,3,4,5]. Let x be a sparse signal in R n and B be an m × n matrix. The measurement vector b R m of the signal x is calculated via
b = A x = sign ( B x )
where sign ( · ) operates componentwise with sign ( z ) = 1 if z 0 and sign ( z ) = 1 otherwise. The measurement operator A : R n { 1 , 1 } m is called the one-bit scalar quantizer. The prime target of one-bit CS is to recover the sparse signal x from the one-bit observation b and the measurement matrix B. We will now review existing models for the one-bit CS problem. The ideal optimization model is the following l 0 -norm minimization:
min x R n { x 0 } , s . t . b = sign ( B x ) ,
where x 0 is the l 0 -norm of x, counting the number of its non-zero entries. The NP-hardness of model (2) makes approximation challenging. The earliest attempt can be traced back to [1], in which the ideal model (2) was relaxed by
min x R n { x 1 } , s . t . Y B x 0 , x = 1 ,
where Y is a diagonal matrix with diagonal entries from b, and x 1 is the sum of the absolute values of the components of x. This model is valid for the noiseless case. For the case in which the measurement B x is contaminated by noise (sign flips), instead of solving model (3), ref. [1] relaxed it as follows:
min x R n { x 1 + λ [ Y B x ] 2 } , s . t . x = 1 ,
where λ > 0 is a parameter, and [ · ] represents the operator projected onto the convex set { u : u R m and u i 0 , i = 1 , 2 , , m } . Since both models (3) and (4) minimize convex objective functions over the non-convex unit sphere, to overcome difficulties resulting from non-convexity, the following model was proposed in [6]:
min x R n { x 1 } , s . t . Y B x 0 , B x 1 = r ,
where r is an arbitrary positive number. Obviously, the set determined by the two constraints is convex. In addition, the model can be efficiently solved using a linear programming method. However, it was shown numerically that solutions of model (5) are not sparse enough when the signal x R n is known to have s-sparsity, that is, x { x R n | x 0 s } . A newer model for the one-bit CS model,
min x R n { h ( [ Y B x ] ) } , s . t . x 0 s , x = 1
was proposed in [7], where function h is either · 1 or · 2 . An algorithm called binary iterative hard thresholding (BIHT) was developed for solving model (6). Based on BIHT, the adaptive outlier pursuit (AOP) technique was introduced in [8] for solving the following one-bit CS model:
min x R n { h ( [ Λ Y B x ] ) } , s . t . i = 1 m ( 1 Λ i i ) L , x 0 s , x = 1 ,
where L is the noise level, that is, there are at most L measurements in y that are wrongly detected (with sign flips), and Λ is an m × m diagonal matrix whose diagonal entries are either 1 or 0. If Λ i i = 1 , then the ith component y i of y is correct; otherwise, it is incorrect. If the matrix Λ is specified, model (7) can be reduced to model (6) by disregarding the incorrect measurements. However, in model (7), the noise level L must be given in advance, a condition that is hardly ever satisfied in practice. The following one-sided l 0 (OSL0) model was proposed in [9] for when L is unavailable:
min x R n , v R m { λ [ v ϵ e ] 0 + γ 2 Y B x v 2 + β 2 x 2 } , s . t . x 0 s ,
where ϵ , λ , γ and β are positive parameters, and e denotes the vector in R m with all components equal to one. In [9], a fixed-point proximity algorithm was proposed for solving model (8). It was shown that the sequence generated from the fixed-point proximity algorithm converges to a local minimizer of the objective function of model (8) and converges to a global minimizer of the function as long as the initial estimate is sufficiently close to any global minimizer of the function. In [10], the authors took advantage of l 1 -regularized least squares to address the one-bit CS problem regardless of the sign information of B x , namely,
min x R n { x 1 + λ b B x 2 } ,
where λ is a given positive parameter. A primal dual active set algorithm was proposed to solve model (9) and proved to converge within one step under two assumptions: the submatrix of B indexed on the nonzero components of the sparse solution is full row rank and the initial point is sufficiently close to the sparse solution. Therefore, the generated sequence again has a local convergence property. To eliminate the assumptions on data B for better convergence results, xiu et al. [11] proposed the following double-sparsity constrained model:
min x R n , y R m { Y B x + y ϵ e 2 + η x 2 } , s . t . x 0 s , y + 0 k ,
where y + 0 = [ y ] 0 , η > 0 is a penalty parameter and s n and k m are two positive integers representing prior information about the upper bounds of the signal sparsity and the number of sign flips, respectively. An algorithm called Gradient projection subspace pursuit was proposed in [11] to solve model (10).
To solve the previous model, either a priori knowledge is required or the algorithm lacks convergence analysis, excluding it from many potential applications. In this study, we introduce a model for the noisy one-bit CS problem that requires neither a priori knowledge of the signal sparsity nor a priori knowledge of the noise level of the measurements. We develop an algorithm to solve the proposed model and analyze the convergence of the proposed algorithm. Our algorithm possesses high time efficiency and recovery accuracy. Moreover, it performs better than other existing algorithms when the the noise level and the sparsity of the signal is known. To summarize, our algorithm is suitable for more many practical scenarios.

2. Elementary Facts

The Euclidean scalar product of R n and its corresponding norm are denoted by · , · , and · , respectively. For a symmetric matrix A, the maximum eigenvalue is denoted by λ max ( A ) . The M-norm of the vector x is denoted by x M : = x T M x if the matrix M is positive definite.
Definition 1 
(Rockafellar and Wets [12]). Let f : R n R { + } be a proper lower semicontinuous function.
(i) 
 The domain of f is defined and denoted by dom f : = { x R n | f ( x ) < + } ;
(ii) 
 For each x dom f , the Fréchet subdifferential of f at x, written ^ f ( x ) , is the set of vectors x R n that satisfies
lim inf y x , y x 1 y x [ f ( y ) f ( x ) x , y x ] 0 .
If x dom f , then ^ f ( x ) = ϕ ;
(iii) 
 The limiting-subdifferential (Mordukhovich [13]), or simply the subdifferential for short, of f at x dom f , written f ( x ) , is defined as follows:
f ( x ) : = { x R n | x n x , f ( x n ) f ( x ) , x n ^ f ( x n ) x } .
A necessary (but insufficient) condition for x R n to be a minimizer of a proper lower semicontinuous function f is 0 f ( x ) . For a proper lower semi-continuous function f : R n R { + } , the proximity operator of f is defined by [14]
prox f ( x ) : = arg min y R n { f ( y ) + 1 2 y x 2 } .
Clearly, for any z prox f ( x ) , by the calculus of the limiting-subdifferential, we have that x z f ( z ) .

3. The l 1 Model

A classical approach to problem (1) is the least squares (LS) approach [15], in which the estimator is chosen to minimize the data error:
min x R n { b sign ( B x ) 2 } .
Notice that the LS solution may have a huge norm and is thus meaningless. Regularization methods are needed to stabilize the solution. The basic concept of regularization is to replace the original problem with a “nearby” problem whose solution approximates the required solution. A popular regularization technique is Tikhonov regularization [16], in which a quadratic penalty is added to the objective function:
min x R n { α 2 b sign ( B x ) 2 + β 2 x 2 } .
The second term in the above minimization problem is a regularization term that controls the norm of the solution. Since x is a sparse signal, l 1 regularization (see, e.g., [17,18]) can be used to induce sparsity in the optimal solution:
min x R n { α 2 b sign ( B x ) 2 + β 2 x 2 + ρ x 1 } .
Notice that for any x R n and b { 1 , 1 } m , the following identity holds:
b sign ( B x ) 2 = 4 b sign ( B x ) 0 .
It follows from [9] (Proposition 3.1) that the function b sign ( B x ) 0 can be majorized by a lower semi-continuous function [ Y B x ϵ e ] 0 . We therefore substitute expression b sign ( B x ) 2 in model (16) by the lower semi-continuous function [ Y B x ϵ e ] 0 . For notational simplicity, we set A : = Y B and φ ( · ) : = [ · ϵ e ] 0 for a fixed ϵ > 0 . As a consequence, model (16) can be recast as
min x R n { λ φ ( A x ) + β 2 x 2 + ρ x 1 } ,
where λ = 2 α . The objective function of model (18) is lower semi-continuous and coercive; therefore, model (18) attains its minimum. By introducing an additional variable y R m , we can rewrite problem (18) as
min x R n , y R m { λ φ ( y ) + β 2 x 2 + ρ x 1 } s . t . A x = y .
Furthermore, model (19) can be approximated by the following unconstrained optimization problem (see, e.g., [19], model (2)):
min x R n , y R m { L ( x , y ) : = ρ x 1 + γ 2 A x y 2 + β 2 x 2 + λ φ ( y ) } ,
where γ is a positive relaxation parameter. model (20) has more than one local minimizer. We are now ready to estimate an upper bound of the number of local minimizers of model (20).
Proposition 1. 
The number of local minimizers for model (20) is no more than 2 m .
Proof. 
Let u = ( x , y ) be a local minimizer of model (20). Then, there is a positive number ε such that
ρ x 1 + γ 2 A x y 2 + β 2 x 2 + λ φ ( y ) ρ x 1 + γ 2 A x y 2 + β 2 x 2 + λ φ ( y )
holds for all u = ( x , y ) R n × R m satisfying u u < ε . For the vector y , we define S : = { j { 1 , 2 , , m } | y j < ϵ } . In association with this set, we further define
S y : = { y R m | y j < ϵ , j S ; y j ϵ , j { 1 , 2 , , m } S }
which is a convex set in R m . Obviously, we have y S y and φ ( y ) = φ ( y ) for all y S y . Hence,
ρ x 1 + γ 2 A x y 2 + β 2 x 2 + δ S y ( y ) ρ x 1 + γ 2 A x y 2 + β 2 x 2 + δ S y ( y )
holds for all u satisfying u u < ε . That is, the function ρ x 1 + γ 2 A x y 2 + β 2 x 2 + δ S y ( y ) attains its local minimum at the vector u . Since the function ρ x 1 + γ 2 A x y 2 + β 2 x 2 + δ S y ( y ) is a strictly convex function on R n + m , it actually attains its global minimum at the vector u . Notice that the number of all possible sets S y is 2 m . Therefore, there are a total 2 m functions of the form ρ x 1 + γ 2 A x y 2 + β 2 x 2 + δ S y ( y ) , each of which has at most one minimizer. As a result, the number of local minimizers for model (20) is no more than 2 m .    □

4. Alternating Proximal Algorithm

In this section, we describe an alternating minimization algorithm (see [19,20,21,22]) for finding local minimizers of the objective function L ( x , y ) of model (20). The alternating discrete dynamical system to be studied is of the following form:
x k + 1 = arg min x R n L x , y k + 1 2 x x k S 2 , y k + 1 arg min y R m L x k + 1 , y + 1 2 y y k T 2 ,
where S and T are entirely positive definite operators. Next, we address the computation of x k + 1 and y k + 1 . To update x k + 1 more easily, we take S = μ I γ A T A with μ > γ λ max ( A T A ) . Then, the x-subproblem can be solved by
x k + 1 = arg min x R n { ρ x 1 + γ 2 A x y k 2 + β 2 x 2 + λ φ ( y k ) + 1 2 x x k S 2 } = arg min x R n { ρ x 1 + μ + β 2 x 1 μ + β ( μ x k + γ A T ( y k A x k ) ) 2 } = prox ρ μ + β · 1 ( 1 μ + β ( μ x k + γ A T ( y k A x k ) ) ) = P ρ μ + β ( 1 μ + β ( μ x k + γ A T ( y k A x k ) ) ) ,
where the operator P r ( · ) with r > 0 at x R n is defined by
[ P r ( x ) ] i : = x i r , x i > r , x i + r , x i < r , 0 , o . w . , i = 1 , 2 , , n .
We take T = ν I with ν > 0 . Then, the y-subproblem can be solved by
y k + 1 arg min y R m { ρ x k + 1 1 + γ 2 A x k + 1 y 2 + β 2 x k + 1 2 + λ φ ( y ) + 1 2 y y k T 2 } = arg min y R m { λ φ ( y ) + ν + γ 2 y 1 ν + γ ( γ A x k + 1 + ν y k ) 2 } = prox λ ν + γ φ ( 1 ν + γ ( γ A x k + 1 + ν y k ) ) .
The proximity operator prox c φ ( · ) with c > 0 at y R m can be presented as follows [9] (Proposition 7.2):
[ prox c φ ( z ) ] i { z i } , z i < ϵ 2 c , { z i , ϵ } , z i = ϵ 2 c , { ϵ } , ϵ 2 c < z i < ϵ , { z i } , o . w . , i = 1 , 2 , , m .
Thus, a complete algorithm for solving model (20) can be presented as follows:
Next, we establish the convergence results of the sequence { ( x k , y k ) | k N } generated by Algorithm 1, where N denotes all natural numbers. We use ω ( x , y ) to denote the set of limit points of the sequence { ( x k , y k ) | k N } and crit L to denote the set of critical points of the function L.
Algorithm 1 Alternating proximal algorithm (APA) for model (20).
  • 1:   Input: the matrix B and the vector b, A = diag ( b ) B , ρ > 0 , β > 0 ,
  •                  λ > 0 , γ > 0 , μ > γ λ max ( A T A ) , ν > 0 and k max > 0 ;
  • 2:   Initialize: choose x 0 and y 0 ;
  • 3:   For  k = 0 , 1 , 2 , , k max 1   do
  •             x k + 1 = P ρ μ + β ( 1 μ + β ( μ x k + γ A T ( y k A x k ) ) ) ,
  •             y k + 1 prox λ ν + γ φ ( 1 ν + γ ( γ A x k + 1 + ν y k ) ) ;
  • 4:   end for
  • 5:   Output:  x ¯ .
Proposition 2. 
Let the sequence { ( x k , y k ) | k N } be generated by Algorithm 1. Then, the following hold:
(i) 
The sequence { L ( x k , y k ) | k N } is nonincreasing and convergent;
(ii) 
The sequence { ( x k , y k ) | k N } is bounded and
lim k + ( x k x k 1 + y k y k 1 ) = 0 ;
(iii) 
For all k 1 , define
( x k , y k ) : = ( γ A T ( y k y k 1 ) , 0 ) ( S ( x k x k 1 ) , T ( y k y k 1 ) ) ,
we then have
( x k , y k ) L ( x k , y k ) and lim k + ( x k , y k ) = ( 0 , 0 ) ;
(iv) 
ω ( x , y ) is a nonempty compact connected set and ω ( x , y ) crit L ;
(v) 
L is finite and constant on ω ( x , y ) , equal to lim k + L ( x k , y k ) .
Proof. 
(i) It follows from (25), (27), and the definition of L ( x , y ) that
L x k , y k 1 + 1 2 x k x k 1 S 2 L x k 1 , y k 1 , L x k , y k + 1 2 y k y k 1 T 2 L x k , y k 1 ,
which implies that
L ( x k , y k ) + 1 2 x k x k 1 S 2 + 1 2 y k y k 1 T 2 L ( x k 1 , y k 1 ) .
Since S and T are positive definite, the sequence { L ( x k , y k ) | k N } does not increase. Since L ( x , y ) 0 , x R n , y R m , the sequence { L ( x k , y k ) | k N } is convergent.
(ii) Since the function L ( x , y ) is coercive, proper lower semi-continuous, and bounded below, the sequence { ( x k , y k ) | k N } is bounded. In addition, we have lim k + ( x k x k 1 + y k y k 1 ) = 0 from inequality (33) and the convergence of { L ( x k , y k ) | k N } .
(iii) By the very definition of x k and y k , we have that for all k 1 ,
0 ρ x k 1 + γ A T A x k y k 1 + β x k + S x k x k 1 , 0 γ y k A x k + λ φ y k + T y k y k 1 .
Because of the definition of L ( x , y ) , we have
x L x k , y k = ρ x k 1 + γ A T A x k y k + β x k , y L x k , y k = γ y k A x k + λ φ y k .
Hence, we can obtain from (34) and (35) that
γ A T y k y k 1 S x k x k 1 x L x k , y k , T y k y k 1 y L x k , y k .
This yields ( x k , y k ) L ( x k , y k ) with [19] (Proposition 2.1). Furthermore, it follows from (29) that lim k + ( x k , y k ) = ( 0 , 0 ) .
(iv) It follows from (ii) and the results of point set topology that ω ( x , y ) is nonempty compact connected. Let ( x ¯ , y ¯ ) be a point in ω ( x , y ) , where there exists a subsequence { ( x k , y k ) } of { ( x k , y k ) } converging to ( x ¯ , y ¯ ) . Furthermore, by the definition of y k , we have for all k 1 and y R m ,
γ 2 A x k y k 2 + λ φ ( y k ) + 1 2 y k y k 1 T 2 γ 2 A x k y 2 + λ φ ( y ) + 1 2 y y k 1 T 2 .
It follows from (29) that by replacing k with k in (37) and letting k + , we can deduce
lim inf k + { λ φ ( y k ) } + γ 2 A x ¯ y ¯ 2 λ φ ( y ) + γ 2 A x ¯ y 2 + 1 2 y y ¯ T 2 , y R m .
In particular, for y = y ¯ , we have that lim inf k + { λ φ ( y k ) } λ φ ( y ¯ ) . Since the function φ ( y ) is lower semicontinuous, we have lim inf k + { λ φ ( y k ) } = λ φ ( y ¯ ) . There is no loss of generality in assuming that the whole sequence { λ φ ( y k ) } converges to λ φ ( y ¯ ) , i.e.,
lim k + { λ φ ( y k ) } = λ φ ( y ¯ ) .
Notice that the function ρ x 1 + γ 2 A x y 2 + β 2 x 2 is continuous, so we have
lim k + { ρ x k 1 + γ 2 A x k y k 2 + β 2 x k 2 } = ρ x ¯ 1 + γ 2 A x ¯ y ¯ 2 + β 2 x ¯ 2 .
Thus, lim k + L ( x k , y k ) = L ( x ¯ , y ¯ ) . It follows from (31) that ( x k , y k ) L ( x k , y k ) and lim k + ( x k , y k ) = ( 0 , 0 ) . Owing to the closedness of L , we determine that ( 0 , 0 ) L ( x ¯ , y ¯ ) . Hence, ω ( x , y ) crit L .
(v) Let ( x ¯ , y ¯ ) be a point in ω ( x , y ) so that there is a subsequence { ( x k , y k ) } of { ( x k , y k ) } , with lim k + L ( x k , y k ) = L ( x ¯ , y ¯ ) . Since the sequence { L ( x k , y k ) | k N } is convergent, we have L ( x ¯ , y ¯ ) = lim k + L ( x k , y k ) independent of ( x ¯ , y ¯ ) , i.e., L is finite and constant on ω ( x , y ) . □
The next result shows that the sequence { ( x k , y k ) | k N } generated by Algorithm 1 converges to a local minimizer of (20).
Theorem 1. 
The sequence { ( x k , y k ) | k N } generated by Algorithm 1, with the initial point ( x 0 , y 0 ) , converges to a local minimizer of (20).
Proof. 
Let ( x ¯ , y ¯ ) be a point in ω ( x , y ) . Then, there exists a subsequence { ( x k , y k ) } of { ( x k , y k ) } converging to ( x ¯ , y ¯ ) , and (39) holds. Define
S : = { j { 1 , 2 , , m } | y ¯ j < ϵ } , S + : = { j { 1 , 2 , , m } | y ¯ j > ϵ } , S 0 : = { j { 1 , 2 , , m } | y ¯ j = ϵ } , δ 1 : = 1 2 min { | y ¯ j ϵ | | j S S + } .
For all y R m satisfying y y ¯ < δ 1 , we deduce that the entries of both y and y ¯ are all less than ϵ on the index set S and are all greater than ϵ on the index set S + . Hence, by the definition of φ , we have
φ ( y ¯ ) = φ ( y ) j S 0 [ y j ϵ ] 0 .
On one hand, there at least exists one index j S 0 such that y j < ϵ . Then, j S 0 [ y j ϵ ] 0 1 . Thus, we have
φ ( y ¯ ) φ ( y ) 1 .
Further, we denote the function ρ x 1 + γ 2 A x y 2 + β 2 x 2 by f ( x , y ) . Since f ( x , y ) is continuous, there is a constant δ 2 > 0 such that
f ( x ¯ , y ¯ ) < f ( x , y ) + λ
holds for all ( x , y ) satisfying ( x , y ) ( x ¯ , y ¯ ) < δ 2 . Choose δ : = min { δ 1 , δ 2 } . It follows from (44) and (43) that for all ( x , y ) satisfying ( x , y ) ( x ¯ , y ¯ ) < δ , there exists at least one index j S 0 such that y j < ϵ , and we have L ( x ¯ , y ¯ ) < L ( x , y ) .
On the other hand, for all ( x , y ) satisfying ( x , y ) ( x ¯ , y ¯ ) < δ and y j ϵ , j S 0 , we have
φ ( y ¯ ) = φ ( y ) .
By the definition of y k , we get
λ φ ( y k + 1 ) + γ 2 y k + 1 A x k + 1 2 + 1 2 y k + 1 y k T 2 λ φ ( y ) + γ 2 y A x k + 1 2 + 1 2 y y k T 2 ,
which, together with (29) and (39), implies that
λ φ ( y ¯ ) + γ 2 y ¯ A x ¯ 2 λ φ ( y ) + γ 2 y A x ¯ 2 + 1 2 y y ¯ T 2 .
Thus, we have that
γ 2 y ¯ A x ¯ 2 γ 2 y A x ¯ 2 + 1 2 y y ¯ T 2 .
For any t [ 0 , 1 ] , define y ˜ : = t y + ( 1 t ) y ¯ , then y ˜ y ¯ < δ and y ˜ j ϵ , j S 0 . We denote the function γ 2 y A x ¯ 2 by g ( y ) . Then, g ( y ) is a convex function and
g ( y ¯ ) g ( y ˜ ) + 1 2 y ˜ y ¯ T 2 .
Hence, we have
g ( y ¯ ) t g ( y ) + ( 1 t ) g ( y ¯ ) + t 2 2 y y ¯ T 2 ,
which can be reduced to
g ( y ¯ ) g ( y ) + t 2 y y ¯ T 2 .
By letting t 0 + , it yields that
g ( y ¯ ) g ( y ) .
Furthermore, by the definition of x k , we have
x k + 1 = prox ρ μ + β · 1 ( 1 μ + β ( μ x k + γ A T ( y k A x k ) ) ) .
It follows from (29) and the continuity of the operator prox ρ μ + β · 1 ( · ) that
x ¯ = prox ρ μ + β · 1 ( 1 μ + β ( μ x ¯ + γ A T ( y ¯ A x ¯ ) ) ) ,
which is equivalent to
x ¯ = arg min x R n { ρ x 1 + γ 2 A x y ¯ 2 + β 2 x 2 } .
Hence, we obtain
ρ x ¯ 1 + γ 2 A x ¯ y ¯ 2 + β 2 x ¯ 2 ρ x 1 + γ 2 A x y ¯ 2 + β 2 x 2 .
It follows from (52) that
A x y ¯ 2 A x y ¯ 2 + y A x ¯ 2 y ¯ A x ¯ 2 A x y 2 + A ( x x ¯ ) 2 + y y ¯ 2 ,
which, together with (56), implies that
ρ x ¯ 1 + γ 2 A x ¯ y ¯ 2 + β 2 x ¯ 2 ρ x 1 + γ 2 A x y 2 + β 2 x 2 + γ 2 A ( x x ¯ ) 2 + γ 2 y y ¯ 2 .
We use h ( x , y ) to denote the function ρ x 1 + γ 2 A x y 2 + β 2 x 2 . Then h ( x , y ) is a convex function. For any t [ 0 , 1 ] , define ( x ˜ , y ˜ ) : = t ( x , y ) + ( 1 t ) ( x ¯ , y ¯ ) . Then, ( x ˜ , y ˜ ) ( x ¯ , y ¯ ) < δ and y j ϵ , j S 0 . It follows from (58) that
h ( x ¯ , y ¯ ) h ( x ˜ , y ˜ ) + γ 2 A ( x ˜ x ¯ ) 2 + γ 2 y ˜ y ¯ 2 ,
which can be reduced to
h ( x ¯ , y ¯ ) t h ( x , y ) + ( 1 t ) h ( x ¯ , y ¯ ) + γ t 2 2 A ( x x ¯ ) 2 + γ t 2 2 y y ¯ 2 ,
i.e.,
h ( x ¯ , y ¯ ) h ( x , y ) + γ t 2 A ( x x ¯ ) 2 + γ t 2 y y ¯ 2 .
By letting t 0 + , we can obtain
h ( x ¯ , y ¯ ) h ( x , y ) .
It follows from (45) and (62) that we have L ( x ¯ , y ¯ ) L ( x , y ) for all ( x , y ) ( x ¯ , y ¯ ) < δ and y j ϵ , j S 0 . Thus, ( x ¯ , y ¯ ) is a local minimizer of (20). By Proposition 1 and (iv) of Proposition 2, we have that the sequence { ( x k , y k ) | k N } converges to a local minimizer of (20). □
Remark 1. 
It follows from the proof of Theorem 1 that if ( A x ¯ ) j ϵ , j S 0 , then we have A x ¯ = y ¯ . At this point, we have that ( x ¯ , y ¯ ) is also a local minimizer of (19).
The next theorem shows that if the initial point of Algorithm 1 is sufficiently close to any one of the global minimizers of the function L given in (20), then the sequence generated by Algorithm 1 converges to a global minimizer of model (20).
Theorem 2. 
Let the sequence { ( x k , y k ) | k N } be generated by Algorithm 1 with the initial point ( x 0 , y 0 ) , and let ( x , y ) be global minimizers of (20). If ( x 0 , y 0 ) is sufficiently close to ( x , y ) with y 0 Ω : = { y R m | y j ϵ for j satisfying y j = ϵ } , then the sequence { ( x k , y k ) | k N } converges to a global minimizer of (20).
Proof. 
We know from Proposition 1 that model (20) has a finite number of local minimizers. If the function L ( x , y ) has a unique local minimal value, then by Theorem 1, we have that the sequence { ( x k , y k ) | k N } converges to a global minimizer of (20). Otherwise, we suppose that L ( x , y ) has at least two local minimal values, and we denote the second smallest minimal value by M. Then, we have that h ( x , y ) = L ( x , y ) λ φ ( y ) < M λ φ ( y ) . Since the function h ( x , y ) is continuous, we can determine that h ( x 0 , y 0 ) < M λ φ ( y ) if ( x 0 , y 0 ) is sufficiently close to ( x , y ) . It follows from (42) that if y 0 y < δ 1 with y 0 Ω , we have L ( x 0 , y 0 ) < M . By (i) of Proposition 2, it holds that L ( x k , y k ) L ( x 0 , y 0 ) < M for all k. By Theorem 1, we know that the sequence { ( x k , y k ) | k N } converges to ( x ¯ , y ¯ ) which is a local minimizer of L ( x , y ) . By the lower semi-continuity of L ( x , y ) , we get L ( x ¯ , y ¯ ) < M . This implies that ( x ¯ , y ¯ ) must be a global minimizer of (20). This completes the proof of the desired result. □

5. Numerical Simulations

In this section, we describe simulation experiments conducted to demonstrate the effectiveness of our proposed Algorithm 1. Our code was written in Matlab 2015b and executed on a Dell personal computer with 11th Gen Intel(R) Core(TM) i5-1135G7 @ 2.40 GHz 2.42 GHz and 16 G memory.
We generated an m × n matrix B whose entries were subjected to the independent and identically distributed (i.i.d.) standard Gaussian distribution. We then generated an s-sparse vector x R n , whose nonzero components were drawn from the i.i.d. samples of the standard Gaussian distribution. To avoid tiny nonzero entries of x , we let x i = x i + sign ( x i ) for nonzero x i and then normalized the resulting vector to be a unit vector. The ideal one-bit measurement vector b is given by b = sign ( B x ) . To simulate sign flips in actual one-bit measurements, we randomly selected a m components in b and flipped their signs, denoting the resulting vector b. And a ( 0 , 1 ) denotes the flipping ratio in the measurement.
Three metrics were chosen to evaluate the quality of the signals reconstructed by one-bit compressive algorithms. They are the signal-to-noise ratio (SNR), Hamming error (HE), and Hamming distance (HD), defined, respectively, by
SNR : = 20 log 10 ( x ¯ x ) , HE : = 1 m b sign ( B x ¯ ) 0 , HD : = 1 m b sign ( B x ¯ ) 0 ,
where x ¯ is the reconstructed signal with norm 1. The higher the SNR value, the better the reconstructed signal. The values of HE and HD are within the range [0, 1]. The smaller these values, the better the reconstructed signal.
The larger parameter γ in model (20) is set, the closer model (20) will be to model (19). Hence, in subsequent numerical experiments, we initially set γ = 500 in Algorithm 1, doubled it in every ten iterations, and fixed it when the total number of iterations exceeds 60. This is the same approach described in [9]. For other parameters, we chose μ = 1.001 γ λ max ( A T A ) , ν = 0.005 , λ = 80 , ϵ = 0.05 , β = 10 5 , and ρ = 0.005 ( μ + β ) . We set the initial estimates ( x 0 , y 0 ) = ( 0 , 0 ) and k max = 500 .
First, we explored the performance ofAlgorithm 1 without the noise level or a priori information about signal sparsity. We tried problems of sizes n = 1000 , 2000 , 3000 , and 5000. For each n, we generated 100 input vectors and reported the average results over 100 runs in Table 1. We defined stopc : = 1 m b sign ( B ( x k x k ) ) 0 , with k = 1 , 2 , 3 , . In all tests, if stopc < 10 6 or k k max , the Algorithm 1 terminated and outputted x ¯ = Π r ( x k ) / Π r ( x k ) , where the operator Π r ( · ) with r > 0 at x R n is defined by
[ Π r ( x ) ] i : = 0 , r < x i < r , x i , o . w . , i = 1 , 2 , , n .
The parameter r is set to ( 40 + 50 a ) ρ μ + β if n 2000 ; otherwise, ( 25 + 50 a ) ρ μ + β .
From Table 1, we can see that Algorithm 1 can effectively recover the sparse signal x from the one-bit observation b if the flipping ratio a [ 0 , 0.05 ] .
Second, we assumed that the sparsity s was known and compared the performance of the Algorithm 1 (APA) with two state-of-the-art algorithms, namely, OSL0 in [9] and AOP in [8]. In our experiments, the parameters for OSL0, AOP, and their variants were determined as suggested in [8,9]. All methods start with the same initial points ( x 0 , y 0 ) = ( 0 , 0 ) and k max = 300 . In Algorithm 1, we projected x onto set C = { x R n | x 0 s } every 25 iterations and outputted x ¯ = x k max / x k max .
Three configurations were considered to test the robustness of Algorithm 1 for the one-bit compressive sensing problem. They were designed to test cases of different levels of noise in the measurements, the size of the sensing matrix B, and the true sparsity s of the vector x . In the first configuration, we fixed m = n = 1000 and sparsity at s = 10 and varied the noise level a { 0 , 0.05 , 0.1 , 0.15 , 0.2 , 0.25 , 0.3 } . In the second configuration, we fixed n = 1000 , sparsity at s = 10 , and the noise level at a = 0.05 and varied m such that the ratio m / n was in { 0.05 , 0.1 , 0.5 , 1 , 1.5 , 2 , 2.5 } . In the third configuration, we fixed m = n = 1000 and the noise level at a = 0.05 and changed the sparsity s { 5 , 8 , 10 , 12 , 15 , 18 , 20 } .
For the first configuration, the average values of SNR, HE, and HD over 50 trials for the signals reconstructed by all algorithms against the noise levels are depicted in Figure 1a, Figure 1b, and Figure 1c, respectively. We can observe from Figure 1 that Algorithm 1 performs better than the other algorithms when the noise level is greater than 0.05 .
For the second configuration, the average values of SNR, HE, and HD over 50 trials for the signals reconstructed by all algorithms against the ratio of m / n are depicted in Figure 2a, Figure 2b, and Figure 2c, respectively. We can see that Algorithm 1 performs better than the other algorithms when the ratio of m / n is greater than 0.5 .
For the third configuration, the average values of SNR, HE, and HD over 50 trials for the signals reconstructed by all algorithms against the sparsity of the ideal signals are depicted in Figure 3a, Figure 3b, and Figure 3c, respectively. We can see that Algorithm 1 performs best among all the algorithms tested in our experiments.

6. Conclusions

In this paper, we propose a robust model for the one-bit CS problem and prove that the model has a finite number of local minimizers. We propose an alternating proximal algorithm for solving the proposed model and prove the following: the sequence of the objective function decreases and converges, the set of limit points of the sequence generated from the algorithm is included in the set of critical points of the objective function, the sequence generated from the algorithm converges to a local minimizer of the objective function of the proposed model, and the sequence generated from the algorithm converges to a global minimizer of the function as long as the initial estimate is sufficiently close to any global minimizer of the function. The proposed algorithm is suitable for practical scenarios in which neither the noise level nor the sparsity of the signal are known in advance. Our algorithm possesses high time efficiency and recovery accuracy. Moreover, it performs better than other algorithms tested in our experiments when the the noise level and the sparsity of the signal is known.

Author Contributions

J.-J.W. is responsible for conceptualization, data curation, formal analysis, investigation, software, visualization, and writing—orinitial draft. Y.-H.H. is responsible for conceptualization, funding acquisition, methodology, validation, writing—original draft and writing—review & editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Sciences Grant (No. 11871182) and the program for scientific research start-up funds of Guangdong Ocean University (060302102004 and 060302102005).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors have no competing interests to declare that are relevant to the content of this article.

References

  1. Boufounos, P.T.; Baraniuk, R.G. 1-bit Compressive Sensing. In Proceedings of the Conference on Information Science and Systems (CISS), Princeton, NJ, USA, 19 March 2008. [Google Scholar] [CrossRef]
  2. Zhou, Z.; Chen, X.; Guo, D.; Honig, M.L. Sparse channel estimation for massive MIMO with 1-bit feedback per dimension. In Proceedings of the 2017 IEEE Wireless Communications and Networking Conference (WCNC), San Francisico, CA, USA, 19–22 March 2017. [Google Scholar] [CrossRef]
  3. Chen, C.H.; Wu, J.Y. Amplitude-aided 1-bit compressive sensing over noisy wireless sensor networks. IEEE Wirel. Commun. Lett. 2015, 4, 473–476. [Google Scholar] [CrossRef]
  4. Fu, N.; Yang, L.; Zhang, J. Sub-nyquist 1-bit sampling system for sparse multiband signals. In Proceedings of the 2014 22nd European Signal Processing Conference (EUSIPCO), Lisbon, Portugal, 1–5 September 2014. [Google Scholar]
  5. Li, Z.; Xu, W.; Zhang, X.; Lin, J. A survey on one-bit compressed sensing: Theory and applications. Front. Comput. Sci. 2018, 12, 217–230. [Google Scholar] [CrossRef]
  6. Plan, Y.; Vershynin, R. One-bit compressed sensing by linear programming. Commun. Pure Appl. Math. 2013, 66, 1275–1297. [Google Scholar] [CrossRef]
  7. Jacques, L.; Laska, J.; Boufounos, P.T.; Baraniuk, R.G. Roubust 1-bit compressive sensing via binary stable embeddings of sparse vectors. IEEE Trans. Inf. Theory 2013, 59, 2082–2102. [Google Scholar] [CrossRef]
  8. Yan, M.; Yang, Y.; Osher, S. Robust 1-bit compressive sensing using adaptive outlier pursuit. IEEE Trans. Signal. Process. 2012, 60, 3868–3875. [Google Scholar] [CrossRef]
  9. Dai, D.Q.; Shen, L.; Xu, Y.; Zhang, N. Noisy 1-bit compressive sensing: Models and algorithms. Appl. Comput. Harmon. Anal. 2016, 40, 1–32. [Google Scholar] [CrossRef]
  10. Huang, J.; Jiao, Y.; Lu, X.; Zhu, L. Robust decoding from 1-bit compressive sampling with ordinary and regularized least squares. SIAM J. Sci. Comput. 2018, 40, 2062–2786. [Google Scholar] [CrossRef]
  11. Zhou, S.; Luo, Z.; Xiu, N.; Li, G.Y. Computing one-bit compressive sensing via double-sparsity constrained optimization. IEEE Trans. Signal. Process. 2022, 70, 1593–1608. [Google Scholar] [CrossRef]
  12. Rockafellar, R.T.; Wets, R. Variational Analysis, Grundlehren der Mathematischen Wissenschaften; Springer: Berlin, Germany, 1998; Volume 317. [Google Scholar]
  13. Mordukhovich, B. Variational Analysis and Generalized Differentiation. I. Basic Theory; Grundlehren der Mathematischen Wissenschaften; Springer: Berlin, Germany, 2006; Volume 330. [Google Scholar]
  14. Bauschke, H.L.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces; AMS Books in Mathematics; Springer: New York, NY, USA, 2011. [Google Scholar]
  15. Björck, A. Numerical Methods for Least Squares Problems; SIAM: Philadelphia, PA, USA, 1996. [Google Scholar]
  16. Tikhonov, A.N.; Arsenin, V.Y. Solution of Ill-Posed Problems; V. H. Winston: Washington, DC, USA, 1977. [Google Scholar]
  17. Daubechies, I.; Defrise, M.; Mol, C.D. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Comm. Pure Appl. Math. 2004, 57, 1413–1457. [Google Scholar] [CrossRef]
  18. Figueiredo, M.A.T.; Nowak, R.D.; Wright, S.J. Gradient projection for sparse reconstruction: Application to compressed sensing and other inverse problems. IEEE J.-STSP. 2007, 1, 586–597. [Google Scholar] [CrossRef]
  19. Attouch, H.; Bolte, J.; Redont, P.; Soubeyran, A. Proximal alternating minimization and projection methods for nonconvex problems: An approach based on the Kurdyka-Łojasiewicz inequality. Math. Oper. Res. 2010, 35, 438–457. [Google Scholar] [CrossRef]
  20. Attouch, H.; Soubeyran, A. Inertia and reactivity in decision making as cognitive variational inequalities. J. Convex. Anal. 2006, 13, 207–224. [Google Scholar]
  21. Attouch, H.; Redont, P.; Soubeyran, A. A new class of alternating proximal minimization algorithms with costs to move. SIAM J. Optimiz. 2007, 18, 1061–1081. [Google Scholar] [CrossRef]
  22. Attouch, H.; Bolte, J.; Redont, P.; Soubeyran, A. Alternating proximal algorithms for weakly coupled convex minimization problems. Applications to dynamical games and PDE’s. J. Convex. Anal. 2008, 15, 485–506. [Google Scholar]
Figure 1. Average values of (a) SNR, (b) HE, and (c) HD over 50 trials vs. the noise level for all tested algorithms. We fixed m = n = 1000 and sparsity at 10.
Figure 1. Average values of (a) SNR, (b) HE, and (c) HD over 50 trials vs. the noise level for all tested algorithms. We fixed m = n = 1000 and sparsity at 10.
Mathematics 13 02926 g001
Figure 2. Average values of (a) SNR, (b) HE, and (c) HD over 50 trials vs. m / n for all tested algorithms. We fix n = 1000 , the noise level at 0.05 , and sparsity at 10.
Figure 2. Average values of (a) SNR, (b) HE, and (c) HD over 50 trials vs. m / n for all tested algorithms. We fix n = 1000 , the noise level at 0.05 , and sparsity at 10.
Mathematics 13 02926 g002
Figure 3. Average values of (a) SNR, (b) HE, and (c) HD over 50 trials vs. true sparsity s for all tested algorithms. We fix m = n = 1000 and the noise level at 0.05 .
Figure 3. Average values of (a) SNR, (b) HE, and (c) HD over 50 trials vs. true sparsity s for all tested algorithms. We fix m = n = 1000 and the noise level at 0.05 .
Mathematics 13 02926 g003
Table 1. Numerical results of Algorithm 1.
Table 1. Numerical results of Algorithm 1.
m = n saSNRHEHDSparsity of x ¯
1000100.0030.570.0010.00110
1000100.0125.940.0130.02110
1000100.0321.060.0270.05110
1000100.0520.400.0230.06510
1000100.0616.640.020.07610
1000100.0816.550.0360.10211
1000100.1018.190.0370.13011
1000100.1512.890.0660.18413
1000100.204.4800.1720.27217
2000200.0028.490.00550.005520
2000200.0123.320.02950.037520
2000200.0321.520.0190.04720
2000200.0520.470.02450.073520
2000200.0619.920.0260.08220
2000200.0816.090.03450.108519
2000200.1016.280.04450.135521
2000200.1511.780.07450.197518
2000200.205.0810.1530.37117
3000300.0029.640.0080.00830
3000300.0122.230.02130.030730
3000300.0320.200.02770.054330
3000300.0519.710.0250.071730
3000300.0617.380.03360.08530
3000300.0814.010.0570.121329
3000300.1014.190.0610.14232
3000300.158.620.10130.208734
3000300.206.3690.14230.267727
5000500.0022.360.0170.01750
5000500.0120.510.02780.036250
5000500.0320.040.02950.060250
5000500.0519.950.03020.076650
5000500.0617.140.03880.09249
5000500.0816.050.04280.116848
5000500.1014.320.05440.139646
5000500.1510.550.08920.206445
5000500.207.8920.11640.269647
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.-J.; Hu, Y.-H. Computing One-Bit Compressive Sensing via Alternating Proximal Algorithm. Mathematics 2025, 13, 2926. https://doi.org/10.3390/math13182926

AMA Style

Wang J-J, Hu Y-H. Computing One-Bit Compressive Sensing via Alternating Proximal Algorithm. Mathematics. 2025; 13(18):2926. https://doi.org/10.3390/math13182926

Chicago/Turabian Style

Wang, Jin-Jiang, and Yan-Hong Hu. 2025. "Computing One-Bit Compressive Sensing via Alternating Proximal Algorithm" Mathematics 13, no. 18: 2926. https://doi.org/10.3390/math13182926

APA Style

Wang, J.-J., & Hu, Y.-H. (2025). Computing One-Bit Compressive Sensing via Alternating Proximal Algorithm. Mathematics, 13(18), 2926. https://doi.org/10.3390/math13182926

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop