Open Access
This article is

- freely available
- re-usable

*Sensors*
**2016**,
*16*(5),
686;
https://doi.org/10.3390/s16050686

Article

Wireless Sensor Array Network DoA Estimation from Compressed Array Data via Joint Sparse Representation

^{1}

State Key Laboratory of Industrial Control Technology, Zhejiang University, Hangzhou 310027, China

^{2}

Institute of Information and Control, Hangzhou Dianzi University, Hangzhou 310018, China

^{3}

Shanghai Institute of Microsystem and Information Technology, Chinese Academy of Sciences, Shanghai 200050, China

^{4}

Institute of Acoustics, Chinese Academy of Sciences, Beijing 100190, China

^{5}

Department of Electrical and Computer Engineering, University of Wisconsin-Madison, Madison, WI 53706, USA

^{*}

Author to whom correspondence should be addressed.

Academic Editor:
Xue-Bo Jin

Received: 17 March 2016 / Accepted: 29 April 2016 / Published: 23 May 2016

## Abstract

**:**

A compressive sensing joint sparse representation direction of arrival estimation (CSJSR-DoA) approach is proposed for wireless sensor array networks (WSAN). By exploiting the joint spatial and spectral correlations of acoustic sensor array data, the CSJSR-DoA approach provides reliable DoA estimation using randomly-sampled acoustic sensor data. Since random sampling is performed at remote sensor arrays, less data need to be transmitted over lossy wireless channels to the fusion center (FC), and the expensive source coding operation at sensor nodes can be avoided. To investigate the spatial sparsity, an upper bound of the coherence of incoming sensor signals is derived assuming a linear sensor array configuration. This bound provides a theoretical constraint on the angular separation of acoustic sources to ensure the spatial sparsity of the received acoustic sensor array signals. The Cram$\stackrel{\xb4}{e}$r–Rao bound of the CSJSR-DoA estimator that quantifies the theoretical DoA estimation performance is also derived. The potential performance of the CSJSR-DoA approach is validated using both simulations and field experiments on a prototype WSAN platform. Compared to existing compressive sensing-based DoA estimation methods, the CSJSR-DoA approach shows significant performance improvement.

Keywords:

wireless sensor network; sensor array; compressive sensing; array processing## 1. Introduction

Direction of arrival (DoA) estimation using acoustic sensor arrays has attracted significant interest due to its wide applications [1,2,3,4]. Traditionally, DoA is realized using sensor arrays, such as passive towed-array sonar systems, where all sensors are wire-connected to a fusion center (FC) [5]. For wireless sensor array networks (WSAN) [6,7], arrays (locally wire/wireless-connected clusters) of dispersed sensors are deployed over large sensor fields and communicate wirelessly to the FC. However, remote sensor arrays often suffer from restricted resources, such as local power supply, local computational capacity and wireless transmission bandwidth. Therefore, data samples at remote sensor arrays need to be compressed before transmitting to the FC. On the other hand, executing sophisticated data compression algorithms on remote sensors is also restricted by local power consumption and computational ability.

To meet those resource requirements of the WSAN, the newly-emerged compressive sensing (CS) technology [8,9,10] has great potential because of its ability to reconstruct the raw signal from only a small number of random measurements. In this work, we propose a novel DoA estimation approach that accomplishes low power, robust DoA estimation over a WSAN platform, which reduces the volume of transmitted data without complicated local data compression operations. Incorporating a CS-based formulation, this approach is able to directly estimate the incident angles of acoustic sources from randomly-sampled acoustic signals. This is made possible by fully exploiting the joint spatial and spectral sparse structure of the acoustic signals acquired by the sensor array. Hence, we refer to this approach as compressive sensing joint sparse representation direction of arrival (CSJSR-DoA) estimation method.

Existing CS-based source DoA estimation methods can be categorized into two different approaches: compressive bearing estimation (COBE) [11,12,13] and compressive sensing array DoA estimation (CSA-DoA) [14]. With COBE, the incident angles (DoAs) are modeled as a sparse angle vector, with non-zero entries indicating the presence of (a few well separated) acoustic sources. The acoustic signals received at each sensor can then be represented as the product of a redundant steering matrix and a sparse angle vector. The steering matrix, consisting of steering vectors of all possible incident angles, must be estimated using the received noisy acoustic signal at a reference node. Thus, the raw data at the reference node and randomly-projected measurements of the non-reference nodes need to be transmitted to the remote FC. This imposes a heavy communication cost and excessive energy consumption at remote sensor arrays. With CSJSR-DoA, individual sensor nodes only perform data acquisition and random subsampling, while the DoA estimation is performed at the FC. Therefore, no local computation is required at the remote sensor nodes, and the CSJSR-DoA approach is more suitable for a resource-restricted WSAN system.

With the CSA-DoA [15,16,17] method, analog sensor array data are randomly projected onto a lower dimensional subspace in the analog domain before being converted into digital samples using an analog to digital converter (ADC). As such, fewer ADCs are required. This approach is similar to the CS-camera projection and was originally proposed in [18]. The steering vectors of different incident angles at a certain frequency are used to form the reconstruction dictionary. Further work extends this type of array processing method to broadband scenarios [19,20]. Since the data volume reduction is realized in the analog domain, special analog electronic circuits are required to implement the CSA-DoA approach. The proposed CSJSR-DoA approach, on the other hand, requires no special-purpose analog hardware. Some similar work [21] uses the angle domain sparsity of sources and formulates the narrow band signal of an antenna array within the BCS [22,23] framework. However, it is utilized as an alternative DoA estimation approach, and no data compression is considered.

Another distinction between the proposed CSJSR-DoA approach and these existing CS-based DoA estimation approaches is that the CSJSR-DoA approach exploits both the spatial and spectral domain structure of the acoustic sensor array signals. We argue that many practical broadband acoustic signals can be characterized by a few dominant frequency entries. With the purpose of DoA estimation, it would be sufficient to focus on these few dominant frequency entries and to exploit their frequency sparsity to enhance the performance. For example, Figure 1 shows the spectrograms of two types of broadband sources: an engine sound of a Porsche vehicle and a bird chirping [24]. It is clear that these broadband acoustic signals are dominated by multiple dominant frequency entries. The frequency domain sparse structure [25] illustrated in these figures may be utilized to realize efficient compressive sensing.

The main contributions are:

- A joint sparse representation-based DoA estimation approach is proposed, that exposes the joint spatial and spectral domain sparse structure of array signals and incorporates the multiple measurement vector [26] approach to solve the DoA estimation problem;
- A theoretical mutual coherence bound of a uniform linear sensor array is provided, which defines the minimum angular separation of the sources that are required for the CSJSR-DoA approach to yield a reliable solution with a high probability;
- The Cram$\stackrel{\xb4}{e}$r–Rao bound of the CSJSR-DoA estimator is derived to quantify the theoretical DoA estimation performance;
- A prototype acoustic WSAN platform is developed to validate the effectiveness of the proposed CSJSR-DoA approach.

The remainder of this paper is organized as follows. In Section 2, the array signal model, the sparsity model of narrowband array processing and the CS theory are briefly reviewed. The CSJSR-DoA approach is derived in Section 3. In Section 4, the theoretical analysis of the DoA estimation performance, as well as the Cram$\stackrel{\xb4}{e}$r–Rao bound are derived. In Section 5, the performance of the CSJSR-DoA approach is evaluated by simulations and field experiments using a prototype wireless sensor array platform. At last, the conclusion is summarized in Section 6.

## 2. Background

In this paper, the transposition and complex-conjugate transposition operations are denoted by superscripts ${(\xb7)}^{T}$ and ${(\xb7)}^{H}$, respectively. Lowercase and uppercase bold symbols denote vectors and matrices, respectively. The major symbols used in this paper are summarized in Table 1.

#### 2.1. Signal Model and Joint Spatial-Spectral Sparse Structure

A WSAN consists of one or more sensor arrays and one FC. Sensors within the same sensor array are arbitrarily deployed and connected to the FC via wireless channels. Sensor arrays are battery powered and have limited processing capabilities. The FC is connected to the infrastructure and has stronger processing capabilities. To conserve energy at remote sensor arrays, it is desirable to reduce the volume of data to be transmitted via wireless channels and to move as much of the processing tasks to the FC as possible.

For simplicity of notation, in this work, we assume a single sensor array consisting of H acoustic sensor nodes, and it communicates wirelessly via radio channels to a single FC. Each sensor node is equipped with an omni-directional microphone. The received acoustic signals will be sampled at a sampling frequency of ${f}_{s}$ Hz. N samples will be grouped into a frame (a snapshot) from which a DoA estimate will be made. We assume that state-of-the-art wireless synchronization protocols, such as Reference Broadcast Synchronization (RBS) [7], are applied so that clocks at sensor nodes can be synchronized at a precision in the order of micro-seconds, which is sufficiently accurate for acoustic DoA estimation.

We assume there are Q targets. The q-th source emits an acoustic signal and contains no more than ${R}_{q}$ ($N>{R}_{q}$) dominant frequency entries. Denote ${\mathbf{s}}_{q}$ to be a $N\times 1$ frame vector of the q-th source received at the reference node of the sensor array. Its N-point DFT (discrete Fourier transform) may be decomposed into two components:
where

$${\mathbf{s}}_{q}=\mathsf{\Psi}({\mathbf{r}}_{q}+{\mathbf{\nu}}_{q})$$

**Ψ**is an $N\times N$ inverse DFT matrix, ${\mathbf{r}}_{q}$ is an $N\times 1$ and ${R}_{q}$-sparse vector, containing no more than ${R}_{q}$ DFT coefficients with larger magnitude, and ${\mathbf{\nu}}_{q}$ contains the remaining smaller DFT coefficients.The h-th sensor node of the same sensor array will receive the same ${\mathbf{s}}_{q}$ with a relative delay ${\tau}_{h,q}$. Thus, the data received at the h-th sensor are the time-delayed summation of Q sources. That is:
where ${\omega}_{h}\left(t\right)$ is the zero mean white Gaussian noise at the h-th sensor.

$${x}_{h}\left(t\right)=\sum _{q=1}^{Q}{s}_{q}(t-{\tau}_{h,q})+{\omega}_{h}\left(t\right)$$

Figure 2 is the time delay model of a sensor array. Define ${\theta}_{q}$ as the incident angle of the q-th source, c as the speed of acoustic signal and $[{u}_{h},{v}_{h}]$ as the position of the h-th sensor with respect to the array centroid (${\sum}_{h=1}^{H}[{u}_{h},{v}_{h}]=[0,0]$). The time delay ${\tau}_{h,q}$ is given by:
where ${\rho}_{h}$ is the direction of the h-th sensor with respect to the array centroid. After exploiting N-point DFT to ${\mathbf{x}}_{h}={[{x}_{h}\left(1\right),\cdots ,{x}_{h}\left(N\right)]}^{T}$, its frequency domain expression is given by:
where ${\tilde{\mathbf{d}}}_{h}={\left[{d}_{h}\left(0\right),{d}_{h}\left(1\right),\cdots ,{d}_{h}(N-1)\right]}^{T}$, and each component ${d}_{h}\left(k\right)$ is equivalent to multiplying the k-th entry of the acoustic spectrum ${\mathbf{r}}_{q}+{\mathbf{\nu}}_{q}$ by a phase shift $exp(-j{\omega}_{k}{\tau}_{h,q})$. That is:
where ${w}_{h}\left(k\right)$ is the frequency domain expression of ${\omega}_{h}\left(t\right)$. Since ${\mathbf{x}}_{h}$ consists of all acoustic signals, the $N\times 1$ vector ${\tilde{\mathbf{d}}}_{h}$ will be R-sparse ($max\left({R}_{q}\right)\le R\le {\sum}_{q=1}^{Q}{R}_{q}$). This is because the dominant components of Q sources may fall within overlapped or different spectrum bands.

$${\tau}_{h,q}={d}_{h}cos({\theta}_{q}-{\rho}_{h})/c={d}_{h}\left[cos\left({\theta}_{q}\right)cos\left({\rho}_{h}\right)+sin\left({\theta}_{q}\right)sin\left({\rho}_{h}\right)\right]/c=\frac{1}{c}\left(cos\left({\theta}_{q}\right){u}_{h}+sin\left({\theta}_{q}\right){v}_{h}\right)$$

$${\tilde{\mathbf{d}}}_{h}={\mathbf{\Psi}}^{-1}{\mathbf{x}}_{h}$$

$${d}_{h}\left(k\right)=\sum _{q=1}^{Q}exp(-j{\omega}_{k}{\tau}_{h,q})({r}_{q}\left(k\right)+{\nu}_{q}\left(k\right))+{w}_{h}\left(k\right)$$

Consider the array data spectrum $\mathbf{d}\left(k\right)={[{d}_{1}\left(k\right),{d}_{2}\left(k\right),\cdots ,{d}_{H}\left(k\right)]}^{T}$ of H sensors at the k-th frequency; one may write:
where ${\mathbf{a}}_{q}\left(k\right)={[exp(-j{\omega}_{k}{\tau}_{1,q}),\cdots ,exp(-j{\omega}_{k}{\tau}_{H,q})]}^{T}$ is the steering vector of the q-th source, $\mathbf{w}\left(k\right)=[{w}_{1}^{\prime}\left(k\right),{w}_{2}^{\prime}\left(k\right),\cdots $$,{w}_{H}^{\prime}{\left(k\right)]}^{T}$, ${w}_{h}^{\prime}\left(k\right)={\sum}_{q=1}^{Q}exp(-j{\omega}_{k}{\tau}_{h,q}){\nu}_{q}\left(k\right)+{w}_{h}\left(k\right)$.

$$\mathbf{d}\left(k\right)=\sum _{q=1}^{Q}{\mathbf{a}}_{q}\left(k\right){r}_{q}\left(k\right)+\mathbf{w}\left(k\right)$$

The relative delay ${\tau}_{h,q}$ is the time for the acoustic wave of the q-th source traveling from the array centroid to the h-th node in the array along the incident angle ${\theta}_{q}$. Suppose that the entire range of incident angles is divided into $L(\gg Q)$ divisions with each division corresponding to a quantized incident angle ${\theta}_{\ell}=\ell /L$, $0\le \ell \le L-1$. We can construct a redundant $H\times L$ steering matrix ${\mathbf{A}}_{L}\left(k\right)=[{\mathbf{a}}_{1}\left(k\right),{\mathbf{a}}_{2}\left(k\right),\cdots ,{\mathbf{a}}_{L}\left(k\right)]$ that includes L steering vectors. Here, ${\mathbf{a}}_{\ell}\left(k\right)={[exp(-j{\omega}_{k}{\tau}_{1,\ell}),exp(-j{\omega}_{k}{\tau}_{2,\ell}),\cdots ,exp(-j{\omega}_{k}{\tau}_{H,\ell})]}^{T}$ is defined as the steering vector corresponding to an incident angle ${\theta}_{\ell}$. Furthermore, define a $L\times 1$, Q-sparse source energy vector $\mathbf{r}\left(k\right)$ such that:

$${r}_{\ell}\left(k\right)=\left\{\begin{array}{cc}{r}_{q}\left(k\right)\hfill & {\theta}_{\ell}\le {\theta}_{q}\le {\theta}_{\ell +1}\hfill \\ 0\hfill & \mathrm{otherwise}.\hfill \end{array}\right.$$

If L is sufficiently large, it is safe to assume that different ${\theta}_{q}$ will fall into different angle bins. With this change of notation, Equation (6) can be rewritten as:
where $\mathbf{r}\left(k\right)={[{r}_{1}\left(k\right),{r}_{2}\left(k\right),\cdots ,{r}_{L}\left(k\right)]}^{T}$. With Equations (4) and (6), one may construct an $H\times N$ matrix:

$$\mathbf{d}\left(k\right)={\mathbf{A}}_{L}\left(k\right)\mathbf{r}\left(k\right)+\mathbf{w}\left(k\right)$$

$$\mathbf{D}=\left[\mathbf{d}\left(0\right),\mathbf{d}\left(1\right),\cdots ,\mathbf{d}(N-1)\right]={\left[{\tilde{\mathbf{d}}}_{1},{\tilde{\mathbf{d}}}_{2},\cdots ,{\tilde{\mathbf{d}}}_{H}\right]}^{T}$$

By performing row-vectorization of matrix $\mathbf{D}$, we have an $H\xb7N\times 1$ vector:

$$\tilde{\mathbf{d}}={\left[{\tilde{\mathbf{d}}}_{1}^{T},{\tilde{\mathbf{d}}}_{2}^{T},\cdots ,{\tilde{\mathbf{d}}}_{H}^{T}\right]}^{T}$$

Similarly, by performing column-vectorization of matrix $\mathbf{D}$, we have another $H\xb7N\times 1$ vector:

$$\overline{\mathbf{d}}={\left[{\mathbf{d}}^{T}\left(0\right),{\mathbf{d}}^{T}\left(1\right),\cdots ,{\mathbf{d}}^{T}(N-1)\right]}^{T}$$

Since the entries of both $\tilde{\mathbf{d}}$ and $\overline{\mathbf{d}}$ are associated with the same $\mathbf{D}$ matrix, there exists an $H\xb7N\times H\xb7N$ permutation matrix $\mathbf{M}$ such that:

$$\tilde{\mathbf{d}}=\mathbf{M}\overline{\mathbf{d}}$$

Equation (12) provides a link between $\tilde{\mathbf{d}}$, which exhibits the R-sparsity in the frequency domain, and $\overline{\mathbf{d}}$, which exploits the Q-sparsity in the spatial (incident angle) domain. In other words, the array samples, as summarized in matrix $\mathbf{D}$, exhibit a joint frequency and spatial (angle) domain sparsity that will be exploited later in this paper.

#### 2.2. Compressive Sensing and Random Sub-Sampled Measurements

The presence of sparsity, as described in the models above, promises great potential for introducing CS to reduce the amount of transmitted sensor data.

The compressive sensing theory [27] states that a signal $\mathbf{x}$ may be reconstructed perfectly from a dimension reduced measurement $\mathbf{y}\in {R}^{M\times 1}$ through a linear measurement system $\mathbf{y}=\mathsf{\Phi}\mathbf{x}+\mathbf{n}$ when $\mathbf{\alpha}={\mathsf{\Psi}}^{-1}\mathbf{x}$ is a K-sparse vector. This is often accomplished by solving a constrained optimization problem:
where ${\left|\right|\mathbf{\alpha}\left|\right|}_{1}$ is the ${\ell}_{1}$ norm of vector $\mathbf{\alpha}$,

$${min\phantom{\rule{1.0pt}{0ex}}\phantom{\rule{1.0pt}{0ex}}\left|\right|\mathbf{\alpha}\left|\right|}_{1}\phantom{\rule{1.0pt}{0ex}}\phantom{\rule{1.0pt}{0ex}}\phantom{\rule{1.0pt}{0ex}}s.t.\phantom{\rule{1.0pt}{0ex}}\phantom{\rule{1.0pt}{0ex}}\left|\right|\phantom{\rule{1.0pt}{0ex}}\mathbf{y}-\mathsf{\Phi}\mathsf{\Psi}\mathbf{\alpha}\left|\right|<\sigma $$

**Φ**is the measurement matrix,**Ψ**is the sparse matrix and σ is a noise threshold. To confirm stable reconstruction, the measurement matrix**Φ**is usually chosen to be a random Gaussian matrix or a random binary matrix.The measurement vector in traditional compressive sensing problem settings is often computed as a weighted linear combination of the observed high dimensional signals. For example, in

**COBE**[13], the $N\times 1$ sensor data vector in a frame will be multiplied to an $M\times N$ measurement matrix ($M<\phantom{\rule{3.33333pt}{0ex}}N$) digitally to yield a measurement vector**y**and then transmitted via a wireless channel to the FC. This obviously requires much computation and is likely to further deplete the energy reserve on sensor nodes. In**CSA-DoA**[15], an analog filter needs to be inserted before the ADC to realize the measurement operation. This approach requires special-purpose hardware and can be quite expensive.In this work, time domain sensor data will be purposely discarded and will not be transmitted to the FC. By doing so, data reduction may be achieved without incurring expensive compression computations on the sensor nodes. Specifically, we will use a random sub-sampling matrix [28] defined on a set of non-uniform, yet known random sampling intervals $\{{\tau}_{m};1\le m\le M\}$. Let a sequence be $\mathbf{u}=\left\{u\right(1),u(2),\cdots ,u(M\left)\right\}$ such that $u\left(1\right)=1$, and:
where $\mathrm{r}(\xb7)$ is the rounding operation. Then, the $(m,n)$-th element of this proposed random subsampling matrix is given by:

$$u\left(m\right)=u(m-1)+\mathrm{r}\left({\tau}_{m}\right),\phantom{\rule{1.em}{0ex}}2\le m\le M$$

$$\begin{array}{c}\mathsf{\Phi}(m,n)=\left\{\begin{array}{cc}1\hfill & 1\le n=u\left(m\right)\le N\hfill \\ 0\hfill & \mathrm{otherwise}\hfill \end{array}\right.\hfill \end{array}$$

The selection of the proposed random subsampling matrix is because the product of $\mathsf{\Phi}\mathsf{\Psi}$ is a partial Fourier matrix, which has been proven to satisfy the restricted isometry property (RIP) property with a high probability [29].

Furthermore, in this work, the lossy nature of wireless channels will be exploited by modeling the packet loss during wireless transmission over noisy channels as a form of (involuntary) random data sub-sampling. It is well known that due to different link types (amplify forward and direct link) [30], power allocation [31] and time-varying channel conditions, wireless transmissions are likely to suffer from packet loss, and the identities of lost data packets are known at the FC. Hence, this type of packet loss can be modeled as another form of random sub-sampling. In our recent work [32], the data packet loss is modeled with a random selection matrix.

In this work, both the random sub-sampling matrix and the random selection matrix will be incorporated as a combined measurement matrix that requires no computation on the sensor nodes while achieving the desired data reduction. Denote the random sampling matrix at the h-th sensor node as ${\mathsf{\Phi}}_{h}$ and the random selection matrix over the wireless channels between the h-th sensor node and the FC as ${\mathsf{\Phi}}_{loss}^{h}$. The combined equivalent measurement matrix may then be obtained as:

$${\mathsf{\Phi}}_{e}^{h}={\mathsf{\Phi}}_{loss}^{h}\xb7{\mathsf{\Phi}}_{h}.$$

## 3. CSJSR-DoA Formulation

The proposed CSJSR-DoA algorithm consists of three major steps: (1) at the h-th wireless sensor node, the received and digitized acoustic signal ${\mathbf{x}}_{h}$ will be sub-sampled using random compressive sampling as described in Equation (15) to yield a node measurement vector ${\mathsf{\Phi}}_{h}{\mathbf{x}}_{h}$; (2) these node measurement vectors will then be transmitted through lossy wireless channels to the FC; (3) The FC will receive an overall measurement vector $\mathbf{y}={\left[{\mathbf{y}}_{1}^{T},{\mathbf{y}}_{2}^{T},\cdots ,{\mathbf{y}}_{H}^{T}\right]}^{T}$, which will then be processed by the CSJSR-DoA approach to directly estimate a sparse DoA indicator vector. The schematic diagram of the proposed CSJSR-DOA approach is summarized in Figure 3.

Note that at the FC, the received compressive measurement of the h-th sensor node may be expressed as:

$${\mathbf{y}}_{h}={\mathsf{\Phi}}_{loss}\xb7{\mathsf{\Phi}}_{h}{\mathbf{x}}_{h}={\mathsf{\Phi}}_{e}^{h}\mathsf{\Psi}{\tilde{\mathbf{d}}}_{h}$$

Hence, the overall measurement vector $\mathbf{y}$ may be expressed as:
or in matrix form,
where $\mathsf{\Theta}=\mathrm{diag}({\mathsf{\Phi}}_{e}^{1}\mathsf{\Psi},{\mathsf{\Phi}}_{e}^{2}\mathsf{\Psi},\cdots ,{\mathsf{\Phi}}_{e}^{H}\mathsf{\Psi})$ is the joint sparse representation matrix. As discussed earlier, $\{{\tilde{\mathbf{d}}}_{h},1\le h\le H\}$ shares a common R-sparse structure, since the signal measured at one sensor node would be a time-delayed version of those at other sensors.

$$\left[\phantom{\rule{-0.166667em}{0ex}}\begin{array}{c}{\mathbf{y}}_{1}\\ {\mathbf{y}}_{2}\\ \vdots \\ {\mathbf{y}}_{H}\end{array}\phantom{\rule{-0.166667em}{0ex}}\right]\phantom{\rule{-0.166667em}{0ex}}=\phantom{\rule{-0.166667em}{0ex}}\left[\phantom{\rule{-0.166667em}{0ex}}\begin{array}{cccc}\phantom{\rule{-0.166667em}{0ex}}{\mathsf{\Phi}}_{e}^{1}\mathsf{\Psi}\phantom{\rule{-0.166667em}{0ex}}& & & \\ & \phantom{\rule{-0.166667em}{0ex}}{\mathsf{\Phi}}_{e}^{2}\mathsf{\Psi}\phantom{\rule{-0.166667em}{0ex}}& & \\ & & \phantom{\rule{-0.166667em}{0ex}}\ddots \phantom{\rule{-0.166667em}{0ex}}& \\ & & & \phantom{\rule{-0.166667em}{0ex}}{\mathsf{\Phi}}_{e}^{H}\mathsf{\Psi}\phantom{\rule{-0.166667em}{0ex}}\end{array}\phantom{\rule{-0.166667em}{0ex}}\right]\phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}\left[\phantom{\rule{-0.166667em}{0ex}}\begin{array}{c}{\tilde{\mathbf{d}}}_{1}\\ {\tilde{\mathbf{d}}}_{2}\\ \vdots \\ {\tilde{\mathbf{d}}}_{H}\end{array}\phantom{\rule{-0.166667em}{0ex}}\right]$$

$$\mathbf{y}=\mathsf{\Theta}\tilde{\mathbf{d}}$$

Next, from Equations (8) and (11), one may write:
where ${\mathbf{r}}_{L}={[{\mathbf{r}}_{L}{\left(0\right)}^{T},{\mathbf{r}}_{L}{\left(1\right)}^{T},\cdots ,{\mathbf{r}}_{L}{(N-1)}^{T}]}^{T}$, $\mathbf{w}={[\mathbf{w}{\left(0\right)}^{T},\mathbf{w}{\left(1\right)}^{T},\cdots ,\mathbf{w}{(N-1)}^{T}]}^{T}$, ${\mathbf{A}}_{L}$ is the $HN\times NL$ joint steering matrix:

$$\overline{\mathbf{d}}={\mathbf{A}}_{L}{\mathbf{r}}_{L}+\mathbf{w},$$

$${\mathbf{A}}_{L}=\left[\phantom{\rule{-0.166667em}{0ex}}\begin{array}{cccc}\phantom{\rule{-0.166667em}{0ex}}{\mathbf{A}}_{L}\left(0\right)\phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\\ \phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}{\mathbf{A}}_{L}\left(1\right)\phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\\ \phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\ddots \phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\\ \phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}{\mathbf{A}}_{L}(N-1)\phantom{\rule{-0.166667em}{0ex}}\end{array}\phantom{\rule{-0.166667em}{0ex}}\right]$$

Substituting Equations (12) and (20) into Equation (19) leads to:
where ${\mathbf{w}}_{c}=\mathsf{\Theta}\mathbf{H}\mathbf{w}$ is the subsampled Gaussian white noise received in the FC. Here, $\mathsf{{\rm Y}}=\mathsf{\Theta}\mathbf{M}{\mathbf{A}}_{L}$ is the joint sparse representation matrix that combines the joint sparse matrices
where ${\mathbf{\varphi}}_{e,n}^{h}$ is the n-th column of ${\mathsf{\Phi}}_{e}^{h}\mathsf{\Psi}$ and ${\mathbf{a}}_{L,h}\left(n\right)$ is the h-th row of ${\mathbf{A}}_{L}\left(n\right)$. The significance of Equation (22) is that one may reconstruct the joint sparse indicating vector ${\mathbf{r}}_{L}$ from the joint measurement vector $\mathbf{y}$ directly without reconstructing the raw data ${\mathbf{x}}_{h},1\le h\le H$ at individual sensor nodes. ${\mathbf{r}}_{L}$ is a joint sparse vector with no more than R non-zero blocks, and there are at most Q non-zero entries in each of the R non-zero blocks.

$$\begin{array}{cc}\hfill \mathbf{y}& =\mathsf{\Theta}\tilde{\mathbf{d}}=\mathsf{\Theta}\mathbf{M}\overline{\mathbf{d}}\hfill \\ \hfill & =\mathsf{\Theta}\mathbf{M}({\mathbf{A}}_{L}{\mathbf{r}}_{L}+\mathbf{w})=\mathsf{{\rm Y}}{\mathbf{r}}_{L}+{\mathbf{w}}_{c}\hfill \\ \hfill \end{array}$$

**Θ**and ${\mathbf{A}}_{L}$. The details of this joint sparse representation matrix are:
$$\mathsf{{\rm Y}}\phantom{\rule{-0.166667em}{0ex}}=\phantom{\rule{-0.166667em}{0ex}}\underset{\mathsf{{\rm Y}}\left[1\right]}{\underbrace{\left[\phantom{\rule{-0.166667em}{0ex}}\begin{array}{c}{\mathbf{\varphi}}_{e,1}^{1}{\mathbf{a}}_{L,1}^{T}\left(0\right)\\ {\mathbf{\varphi}}_{e,1}^{2}{\mathbf{a}}_{L,2}^{T}\left(0\right)\\ \vdots \\ {\mathbf{\varphi}}_{e,1}^{H}{\mathbf{a}}_{L,H}^{T}\left(0\right)\end{array}\right.}}\underset{\mathsf{{\rm Y}}\left[2\right]}{\underbrace{\begin{array}{c}{\mathbf{\varphi}}_{e,2}^{1}{\mathbf{a}}_{L,1}^{T}\left(1\right)\\ {\mathbf{\varphi}}_{e,2}^{2}{\mathbf{a}}_{L,2}^{T}\left(1\right)\\ \vdots \\ {\mathbf{\varphi}}_{e,2}^{H}{\mathbf{a}}_{L,H}^{T}\left(1\right)\end{array}}}\begin{array}{c}\\ \\ \ddots \\ \end{array}\underset{\mathsf{{\rm Y}}\left[N\right]}{\underbrace{\left.\begin{array}{c}{\mathbf{\varphi}}_{e,N}^{1}{\mathbf{a}}_{L,1}^{T}\left({\scriptscriptstyle N-1}\right)\\ {\mathbf{\varphi}}_{e,N}^{2}{\mathbf{a}}_{L,2}^{T}\left({\scriptscriptstyle N-1}\right)\\ \vdots \\ {\mathbf{\varphi}}_{e,N}^{H}{\mathbf{a}}_{L,H}^{T}\left({\scriptscriptstyle N-1}\right)\end{array}\phantom{\rule{-0.166667em}{0ex}}\right]\phantom{\rule{-0.166667em}{0ex}}}}$$

With these relations, we may now formulate the joint sparse representation-based DoA estimation problem (group sparse reconstruction [33]) in a multiple measurements vector (MMV) formulation [26,34,35]:
where $\mathbf{s}\left(\ell \right)={\sum}_{n=0}^{N-1}{\mathbf{r}}_{\ell}{\left(n\right)}^{2}$ and γ is a noise threshold. To solve this optimization problem, a second-order cone programming (SOCP) [36] approach has been proposed. The joint sparse representation of the wideband array signal in Equation (23) helps to improve the DoA estimation performance of wideband signals, as long as the same DoA of a signal is shared by different frequencies [37]. Note that the group sparsity-based DoA estimation for wideband signals is advantageous as compared to the traditional wideband DoA estimation methods, because it achieves near-coherent DoA estimation performance without the requirement of coherent signal projection through, e.g., the well-known focusing technique [38].

$$\underset{{\mathbf{r}}_{L}}{min}\sum _{\ell =1}^{L}\left|\mathbf{s}\left(\ell \right)\right|\phantom{\rule{4.pt}{0ex}}\text{subject}\phantom{\rule{4.pt}{0ex}}\text{to}\phantom{\rule{4.pt}{0ex}}\left|\right|\mathbf{y}-\mathbf{{\rm Y}}{\mathbf{r}}_{L}{\left|\right|}_{2}\le \gamma $$

However, the computation cost is rather expensive. In [39], an MMV-Proxmethod has been proposed. We adopt this method for the underlying problem. In particular, we solve Equation (23) by a two-step process:

- Estimate the frequency domain support $\mathbf{T}$ of array signals by reconstructing the spectral sparse indicative vector $\tilde{\mathbf{d}}$ from the joint measurements $\mathbf{y}$ and prune the joint reconstruction matrix
**Υ**by selecting $\mathsf{{\rm Y}}\left[r\right]$ with non-zero frequency bins; - reconstruct the DoA indicative vector $\mathbf{s}$ directly from the joint measurements $\mathbf{y}$ using the pruned joint reconstruction matrix.

In summary, the first step is to estimate the non-zero block ${\mathbf{r}}_{L}\left(k\right)$ and to simplify the joint sparse matrix by selecting $\mathsf{{\rm Y}}\left[r\right]$ with non-zero ${\mathbf{r}}_{L}\left(r\right)$. The second step is to reconstruct the DoA using the pruned joint sparse representation matrix $\tilde{\mathsf{{\rm Y}}}$. These steps are summarized in the listing of Algorithm 1.

The computational complexity of Equation (23) in the SOCP framework using an interior point implementation is $\mathcal{O}\left({L}^{3}{N}^{3}\right)$ [39]. If we decouple the original problem into two sub-problems (Steps 2, 3 and 4 in the above algorithm), its computational complexity can be reduced to $\mathcal{O}({L}^{3}{R}^{3}+{N}^{3}{H}^{3})$.

Algorithm 1 Decoupled joint sparse reconstruction. |

Input: Received joint random measurement vector $\mathbf{y}$; |

Equivalent random measurement matrix ${\mathsf{\Phi}}_{e}^{h}$; |

Output: DoA indicating vector $\mathbf{s}$; |

1. Estimate the noise level by random sampling in source free scenario, $\gamma =\left|\right|\mathbf{y}|{|}_{2}^{2}$; |

2. Estimate the common support $\mathbf{T}=\mathbf{supp}\left(\tilde{\mathbf{d}}\right)$ by solving: |

$\begin{array}{c}\underset{\tilde{\mathbf{d}}}{min}{\displaystyle \sum _{n=0}^{N-1}}{\displaystyle \sum _{h=1}^{H}}{d}_{h}{\left(n\right)}^{2}\hfill \\ \text{subject to}\left|\right|\mathbf{y}-\mathsf{\Theta}\tilde{\mathbf{d}}{\left|\right|}_{2}^{2}\le \gamma ;\hfill \end{array}$ |

3. Construct the pruned joint reconstruction matrix $\tilde{\mathsf{{\rm Y}}}$; |

$\tilde{\mathsf{{\rm Y}}}=\left[\mathsf{{\rm Y}}\left[{n}_{1}\right],\mathsf{{\rm Y}}\left[{n}_{2}\right],\cdots ,\mathsf{{\rm Y}}\left[{n}_{\left|\mathbf{T}\right|}\right]\right],r\in \mathbf{T}$; |

4. Solve the pruned reconstruction problem: |

$\begin{array}{c}\underset{{\overline{\mathbf{r}}}_{L}}{min}{\parallel \mathbf{s}\parallel}_{1}\text{subject to}\left|\right|\mathbf{y}-\tilde{\mathsf{{\rm Y}}}\overline{\mathbf{r}}{\left|\right|}_{2}^{2}\le \gamma ,\hfill \\ \mathbf{s}\left(\ell \right)={\sum}_{n\in \mathbf{T}}{\mathbf{r}}_{\ell}{\left(n\right)}^{2},\overline{\mathbf{r}}={[{\mathbf{r}}^{T}\left({n}_{1}\right),\cdots ,{\mathbf{r}}^{T}\left({n}_{\left|T\right|}\right)]}^{T};\hfill \end{array}$ |

## 4. Performance Analysis

In this section, the performance of the proposed CSJSR-DoA approach is investigated. First, the relevant coherence, its upper bound and the angular separation problem are studied. Second, the Cram$\stackrel{\xb4}{e}$r–Rao bound of the proposed CSJSR-DoA approach will be discussed.

#### 4.1. Sparse Reconstruction Analysis and Angle Separation

The existence of solution to Equation (23) depends on the property of the reconstruction dictionary

**Υ**. Usually, the mutual coherence and the restricted isometry property (RIP) of a sparse representation matrix are used as the criteria that indicate the reconstruction performance. Consider that the verification of the RIP property requires combinational computational complexity [40]; it is preferable to use the property of coherence that is easily computable to provide more concrete recovery guarantees. In this paper, we try to analyze the coherence of the CSJSR-DoA approach.In the CS theory, the coherence of a matrix is the worst-case linear dependence of any pair of its column vectors. Similarly, in the block sparse signal case, in which a signal is composed by several blocks and only a small number of these blocks is non-zero, the block-coherence [41] has been proposed.

Recall from Equation (21):
where

$$\mathsf{{\rm Y}}=\mathsf{\Theta}\mathbf{M}{\mathbf{A}}_{L}$$

**Θ**is a block diagonal joint sparse sampling matrix, $\mathbf{M}$ is a full rank, unitary permutation matrix and ${\mathbf{A}}_{L}$ is a block diagonal steering matrix. To study the coherence of**Υ**, we need to study the block-coherence of**Υ**.**Theorem 1.**

Assume that H sensor nodes form a sensor array, and the received data in the fusion center are ${\mathbf{y}}_{h}={\mathsf{\Phi}}_{e}^{h}{\mathbf{x}}_{h}({\mathsf{\Phi}}_{e}^{h}\in {R}^{M\times N}),h=1,2,\cdots ,H$. Then, the block-coherence of the joint sparse representation matrix

**Υ**satisfies:
$${\mu}_{B}\left(\mathsf{{\rm Y}}\right)\le \sqrt{\frac{N-M}{(N-1)M}}$$

**Proof.**

See Appendix A. ☐

The block-coherence ${\mu}_{B}$ guarantees stable reconstruction of the block sparse signal ${\mathbf{r}}_{L}$ with a high probability. However, each block ${\mathbf{r}}_{L}\left(n\right)$ indicates the DoA of sources. Thus, the coherence of each sub-matrix $\mathsf{{\rm Y}}\left[n\right]$ should be considered. In this paper, we derive the coherence of $\mathsf{{\rm Y}}\left[n\right]$. This result will provide a lower bound of the source separation (in turns of DoA angle) issue to ensure a reliable DoA estimation. However, the coherence of $\mathsf{{\rm Y}}\left[n\right]$ depends on array geometry, and it is difficult to obtain an analytical bound.

In this paper, we consider the typical uniform linear array geometry and find an interesting result of the coherence bound. The derived result is similar to the array pattern [42] analysis and is validated for both narrowband and broadband acoustic sources. For broadband sources, the highest dominant frequency may be used.

**Theorem 2.**

Assume that H sensor nodes form a uniformly-spaced linear array with $d(\le \lambda /2)$ being the spacing between adjacent sensors where λ is the wavelength of the acoustic signal. If the difference of the incident angles of any pair of sources satisfies:
then the coherence of $\mathsf{{\rm Y}}\left[n\right]$ is bounded by:

$$\mathsf{\Delta}\theta \ge \frac{c}{Hdf}$$

$$\mu \left(\mathsf{{\rm Y}}\left[n\right]\right)\le \frac{1}{Hsin(\pi /H)}\sqrt{\frac{N-M}{(N-1)M}}$$

**Proof.**

See Appendix B. ☐

#### 4.2. Cram$\stackrel{\xb4}{e}$r–Rao Bound of CSJSR-DoA

In this subsection, we derive the Cram$\stackrel{\xb4}{e}$r–Rao bound (CRB) of the proposed CSJSR-DoA method and study its impact on the DoA estimation accuracy due to factors such as the number of measurements, the number of sensors and the noise level ${\sigma}^{2}$, which is defined in Equation (13).

Recall that the CRB is defined as the inverse of the Fisher information matrix (FIM) [43],
where

$$\mathbf{CRB}\left(\mathsf{\Lambda}\right)={\mathbf{F}}^{-1}\left(\mathsf{\Lambda}\right)$$

**Λ**is the unknown parameter set. In this paper, we only derive the CRB of a single source case. Thus, the unknown parameter is $\mathsf{\Lambda}={[\theta ,{\mathbf{r}}^{T},{\mathbf{f}}^{T}]}^{T}$, where $\mathbf{f}={[{f}_{1},{f}_{2},\cdots ,{f}_{R}]}^{T}$ and $\mathbf{r}={[{r}_{0}\left({k}_{1}\right),{r}_{0}\left({k}_{2}\right),\cdots ,{r}_{0}\left({k}_{R}\right)]}^{T}$ are the sparse frequency bins and their corresponding source spectrum, respectively.In the case of Gaussian white noise, the FIM is:
where $\mathbf{G}=\partial \mathbf{y}/\partial {\mathsf{\Lambda}}^{T}$. For simplification, we assume that all H sensors have the same ${\mathsf{\Phi}}_{e}({\mathsf{\Phi}}_{e}^{1}=,\cdots ,={\mathsf{\Phi}}_{e}^{H}={\mathsf{\Phi}}_{e})$. Recall Equations (4), (6) and (17); the measurement vector of the CSJSR-DoA approach in a single source case is:
where ${\mathbf{\varphi}}_{e,{k}_{i}}$ is the ${k}_{i}$-th column of ${\mathsf{\Phi}}_{e}\mathsf{\Psi}$. After some calculations, the CRB of the CSJSR-DoA approach is given by:
where $\mathbf{e}={[{f}_{1}{r}_{0}\left({k}_{1}\right),{f}_{2}{r}_{0}\left({k}_{2}\right),\cdots ,{f}_{R}{r}_{0}\left({k}_{R}\right)]}^{T}$, $\alpha =\frac{4{\pi}^{2}}{{c}^{2}}{\sum}_{h=0}^{H-1}{\left({u}_{h}cos\left(\theta \right)-{v}_{h}sin\left(\theta \right)\right)}^{2}$.

$$\mathbf{F}\left(\mathsf{\Lambda}\right)=\frac{{\sigma}^{2}}{2}\mathrm{Re}\left[{\mathbf{G}}^{H}\mathbf{G}\right]$$

$$\mathbf{y}\left(\mathsf{\Lambda}\right)=\sum _{i=1}^{R}{r}_{0}\left({k}_{i}\right){\mathbf{a}}_{0}\left({k}_{i}\right)\otimes {\mathbf{\varphi}}_{e,{k}_{i}}+{\mathbf{w}}_{c}$$

$$CRB\left(\theta \right)=\frac{{\sigma}^{2}}{{2M\alpha \parallel \mathbf{e}\parallel}_{2}^{2}}$$

**Proof.**

See Appendix C. ☐

## 5. Performance Evaluation

In this section, extensive simulations are carried out to: (1) study the performance of the CSJSR-DoA approach; and (2) compare the performance of the CSJSR-DoA approach against COBE and CSA-DoA, using the L1-SVD [26] algorithm as the baseline. In addition, we also developed a hardware prototype platform and collected data from outdoor experiments to validate the practical applicability of the CSJSR-DoA approach.

#### 5.1. Simulation Settings

We synthesized acoustic signals based on Equation (1). The synthesized signals contain four dominant frequency entries at 300 Hz, 500 Hz, 600 Hz and 800 Hz. These correspond to wavelengths of $1.14$, $0.7$, $0.57$ and $0.43$ m, respectively, assuming the sound speed at 343 m/s. The received acoustic signal also contains additive Gaussian white noise with zero mean. The variances are set so that the resulting SNR ranges from $-10$ dB to 10 dB in 5-dB increments. Such an acoustic signal resembles the dominant entries’ distribution of some practical acoustic sources, such as truck engine sounds. The sampling rate is 2 ksamples/s. At this rate, a frame of 125 ms is selected. The sampling rate is chosen so that it satisfies the Nyquist sampling rate and avoids the frequency aliasing issue.

We assume a single uniform linear array consisting of six ($H=6$) acoustic sensor nodes deployed in a sensor field. The spacing between adjacent sensors is 0.2 m, which is smaller than half of the wavelength ($0.43$ m) of the 800-Hz entry. All acoustic sources are located at the broad side of this linear array, so that the incident angles (directions of arrivals) are constrained within a $-{90}^{\circ}$ to ${90}^{\circ}$ range. We divide this angle range into ($L=360$) partitions, so that each partition equals $0.{5}^{\circ}$.

Consider that data packets transmitted from the sensor array to the FC may suffer from data packet loss. We will simulate data packet loss rates, denoted by ${r}_{\mathrm{loss}}$, at 0, $5\%$, $10\%$, $20\%$ and $30\%$, respectively. To mitigate burst data transmission loss, the acoustic data stream will be interleaved while being assembled into the data packet for transmission. However, other than packet headers, there will be no additional channel coding bits appended. The data loss rate is computed as the percentage of packets that are lost during transmission through wireless channels versus the total number of packets that are sent from the sensor array.

With the CSJSR-DoA approach, both random subsampling and wireless channel packet loss will reduce the amount of received acoustic samples compared to what was originally sampled at the sensor array. In particular, at each sensor node, the acoustic data samples will be randomly subsampled according to Equations (14) and (15). The ratio $M/N$ in these equations can be regarded, in the context of CS, as the ratio of the number of measurements versus the total number of data samples. With channel data packet loss taken into account, we instead define ${r}_{\mathrm{dc}}(\le 1)$ to be the ratio of the number of acoustic samples successfully received at the fusion center versus the total number of samples acquired by the H acoustic sensors. Ignoring the overhead of data packet header, we have:

$${r}_{\mathrm{dc}}\approx (1-{r}_{\mathrm{loss}})\times M/N$$

In this section, the DoA performance will be reported against different values of ${r}_{\mathrm{dc}}$. In practice, if ${r}_{\mathrm{loss}}$ can be estimated in real time, we may adjust M to achieve the desired DoA accuracy.

Four algorithms are implemented: L1-SVD, CSJSR-DoA, CSA-DoA and COBE. The L1-SVD algorithm [26] is applied to the entire set of acoustic samples without random subsampling and is used as a baseline for both simulation and experiment. The COBE algorithm [13] requires a specific reference node to be sampled at the Nyquist rate without any subsampling in order to provide a reference source signal. To have fair comparisons, we performed trial runs to empirically obtain the best parameters for the CSA-DoA and COBE algorithms. All four algorithms are implemented using MATLAB Version 7.9.0. The optimization is performed using the Sedumi 1.2.1 toolbox [44].

#### 5.2. CSJSR-DoA Performance Evaluation

In this simulation, two ($Q=2$) stationary acoustic sources are placed in the far field from the sensor array at $[-{60}^{\circ},{30}^{\circ}]$. The noise levels of the acoustic signal correspond to SNR = $-10$ dB, $-5$ dB, 0 dB, 5 dB and 10 dB, and it is assumed that ${r}_{\mathrm{loss}}=0\%$. Five hundred independent trials are performed. Consider that sparse reconstruction is a probability event; a DoA estimate is considered a success detection if a sparse indicative vector can be reconstructed successfully. In other words, some trials fail to give a solution to the sparse DoA indicative vector. Based on the results of detected trials, the number of impinging signals, Q, and the corresponding DoAs can be estimated. Similar to some spatial spectrum search-based approaches, the spatial spectrum will be calculated using ${\mathbf{r}}_{\mathrm{ss}}=20{log}_{10}\left(\mathbf{s}/max\left(\mathbf{s}\right)\right)$, and some conventional peak-finding approaches [45,46] are used to find $\tilde{Q}$ peaks of the obtained spatial spectrum. However, the estimated $\tilde{Q}$ directions can be divided into three categories: $\tilde{Q}<Q$, $\tilde{Q}=Q$ and $\tilde{Q}>Q$. In this case, a proper performance criterion is hard to choose. Following a similar work [22], which takes into account both the errors in estimating the signal number $\tilde{Q}$ and the corresponding DoAs, the root mean square errors (RMSE) of the n-th detected trial will also be reported using the formula:
where ${N}_{e}$ is the number of detected trials, ${\overline{\theta}}_{j}^{\left(n\right)}=arg\left\{\underset{{\theta}_{q},q\in [1,Q]}{min}\left|{\theta}_{q}-{\tilde{\theta}}_{j}^{\left(n\right)}\right|\right\}$. Here, $\mathsf{\Delta}{\theta}_{max}$ is a penalty term of the maximum admissible DoA error (i.e., ${180}^{\circ}$ for a liner array), and ${\overline{\theta}}_{j}^{\left(n\right)}$ is the least error of additional false DoA.

$$\mathrm{RMSE}=\sqrt{\frac{{\displaystyle \sum _{n=1}^{{N}_{e}}}{\left({\mathrm{RMSE}}^{\left(n\right)}\right)}^{2}}{{N}_{e}}}$$

$$\mathrm{RMS}{\mathrm{E}}^{\left(n\right)}=\left\{\begin{array}{c}\sqrt{\frac{{\sum}_{q=1}^{\tilde{Q}\left(n\right)}{\left|{\theta}_{q}-{\tilde{\theta}}_{q}^{\left(n\right)}\right|}^{2}+\left|Q-{\tilde{Q}}^{\left(n\right)}\right|{\left(\mathsf{\Delta}{\theta}_{max}\right)}^{2}}{Q}},{\tilde{Q}}^{\left(n\right)}\le Q\\ \sqrt{\frac{{\sum}_{q=1}^{Q}{\left|{\theta}_{q}-{\tilde{\theta}}_{q}^{\left(n\right)}\right|}^{2}+{\sum}_{j=Q+1}^{{\tilde{Q}}^{\left(n\right)}}{\left|{\tilde{\theta}}_{j}^{\left(n\right)}-{\overline{\theta}}_{j}^{\left(n\right)}\right|}^{2}}{Q}},{\tilde{Q}}^{\left(n\right)}>Q\end{array}\right.$$

The averaged detection rate and the RMSE are summarized in Figure 4. It may be noted that when the SNR is greater or equal to 0 dB, ${r}_{\mathrm{dc}}\ge 0.25$ yields very satisfactory DoA performance. To put it into context, ${r}_{\mathrm{dc}}=0.25$ roughly equals transmitting 500 samples per second without data loss. The corresponding data volume of Nyquist rate sampling is 1600.

It is also interesting to examine how the RMSE obtained using simulation is compared against the theoretical Cram$\stackrel{\xb4}{e}$r–Rao bound. In Figure 5, it can be seen that for $\mathrm{SNR}>0$ dB, the RMSE is very close to the theoretically-computed CRB.

Next, we want to investigate the impact of varying the data packet loss rate for different values of random subsampling ratio $M/N$. The noise level of the acoustic signal is set at SNR = 0 dB. Two sets of plots are provided in Figure 6. On the top, the averaged probability of successful detection and the RMSE are plotted against $M/N$. The performance degradation due to data packet loss can be seen clearly. On the bottom, the averaged probability of successful detection and the RMSE are plotted against ${r}_{\mathrm{dc}}$, which, according to Equation (32), already takes into account the impact of ${r}_{\mathrm{loss}}$. Hence, the different curves on the bottom panel are merged into an identical one. This is because the random data loss is equivalent to the random subsampling of the raw data.

Our next goal is to investigate the DoA resolution of the proposed CSJSR-DoA approach when two acoustic sources become closer in terms of DoAs. Under the same simulation configuration, we fix one source at ${0}^{\circ}$ and vary the incident angle of the second source from ${2}^{\circ}$ to ${40}^{\circ}$ in ${2}^{\circ}$ increments. We perform 300 independent trials for each setting. The $\mathrm{SNR}=5\phantom{\rule{3.33333pt}{0ex}}\mathrm{dB}$, and the peaks of the $360\times 1$ sparse vector indicate the DoAs of two sources.

We report the averaged angle estimation error versus target DoA separation angles in Figure 7. Based on the discussion on Theorem 2, the steering vectors of two nearby sources are strongly correlated or have larger coherence. Therefore, the reconstruction of a correct sparse DoA indicative vector does not perform well. In simulations, the probability (false DoA probability) that only one source is found when two DoAs exist and the RMSE of each DoA estimation versus target DoA separation angles are shown in Figure 8. Note that when the separation angles are small, the averaged estimation errors, the RMSEs and the probability of false DoA estimation are rather large. However, when the targets are well separated ($\ge {20}^{\circ}$), the DoAs of both targets are accurately estimated (the averaged estimation error and the false DoA probability are close to zero, and the RMSEs decrease to a constant). This validates the theoretical angle separation result of $\frac{c}{Hdf}=\frac{343}{6\times 0.2\times 800}=20.{5}^{\circ}$.

Finally, we use simulation to compare the proposed CSJSR-DoA approach against two other CS-based DoA methods: COBE and CSA-DoA. The simulation conditions are identical to those used in Figure 4. While these three methods differ in how the measurements are made, the definition of ${r}_{\mathrm{dc}}$ would still be applicable to all of them; namely, the number of samples received at the FC versus the total amount of raw data samples at the sensor array. Specifically, the analog projection of the CSA-DoA approach is emulated by projecting the digitized signal. The results are summarized in Figure 9. It shows that the CSJSR-DoA method has favorable performance over others in a number of situations. Additional comparison of these methods using data collected from a prototype WSAN platform will be discussed next.

#### 5.3. Prototype WSAN Platform and Field Experiment

To further validate the capability of the CSJSR-DoA approach, a WSAN [47] was developed. Each sensor node (shown in Figure 10a) is equipped with an omni-directional microphone, a 16-bit ADC and a wireless transceiver operating at a rate of 31.25 KBps using the ZigBee protocol [48]. Random sub-sampled acoustic data samples are transmitted to the FC through the wireless transceiver to the FC (laptop). To guarantee time synchronization within the sensor array, the RBS time synchronization scheme is used. Figure 10b shows the experiment configuration, in which a laptop equipped with a wireless receiver was used as the FC. The sensor array is shown in the background close to the upper right corner.

In the first experiment, the task is to track the DoAs of a single moving acoustic source (digital acoustic signal played through a speaker). The distance between the source and the array varies from 20 m to 40 m, and the measured SNR in the array was about 6 dB. Each sensor node took samples during a period of 125 ms at a 2 Ksampling rate and transmitted them to the FC using the IEEE 802.15.4 protocol. In the FC, the DoA estimate was updated every second. The measured data loss ratio ${r}_{\mathrm{loss}}$ was $0.27$, and the corresponding ${r}_{\mathrm{dc}}$ of the received data in the FC was $0.23$. The result is shown in Figure 11. The red line is the ground truth, and the green circles are CSJSR-DoA estimates. The averaged probability of success detection is $90\%$, and the corresponding RMSE is $2.{76}^{\circ}$. This experiment provides an example of the practical use of the CSJSR-DoA approach.

In the second experiment, two stationary acoustic sources are used. They are placed at distances of $12.94$ m and $14.66$ m with (ground truth) DoA angles at $-6.{5}^{\circ}$ and $31.{0}^{\circ}$ respectively. Six sensors were incorporated into the array with an inter-sensor spacing equal to 0.2 m. Using data collected in this experiment, the performance of the CSJSR-DoA approach is compared against COBE, CSA-DoA and the baseline algorithm L1-SVD. The results are shown in Figure 12 and Table 2. The data compression ratio of all of these algorithms is fixed at ${r}_{\mathrm{dc}}=0.3$. To facilitate the comparison, the corresponding $L\times 1$ sparse DoA angle vector $\mathbf{s}$ is normalized and expressed in units of dB. From Figure 12, it is shown that the CSJSR-DoA approach yields sharper DoA estimates compared to the other three methods.

## 6. Conclusions

In this paper, a compressive sensing joint sparse representation approach (CSJSR-DoA) is presented for DoA estimation on a WSAN platform. By exploiting both frequency domain and spatial domain sparsity, the CSJSR-DoA approach provides a direct DoA angle estimation at the FC, while requiring almost no computation at power-constrained remote sensor nodes. We provided performance analysis in terms of DoA angle resolution and the Cram$\stackrel{\xb4}{e}$r–Rao bound of the estimates. We further conducted extensive simulation and built a prototype experimental WSAN platform to investigate the impacts of various parameters on the DoA performance. We also compared the performance of the proposed CSJSR-DoA approach against two other compressive sensing DoA estimation methods and showed that the CSJSR-DoA approach provides superior performance in both simulation runs and real-world experiments.

## Acknowledgments

This work was supported in part by the National Natural Science Foundation of China (No. NSFC61273079), in part by the National Natural Science Foundation of China Key Projects (U1509215), in part by the Strategic Priority Research Program of the Chinese Academy of Sciences (Grant No. XDA06020201) and in part by the Open Research Project of the State Key Laboratory of Industrial Control Technology, Zhejiang University (No. ICT1600208, No. ICT1600199, No. ICT1600213).

## Author Contributions

Kai Yu, Zhi Wang and Yu-Hen Hu have contributed to the scientific part of this work. All of the authors have contributed to the writing of this article.

## Conflicts of Interest

The authors declare no conflict of interest.

## Appendix A.

Let $\mathbf{C}=[\underset{\mathbf{C}\left[1\right]}{\underbrace{{\mathbf{c}}_{1}\cdots {\mathbf{c}}_{\mathrm{d}}}},\underset{\mathbf{C}\left[2\right]}{\underbrace{{\mathbf{c}}_{\mathrm{d}+1}\cdots {\mathbf{c}}_{2\mathrm{d}}}},\cdots \underset{\mathbf{C}\left[M\right]}{\underbrace{{\mathbf{c}}_{N-d+1}\cdots {\mathbf{c}}_{N}}}]$ be a concatenation of column block $\mathbf{C}\left[m\right]$ of size $L\times d$. Then, the block coherence is defined as:
where $\rho \left(\mathbf{X}\right)$ is the spectral norm of matrix $\mathbf{X}$. For simplification, we assume all of the H sensors have the same measurement matrix (${\mathsf{\Phi}}_{e}={\mathsf{\Phi}}_{e}^{1}=,\cdots ,={\mathsf{\Phi}}_{e}^{H}$); therefore:

$$\begin{array}{c}\hfill {\mu}_{B}=\underset{r,l\ne r}{max}\frac{1}{d}\rho \left(\mathbf{C}{\left[r\right]}^{H}\mathbf{C}\left[l\right]\right)\end{array}$$

$$\begin{array}{c}\hfill \begin{array}{cc}\mathsf{{\rm Y}}{\left[l\right]}^{H}\mathsf{{\rm Y}}\left[r\right]\hfill & =\frac{1}{MH}{\left({\mathbf{A}}_{L}\left(l\right)\otimes {\mathbf{\varphi}}_{e,l}\right)}^{H}\left({\mathbf{A}}_{L}\left(r\right)\otimes {\mathbf{\varphi}}_{e,r}\right)\hfill \\ & =\frac{{\mathbf{\varphi}}_{e,l}^{H}{\mathbf{\varphi}}_{e,r}}{MH}{\mathbf{A}}_{L}{\left(l\right)}^{H}{\mathbf{A}}_{L}\left(r\right)\hfill \end{array}\end{array}$$

Note the structure of ${\mathbf{\varphi}}_{e,r}$; the product ${\mathsf{\Phi}}_{e}\mathsf{\Psi}$ is equivalent to selecting some rows from the discrete Fourier transform (DFT) matrix

**Ψ**. Hence, their product ${\mathsf{\Phi}}_{e}\mathsf{\Psi}$ is a partial Fourier transform matrix (submatrix of a full DFT matrix). For the underlying random non-uniform sampling, the selection of these M rows is determined by $\mathbf{u}$. Applying the Welch Bound inequality [49], when $\mathbf{u}$ is randomly selected from $\{1,2,\cdots ,N\}$,
$$\begin{array}{c}\hfill |{\mathbf{\varphi}}_{e,l}^{H}{\mathbf{\varphi}}_{e,r}|\le \sqrt{\frac{(N-M)M}{(N-1)}}\end{array}$$

According to the Gershgorin circle theorem, ${\mu}_{B}$ is bound by:

$$\begin{array}{c}\hfill \begin{array}{cc}{\mu}_{B}\hfill & =\underset{l,r\ne l}{max}\frac{1}{L}\rho \left(\mathsf{{\rm Y}}{\left[l\right]}^{H}\mathsf{{\rm Y}}\left[r\right]\right)\hfill \\ & \le \underset{l,r\ne l}{max}\frac{1}{LH}\sqrt{\frac{(N-M)}{(N-1)M}}{\lambda}_{max}\left({\mathbf{A}}_{L}{\left(l\right)}^{H}{\mathbf{A}}_{L}\left(r\right)\right)\hfill \\ & =\sqrt{\frac{N-M}{(N-1)M}}\hfill \end{array}\end{array}$$

## Appendix B.

Assume that a uniform linear array consists of H sensors and that the distance between adjacent sensors is d. The orientationof a linear array is set to be orthogonal to the y-axis, and the array centroid is $[0,0]$. The positions of the h-th sensor satisfies ${u}_{h}-{u}_{h-1}=d(h=2,3,\cdots ,H)$, ${v}_{h}=0(h=1,2,\cdots ,H)$ and ${\tau}_{h,q}={u}_{h}sin\left({\theta}_{q}\right),{\theta}_{q}\in (-90,90]$. In this case, the corresponding steering vector is given by:
where $\chi ={e}^{\frac{-j{\omega}_{k}{u}_{1}\mathrm{sin}\left({\theta}_{\ell}\right)}{c}}$ is a constant. Recall Equation (A2); the block sub-matrix is given by:

$$\begin{array}{c}\hfill \begin{array}{cc}{\mathbf{a}}_{\ell}\left(k\right)\hfill & =\phantom{\rule{-0.166667em}{0ex}}{[{e}^{\frac{-j{\omega}_{k}{u}_{1}\mathrm{sin}\left({\theta}_{\ell}\right)}{c}},{e}^{\frac{-j{\omega}_{k}{u}_{2}\mathrm{sin}\left({\theta}_{\ell}\right)}{c}},\cdots ,{e}^{\frac{-j{u}_{H}{\omega}_{k}\mathrm{sin}\left({\theta}_{\ell}\right)}{c}}]}^{T}\hfill \\ & =\chi {[1,{e}^{\frac{-j{\omega}_{k}d\mathrm{sin}\left({\theta}_{\ell}\right)}{c}},\cdots ,{e}^{\frac{-j{u}_{H}{\omega}_{k}(H-1)d\mathrm{sin}\left({\theta}_{\ell}\right)}{c}}]}^{T}\hfill \end{array}\end{array}$$

$$\begin{array}{c}\hfill \mathsf{{\rm Y}}\left[k\right]={\mathbf{A}}_{L}\left(k\right)\otimes {\mathbf{\varphi}}_{e,k}\end{array}$$

The sparse representation matrix in Equation (8) that compactly expresses the spectrum of the H received signals is:
when ${\omega}_{k}d/c>\pi $ ($d>\lambda /2$) [37,50]; there are at least two different angles ${\theta}_{1}$ and ${\theta}_{2}$ that satisfy:
in which ${\theta}_{1}\ne {\theta}_{2}\in (-\pi /2,\pi /2]$. This implies that there will be multiple identical columns in ${\mathbf{A}}_{L}\left(k\right)$, and hence, the coherence of ${\mathbf{A}}_{L}\left(k\right)$ will be one. However, in CS theory, smaller coherence means better reconstruction performance. To confirm the uniqueness of ${\mathbf{A}}_{L}\left(k\right)$ (not the same columns in ${\mathbf{A}}_{L}\left(k\right)$) and, hence, a stable reconstruction of a sparse indicating vector, a small coherence is desirable.

$$\begin{array}{c}\hfill {\mathbf{A}}_{L}\left(k\right)=[{\mathbf{a}}_{1}\left(k\right),{\mathbf{a}}_{2}\left(k\right),\cdots ,{\mathbf{a}}_{L}\left(k\right)]\end{array}$$

$$\begin{array}{c}\hfill {e}^{\frac{-j{\omega}_{k}d\mathrm{sin}\left({\theta}_{1}\right)}{c}}={e}^{\frac{-j{\omega}_{k}d\mathrm{sin}\left({\theta}_{2}\right)}{c}}\end{array}$$

On the condition that $d\le 1/2\lambda $, the coherence of $\mathsf{{\rm Y}}\left[k\right]$ is given by:
where ${\mu}_{a}=\frac{1}{H}\underset{1\le {\ell}_{1}\ne {\ell}_{2}\le L}{\mathrm{max}}\left|\frac{1-exp\left(\frac{-jH{\omega}_{k}dp\left(\theta \right)}{c}\right)}{1-exp\left(\frac{-j{\omega}_{k}dp\left(\theta \right)}{c}\right)}\right|$ is determined by the redundant array manifold matrix, $\mathrm{p}\left(\theta \right)=sin\left({\theta}_{1}\right)-sin\left({\theta}_{2}\right)\approx cos\left({\theta}_{1}\right)\mathsf{\Delta}\theta $, $\mathsf{\Delta}\theta ={\theta}_{1}-{\theta}_{2}$.

$$\begin{array}{c}\hfill \begin{array}{c}\mu \left(\mathsf{{\rm Y}}\left[k\right]\right)\hfill \\ =\frac{1}{MH}\underset{1\le {\ell}_{1}\ne {\ell}_{2}\le L}{\mathrm{max}}\left|{\left({\mathbf{a}}_{{\ell}_{1}}\left(k\right)\otimes {\mathbf{\varphi}}_{e,k}\right)}^{H}\left({\mathbf{a}}_{{\ell}_{2}}\left(k\right)\otimes {\mathbf{\varphi}}_{e,k}\right)\right|\hfill \\ \le \frac{1}{H}\sqrt{\frac{N-M}{M(N-1)}}\underset{1\le {\ell}_{1}\ne {\ell}_{2}\le L}{\mathrm{max}}\left|{\mathbf{a}}_{{\ell}_{1}}^{H}\left(k\right){\mathbf{a}}_{{\ell}_{2}}\left(k\right)\right|\hfill \\ =\frac{1}{H}\sqrt{\frac{N-M}{M(N-1)}}\underset{1\le {\ell}_{1}\ne {\ell}_{2}\le L}{\mathrm{max}}\left|{\sum}_{h=0}^{H-1}exp\left(\frac{-jh{\omega}_{k}dp\left(\theta \right)}{c}\right)\right|\hfill \\ =\sqrt{\frac{N-M}{M(N-1)}}{\mu}_{a}\hfill \end{array}\end{array}$$

Note that $\underset{\mathsf{\Delta}\theta \to 0}{lim}{\mu}_{a}=1$, that is when the incident angles of two sources are close enough, the coherence may approach ${\mu}_{B}$, and hence, their DoAs become unresolvable. If the angle difference between two sources is larger than a certain value, a small coherence value can be guaranteed. Based on such an observation, an upper bound of ${\mu}_{a}$ can be established as ${\mu}_{b}=2/\left(H\right|1-exp\left(\frac{-j\omega d\mathrm{p}\left({\theta}_{1}\right)}{c}\right)\left|\right)$. In Figure B1, the coherence ${\mu}_{a}$ and its upper bound ${\mu}_{b}$ are plotted against $p\left({\theta}_{1}\right)$. It is observed that ${\mu}_{b}$ decreases sharply when $p\left({\theta}_{1}\right)$ is smaller than the first side lobe ($H\omega d\mathrm{p}\left({\theta}_{1}\right)/c=2\pi $). It is not difficult to verify that when:
${\mu}_{a}\le {\mu}_{b}\left(\mathsf{\Delta}\theta \right)=\frac{1}{H}\left|\frac{2}{1-{e}^{(-j2\pi /H)}}\right|=\frac{1}{Hsin(\pi /H)}$. This means a small coherence can be guaranteed when the angle difference of the two sources is larger than a certain threshold.

$$\begin{array}{c}\hfill \left|\mathsf{\Delta}\theta \right|\ge \frac{c}{Hfdcos\left({\theta}_{1}\right)}\ge \frac{c}{Hfd}\end{array}$$

## Appendix C.

Recall from Equation (29) that the details of the FIM matrix are:

$$\begin{array}{c}\hfill \mathbf{F}\left(\mathsf{\Lambda}\right)\phantom{\rule{-0.166667em}{0ex}}=\phantom{\rule{-0.166667em}{0ex}}\frac{2}{{\sigma}^{2}}\mathrm{Re}\left[\begin{array}{ccc}\left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \theta}\frac{\partial \mathbf{y}}{\partial \theta}\phantom{\rule{-0.166667em}{0ex}}\right)\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \theta}\frac{\partial \mathbf{y}}{\partial {\mathbf{r}}^{T}}\phantom{\rule{-0.166667em}{0ex}}\right)\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \theta}\frac{\partial \mathbf{y}}{\partial {\mathbf{f}}^{T}}\phantom{\rule{-0.166667em}{0ex}}\right)\\ \left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{r}}\frac{\partial \mathbf{y}}{\partial \theta}\phantom{\rule{-0.166667em}{0ex}}\right)\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{r}}\frac{\partial \mathbf{y}}{\partial {\mathbf{r}}^{T}}\phantom{\rule{-0.166667em}{0ex}}\right)\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{r}}\frac{\partial \mathbf{y}}{\partial {\mathbf{f}}^{T}}\phantom{\rule{-0.166667em}{0ex}}\right)\\ \left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{f}}\frac{\partial \mathbf{y}}{\partial \theta}\phantom{\rule{-0.166667em}{0ex}}\right)\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{f}}\frac{\partial \mathbf{y}}{\partial {\mathbf{r}}^{T}}\phantom{\rule{-0.166667em}{0ex}}\right)\phantom{\rule{-0.166667em}{0ex}}& \phantom{\rule{-0.166667em}{0ex}}\left(\phantom{\rule{-0.166667em}{0ex}}\frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{f}}\frac{\partial \mathbf{y}}{\partial {\mathbf{f}}^{T}}\phantom{\rule{-0.166667em}{0ex}}\right)\end{array}\right]\end{array}$$

Additionally, each term of the FIM is:
where $\mathbf{b}\left({k}_{i}\right)\phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}=\phantom{\rule{-0.166667em}{0ex}}\phantom{\rule{-0.166667em}{0ex}}-j2\pi {f}_{{k}_{i}}[{u}_{1}cos\left(\theta \right)-{v}_{1}sin\left(\theta \right),{u}_{2}cos\left(\theta \right)-{v}_{2}sin\left(\theta \right),\cdots ,{u}_{H}cos\left(\theta \right)-{v}_{H}sin\left(\theta \right)]$, $\frac{\partial \mathbf{y}}{\partial {f}_{i}}={r}_{0}\left({k}_{i}\right){\mathbf{a}}_{0}\left({k}_{i}\right)\otimes \left({\mathbf{v}}_{{k}_{i}}\odot {\mathbf{\varphi}}_{e,{k}_{i}}\right)$, ${\mathbf{v}}_{{k}_{i}}\phantom{\rule{-0.166667em}{0ex}}=-j2\pi (\mathbf{u}-\mathbf{1})/N$. We assume the coherence of ${\mathsf{\Phi}}_{h}\mathsf{\Psi}$ is very small and can be neglected (${\left({\mathbf{\varphi}}_{e,{k}_{l}}\right)}^{H}{\mathbf{\varphi}}_{e,{k}_{n}}\approx 0,l\ne n$). Based on this approximation, each component of the FIM is given by:
where $\alpha =\frac{4{\pi}^{2}}{{c}^{2}}{\sum}_{h=0}^{H-1}{\left({u}_{h}cos\left(\theta \right)-{v}_{h}sin\left(\theta \right)\right)}^{2}$, $\mathbf{e}={[{k}_{1}{r}_{0}\left({k}_{1}\right),\cdots ,{k}_{R}{r}_{0}\left({k}_{R}\right)]}^{T}$, $\eta =-j\pi (1+M)$, $\epsilon =(4{\pi}^{2}/{N}^{2}){\parallel \mathbf{u}-\mathbf{1}\parallel}_{2}^{2}$. Applying the well-known matrix inverse lemma, the CRB of CSJSR-DoA is:

$$\begin{array}{c}\hfill \begin{array}{cc}\frac{\partial \mathbf{y}}{\partial \theta}\hfill & ={\sum}_{i=1}^{R}{r}_{0}\left({k}_{i}\right)\left(\mathbf{b}\left({k}_{i}\right)\odot {\mathbf{a}}_{0}\left({k}_{i}\right)\right)\otimes {\mathbf{\varphi}}_{e,{k}_{i}},\hfill \\ \frac{\partial \mathbf{y}}{\partial {\mathbf{r}}_{0}^{T}}\hfill & =[{\mathbf{a}}_{0}\left({k}_{\mathbf{1}}\right)\otimes {\mathbf{\varphi}}_{e,{k}_{1}},{\mathbf{a}}_{0}\left({k}_{\mathbf{2}}\right)\otimes {\mathbf{\varphi}}_{e,{k}_{2}},\cdots ,{\mathbf{a}}_{\mathbf{0}}\left({k}_{R}\right)\otimes {\mathbf{\varphi}}_{e,{k}_{R}}]\hfill \\ \frac{\partial \mathbf{y}}{\partial {\mathbf{f}}^{T}}\hfill & =[\frac{\partial \mathbf{y}}{\partial {f}_{1}},\frac{\partial \mathbf{y}}{\partial {f}_{2}},\cdots ,\frac{\partial \mathbf{y}}{\partial {f}_{R}}],\hfill \end{array}\end{array}$$

$$\begin{array}{c}\hfill \begin{array}{cc}\frac{\partial {\mathbf{y}}^{H}}{\partial \theta}\frac{\partial \mathbf{y}}{\partial \theta}\hfill & ={M\alpha \left|\right|\mathbf{e}\left|\right|}_{2}^{2}\hfill \\ \frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{r}}\frac{\partial \mathbf{y}}{\partial {\mathbf{r}}^{T}}\hfill & =MJ\mathbf{I}\hfill \\ \frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{f}}\frac{\partial \mathbf{y}}{\partial {\mathbf{f}}^{T}}\phantom{\rule{1.0pt}{0ex}}\hfill & =\epsilon \mathrm{diag}({r}_{0}{\left({k}_{1}\right)}^{2},\cdots ,{r}_{0}{\left({k}_{R}\right)}^{2})\hfill \\ \frac{\partial {\mathbf{y}}^{H}}{\partial \theta}\frac{\partial \mathbf{y}}{\partial {\mathbf{r}}^{T}}\hfill & =\mathbf{0}\hfill \\ \frac{\partial {\mathbf{y}}^{H}}{\partial \theta}\frac{\partial \mathbf{y}}{\partial {\mathbf{f}}^{T}}\hfill & =\mathbf{0}\hfill \\ \frac{\partial {\mathbf{y}}^{H}}{\partial \mathbf{r}}\frac{\partial \mathbf{y}}{\partial {\mathbf{f}}^{T}}\hfill & =\eta \mathrm{diag}({r}_{0}\left({k}_{1}\right),\cdots ,{r}_{0}\left({k}_{R}\right))\hfill \end{array}\end{array}$$

$$\begin{array}{c}\hfill CRB\left(\theta \right)=\frac{{\sigma}^{2}}{{2M\alpha \parallel \mathbf{e}\parallel}_{2}^{2}}\end{array}$$

## References

- Ilyas, P.; Chen, H.; Tremoulis, G. Tracking of multiple moving speakers with multiple microphone arrays. IEEE Trans. Speech Audio Process.
**2004**, 12, 520–529. [Google Scholar] - Wang, Z.; Luo, J.; Zhang, X. A noval location-penalized maximum likelihood estimator for bearing-only target localization. IEEE Trans. Speech Audio Process.
**2012**, 60, 6166–6181. [Google Scholar] - Wang, Z.; Liao, J.; Cao, Q.; Qi, H.; Wang, Z. Achieving k-barrier Coverage in Hybrid Directional Sensor Networks. IEEE Transactions on Mob. Comput.
**2014**, 13, 1443–1455. [Google Scholar] [CrossRef] - Krim, H.; Viberg, M. Two decades of array signal processing research: the parametric approach. IEEE Signal Process. Mag.
**1996**, 13, 67–94. [Google Scholar] [CrossRef] - Winder, A. Sonar system technology. IEEE Trans. Sonics Ultrason.
**1975**, 22, 291–332. [Google Scholar] [CrossRef] - Chen, J.C.; Yao, K.; Hudson, R.E.; Tung, T.; Reed, C.; Chen, D. Source localization and tracking of a wideband source using a randomly distributed beam-forming sensor array. Int. J. High Perform. Comput. Appl.
**2002**, 16, 259–272. [Google Scholar] [CrossRef] - Chen, J.C.; Yip, L.; Elson, J.; Wang, H.; Maniezzo, D.; Hudson, R.E.; Yao, K.; Estrin, D. Coherent acoustic array processing and localization on wireless sensor networks. IEEE Process.
**2003**, 91, 1154–1162. [Google Scholar] [CrossRef] - Donoho, D.L. Compressive sensing. IEEE Trans. Inf. Theory
**2006**, 52, 1289–1306. [Google Scholar] [CrossRef] - Baraniuk, R. Compressive sensing. IEEE Signal Process. Mag.
**2007**, 2007, 118–120. [Google Scholar] [CrossRef] - Candés, E.J.; Romberg, J.; Tao, T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans. Inf. Theory
**2006**, 52, 489–509. [Google Scholar] - Gurbuz, A.C.; Mcclellan, J.H.; Cevher, V. A compressive beamforming method. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Las Vegas, NV, USA, 31 March–4 April 2008; pp. 2617–2620.
- Cevher, V.; Gurbuz, A.C.; Mcclellan, J.H.; Chellappa, R. Compressive wireless arrays for bearing estimation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Las Vegas, NV, USA, 31 March–4 April 2008; pp. 2497–2500.
- Gurbuz, A.C.; Cevher, V.; Mcclellan, J.H. Bearing estimation via spatial sparsity using compressive sensing. IEEE Trans. Aerosp. Electron. Syst.
**2012**, 48, 1358–1369. [Google Scholar] [CrossRef] - Wang, Y.; Leus, G.; Pandharipande, A. Direction estimation using compressive sampling array processing. In Proceedings of the 15th IEEE Workshop on Statistical Signal Process, Cardiff, UK, 31 August–3 September 2009; pp. 626–629.
- Li, B.; Zou, Y.; Zhu, Y. Direction estimation under compressive sensing framework: A review and experimental results. In Proceedings of the International Conference on Information and Automation (ICIA), Shenzhen, China, 6–8 June 2011; pp. 63–68.
- Zhang, J.; Bao, M.; Li, X. Wideband DOA estimation of frequency sparse sources with one receiver. In Proceedings of the International Conference on Mobile Ad-hoc and Sensor System (IEEE MASS), Hangzhou, China, 14–16 October 2013; pp. 609–613.
- Gu, J.; Zhu, W.; Swamy, M. Compressed sensing for DOA estimation with fewer receivers than sensors. In Proceedings of the International Symposium on Circuits and Systems (ISCAS), Rio de Janeiro, Brazil, 15–18 May 2011; pp. 1752–1755.
- Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Kelly, K.F.; Sun, T.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag.
**2007**, 24, 83–91. [Google Scholar] [CrossRef] - Luo, J.; Zhang, X.; Wang, Z. Direction-of-arrival estimation using sparse variable projection optimization. In Proceedings of the International Symposium on Circuits and Systems (ISCAS), Seoul, Korea, 20–23 May 2012; pp. 3106–3109.
- Luo, J.; Zhang, X.; Wang, Z. A new subband information fusion method for wideband DOA estimation using sparse signal representation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada, 26–31 May 2013; pp. 4016–4020.
- Carlin, M.; Rocca, P.; Oliveri, G.; Viani, F. Directions-of-Arrival Estimation Through Bayesian Compressive Sensing Strategies. IEEE Trans. Antennas Propag.
**2013**, 61, 3828–3838. [Google Scholar] [CrossRef] - Ji, S.; Xue, Y.; Carin, L. Bayesian Compressive Sensing. IEEE Trans. Signal Process.
**2008**, 56, 2346–2356. [Google Scholar] [CrossRef] - Yu, L.; Sun, H.; Barbot, J.P.; Zheng, G. Bayesian compressive sensing for cluster structured sparse signals. Signal Process.
**2012**, 92, 259–269. [Google Scholar] - Zhang, J.; Kossan, G.; Hedley, R.W.; Hudson, R.E.; Taylor, C.E.; Yao, K.; Bao, M. Fast 3D AML-based bird song estimation. Unmanned Syst.
**2014**, 2, 249–259. [Google Scholar] [CrossRef] - Duarte, M.F.; Davenport, M.A.; Wakin, M.B.; Baraniuk, R.G. Sparse signal detection from incoherent projections. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toulouse, France, 14–19 May 2006; pp. 305–308.
- Malioutov, D.; Cetin, M.; Willsky, A.S. A sparse signal reconstruction perspective for source localization with sensor arrays. IEEE Trans. Signal Process.
**2005**, 53, 3010–3022. [Google Scholar] [CrossRef] - Candés, E.J. The restricted isometry property and its implications for compressed sensing. Comptes Rendus Math.
**2008**, 346, 589–592. [Google Scholar] - Yu, K.; Yin, M.; Hu, Y.-H.; Wang, Z. CoSCoR framework for DoA estimation in wireless array sensor network. In Proceedings of the IEEE China Summit and International Conference on Signal and Information Processing (ChinaSIP), Beijing, China, 6–10 July 2013; pp. 215–219.
- Candés, E.J.; Romberg, J.K.; Tao, T. Stable signal recovery from incomplete and inaccurate measurements. Commun. Pure Appl. Math.
**2005**, 19, 410–412. [Google Scholar] - Viswanathan, R. Signal-to-noise ratio comparison of amplify-forward and direct link in wireless sensor networks. In Proceedings of the Communication Systems Software and Middleware, Bangalore, Indian, 7–12 January 2007; pp. 1–4.
- Alirezaei, G. Optimizing Power Allocation in Sensor Networks with Application in Target Classification; Shaker Verlag: Aachen, Germany, 2014. [Google Scholar]
- Wu, L.; Yu, K.; Cao, D.; Hu, Y.-H.; Wang, Z. Efficient sparse signal transmission over lossy link using compressive sensing. Sensors
**2015**, 15, 880–911. [Google Scholar] [CrossRef] [PubMed] - Lv, X.; Bi, G.; Chen, W. The group lasso for stable recovery of block-sparse signal representations. IEEE Trans. Signal Process.
**2011**, 59, 1371–1382. [Google Scholar] [CrossRef] - Jie, C.; Huo, X. Theoretical results on sparse representations of multiple-measurement vectors. IEEE Trans. Signal Process.
**2006**, 54, 4634–4643. [Google Scholar] - Cotter, S.F.; Rao, B.D.; Engan, K.; Kreutz-Delgado, K. Sparse solutions to linear inverse problems with multiple measurement vectors. IEEE Trans. Signal Process.
**2005**, 53, 2477–2488. [Google Scholar] [CrossRef] - Lobo, M.S.; Vandenberghe, L.; Boyd, S.; Lebret, H. Applications of second-order cone programming. Linear Algebra Its Appl.
**1998**, 284, 193–228. [Google Scholar] [CrossRef] - Shen, Q.; Liu, W.; Cui, W.; Wu, S.; Zhang, Y.D.; Amin, M.G. Low complexity direction of arrival estimation based on wideband co-prime arrays. IEEE Trans. Audio Speech Lang. Process.
**2015**, 23, 1445–1456. [Google Scholar] [CrossRef] - Wang, H.; Kaveh, M. Coherent signal-subspace processing for the detection and estimation of angles of arrival of multiple wide-band sources. IEEE Trans. Acoust. Speech Signal Process.
**1985**, 33, 823–831. [Google Scholar] [CrossRef] - Sun, L.; Liu, J.; Chen, J.; Ye, J. Efficient recovery of jointly sparse vectors. In Proceedings of the 2009 Conference on Advances in Neural Information Processing System, Vancouver, BC, Canada, 7–10 December 2009; pp. 1812–1820.
- Davenport, M.; Duarte, M.F.; Eldar, Y.C.; Kutyniok, G. Introduction to Compressed Sensing, Chapter in Compressed Sensing: Theory and Applications; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
- Eldar, Y.; Kuppinger, P.; Bolcskei, H. Block-sparse signals: Uncertainty relations and efficient recovery. IEEE Trans. Signal Process.
**2010**, 58, 3042–3054. [Google Scholar] [CrossRef] - Hansen, R. Array pattern control and synthesis. IEEE Proc.
**1992**, 80, 141–151. [Google Scholar] [CrossRef] - Sengijpta, S.K. Fundamentals of statistical signal processing: Estimation theory. Technometrics
**1995**, 37, 465–466. [Google Scholar] [CrossRef] - Sedumi. Available online: http://sedumi.ie.lehigh.edu/ (accessed on 9 May 2016).
- Likar, A.; Vidmar, T. A peak-search method based on spectrum convolution. J. Phys. D Appl. Phys.
**2003**, 36, 1903–1909. [Google Scholar] [CrossRef] - Guo, Q.; Liao, G. Fast music spectrum peak search via metropolis-hastings sampler. J. Electron.
**2005**, 22, 599–604. [Google Scholar] [CrossRef] - Yu, K.; Yin, M.; Du, T.; Wu, L.; Wang, Z. Demo: AWSAN: A realtime wireless sensor array platform. In Proceedings of the International Conference on Mobile Ad-hoc and Sensor System (IEEE MASS), Hangzhou, China, 14–16 October 2013; pp. 441–442.
- ZigBee Specifications. Available online: http://www.zigbee.org (accessed on 9 May 2016).
- Xia, P.; Zhou, S.; Giannakis, G.B. Achieving the welch bound with difference set. IEEE Trans. Inf. Theory
**2005**, 51, 1900–1907. [Google Scholar] [CrossRef] - Tang, Z.; Blacquiere, G.; Leus, G. Aliasing-free wideband beamforming using sparse signal representation. IEEE Trans. Signal Process.
**2011**, 59, 3464–3469. [Google Scholar] [CrossRef]

**Figure 4.**Performance comparison of the CSJSR-DoA approach under different data volumes: (

**left**) detection rate (

**right**) RMSE.

**Figure 5.**Comparison between the CSJSR-DoA result and the Cram$\stackrel{\xb4}{e}$r–Rao bound (CRB): (

**a**) CRB comparison with ${r}_{dc}=23\%$; (

**b**) CRB comparison with ${r}_{dc}=31\%$; (

**c**) CRB comparison with ${r}_{dc}=39\%$; (

**d**) CRB comparison with ${r}_{dc}=47\%$.

**Figure 6.**Comparison of different lossy transmissions: (

**a**) Detection rate versus $M/N$; (

**b**) RMSE versus $M/N$; (

**c**) Detection rate versus ${r}_{dc}$; (

**d**) RMSE versus ${r}_{dc}$ .

**Figure 9.**Performance comparison among CS-based methods under the same SNR ratio:(

**a**) Detection rate with SNR = 10 dB; (

**b**) RMSE with SNR = 10 dB; (

**c**) Detection rate with SNR = 5 dB;(

**d**) RMSE with SNR = 5 dB; (

**e**) Detection rate with SNR = 0 dB; (

**f**) RMSE with SNR = 0 dB.

Symbol | Explanation |
---|---|

H | number of sensors within an array |

Q | number of sources |

${R}_{q}$ | sparsity of the q-th source signal |

${\mathbf{s}}_{q}$ | the q-th source signal in the time domain |

${\mathbf{r}}_{q}$ | dominant sparse vector in the frequency domain with ${R}_{q}$-sparsity |

${\mathbf{\nu}}_{q}$ | less prominent components of ${\mathbf{s}}_{q}$ in the frequency domain |

${\tau}_{h,q}$ | time delay of the q-th source between the h-th sensor and the array centroid |

${x}_{h}\left(t\right)$ | received time domain signal in the h-th array |

Ψ | $N\times N$ inverse DFT matrix |

${\mathbf{x}}_{h}$ | received time domain signal at the h-th sensor, ${\mathbf{x}}_{h}={[{x}_{h}\left(1\right),\cdots ,{x}_{h}\left(N\right)]}^{T}$ |

${\tilde{\mathbf{d}}}_{h}$ | signal spectrum of ${\mathbf{x}}_{h}$, ${\tilde{\mathbf{d}}}_{h}={[{d}_{h}\left(0\right),{d}_{h}\left(1\right),\cdots ,{d}_{h}(N-1)]}^{T}$ |

$\mathbf{d}\left(k\right)$ | array data spectrum at the k-th frequency, $\mathbf{d}\left(k\right)={[{d}_{1}\left(k\right),{d}_{2}\left(k\right),\cdots ,{d}_{H}\left(k\right)]}^{T}$ |

${\mathbf{a}}_{\ell}\left(k\right)$ | steering vector of ${\theta}_{\ell}$ at the k-th frequency, |

${\mathbf{a}}_{\ell}\left(k\right)={[exp(-j{\omega}_{k}{\tau}_{1,q}),\cdots ,exp(-j{\omega}_{k}{\tau}_{H,q})]}^{T}$ | |

$\mathbf{r}\left(k\right)$ | source signal vector of L directions, $\mathbf{r}\left(k\right)={[{r}_{1}\left(k\right),{r}_{2}\left(k\right),\cdots ,{r}_{L}\left(k\right)]}^{T}$ |

${\mathbf{A}}_{L}\left(k\right)$ | steering matrix of L directions at the k-th frequency |

${\mathbf{A}}_{L}\left(k\right)=[{\mathbf{a}}_{1}\left(k\right),{\mathbf{a}}_{2}\left(k\right),\cdots ,{\mathbf{a}}_{L}\left(k\right)]$ | |

$\mathbf{D}$ | array data spectrum matrix |

$\mathbf{M}$ | permutation matrix |

$\overline{\mathbf{d}}$ | wideband array data spectrum, $\overline{\mathbf{d}}={[{\mathbf{d}}^{T}\left(0\right),{\mathbf{d}}^{T}\left(1\right),\cdots ,{\mathbf{d}}^{T}(N-1)]}^{T}$ |

$\tilde{\mathbf{d}}$ | array spectrum of H sensors, $\tilde{\mathbf{d}}={[{\tilde{\mathbf{d}}}_{1}^{T},{\tilde{\mathbf{d}}}_{2}^{T},\cdots ,{\tilde{\mathbf{d}}}_{H}^{T}]}^{T}$ |

$\mathbf{u}$ | sample intervals, $\mathbf{u}={[u\left(1\right),u\left(2\right),\cdots ,u\left(M\right)]}^{T}$ |

$\mathrm{r}(\xb7)$ | rounding operation |

Φ | random sub-sampling matrix |

${\mathsf{\Phi}}_{loss}^{h}$ | channel loss matrix |

${\mathbf{y}}_{h}$ | received measurement of the h-th sensor in the fusion center |

$\mathbf{y}$ | joint measurement vector of H sensors, $\mathbf{y}={[{\mathbf{y}}_{1}^{T},\cdots ,{\mathbf{y}}_{H}^{T}]}^{T}$ |

$\mathbf{w}$ | joint noise vector of N frequencies, $\mathbf{w}={[\mathbf{w}\left(0\right),\mathbf{w}\left(1\right),\cdots ,\mathbf{w}(N-1)]}^{T}$ |

Θ | joint measurement matrix of H sensors, $\mathsf{\Theta}=\mathrm{diag}({\mathsf{\Phi}}_{e}^{1}\mathsf{\Psi},{\mathsf{\Phi}}_{e}^{2}\mathsf{\Psi},\cdots ,{\mathsf{\Phi}}_{e}^{H}\mathsf{\Psi})$ |

$\mathrm{diag}(.)$ | block diagonal matrix operation |

Υ | joint sparse matrix |

$\mathbf{s}$ | direction indicative vector, $\mathbf{s}\left(\ell \right)={\sum}_{n=0}^{N-1}{\mathbf{r}}_{\ell}{\left(n\right)}^{2}$ |

$\mathbf{supp}(.)$ | nonzero index of a vector, $\mathrm{supp}\left(\mathbf{x}\right)=\left\{n\left|\mathbf{x}\right(n)>0\right\}$ |

$\tilde{\mathsf{{\rm Y}}}$ | pruned joint sparse matrix, $\tilde{\mathsf{{\rm Y}}}=\left[\mathsf{{\rm Y}}\left[{n}_{1}\right],\mathsf{{\rm Y}}\left[{n}_{2}\right],\cdots ,\mathsf{{\rm Y}}\left[{n}_{\left|\mathbf{T}\right|}\right]\right],r\in \mathbf{T}$ |

$\mathbf{F}\left(\mathsf{\Lambda}\right)$ | Fisher information matrix of parameter Λ |

$\mathbf{CRB}$ | Cram$\stackrel{\xb4}{e}$r–Rao bound of the CSJSR algorithm |

**Table 2.**Result of the prototype system experiment. CSA, compressive sensing array; COBE, compressive bearing estimation.

L1-SVD | CSJSR-DoA | CSA-DoA | COBE | |
---|---|---|---|---|

DoA (${}^{\circ}$) | [$-7$,29] | [$-7.5$,27.5] | [$-2.5$,30] | [$-9.5$,25] |

© 2016 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/).