Next Article in Journal
Time Series Feature Selection Method Based on Mutual Information
Previous Article in Journal
Cross-Modality Interaction-Based Traffic Accident Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Method for Sparse Representation of Complex Data Based on Overcomplete Basis, l1 Norm, and Neural MFNN-like Network

by
Nikolay V. Panokin
,
Artem V. Averin
,
Ivan A. Kostin
,
Alexander V. Karlovskiy
,
Daria I. Orelkina
and
Anton Yu. Nalivaiko
*
Center for Advanced Development of Autonomous Systems, Moscow Polytechnic University, 107023 Moscow, Russia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(5), 1959; https://doi.org/10.3390/app14051959
Submission received: 29 December 2023 / Revised: 20 February 2024 / Accepted: 21 February 2024 / Published: 27 February 2024

Abstract

:
The article presents the results of research into a method for representing complex data based on an overcomplete basis and l0/l1 norms. The proposed method is an extended modification of the neural-like MFNN (minimum fuel neural network) for the case of complex data. The influence of the choice of activation function on the performance of the method is analyzed. The results of the numerical simulation demonstrate the effectiveness of the proposed method for the case of sparse representation of complex data and can be used to determine the direction of arrival (DOA) for a uniform linear array (ULA).

1. Introduction

Classical methods for presenting data based on a selected basis include fast Fourier transform algorithms, cosine transform, wavelet transform, etc. The concept of an overcomplete basis is the use of a basis with a significantly larger number of components compared to the number of components in the analyzed data, which leads to the so-called SR (sparse representation). Unlike classical methods, the elements of the basis may not necessarily be functions orthogonal to each other, but the method of data representation is based on their decomposition into a number of optimal basic components, which are found from an overly complete basis dictionary using optimization algorithms such as, for example, MP (matching pursuit) and MOF (method of frame), using different norms [1,2,3,4,5,6,7,8,9,10]. The overcomplete basis or basis components are specified a priori or found using some optimization algorithms, such as correspondence search, frame methods, and basis search, depending on the current characteristics of the received data.
One of the important classes of norms [11] is the so-called lp norms for vector x. representing dimension I:
l p = i = 1 I x i p 1 / p   for   p 1
For the occasion p = 0:
l 0 = i = 1 I x i 0 ,       x i 0 = 0 , if   x i = 0 1 , otherwise
Thus, l0 is the number of non-zero elements in x. Minimizing the l0 norm leads to maximizing the number of zero values and to sparse data [12].
For the occasion p = 1:
l 1 = i = 1 I x i 1
Norm l1 represents the sum of the absolute values of elements in x and is widely used in algorithms such as LASSO (least absolute shrinkage and selection operator) in statistics BP (basis pursuit) in signal processing [12].
It should be noted that methods using SR can effectively separate coherent sources because data evaluation is based on an optimization method, such as l1 norm minimization, rather than on sub-space orthogonality, as in MUSIC (multiple signal classification) [13] or MVDR (minimum variance distortion less response) [14] and their use [15]. In addition, if the noise variance value used as the regularization parameter does not match the environmental noise variance, the evaluation performances of the corresponding methods are significantly degraded. An essential point is that when solving highly sparse linear optimization problems, the l1 norm is equivalent to the l0 norm, including for complex data [16,17].
The solutions given in [18,19] also allow the representation of data in an overcomplete dictionary basis. Along with the methods mentioned earlier, the decomposition of data into weighted dictionary components can be presented in the form of the so-called “fuel minimization” problem or MFNN [19], which is an optimization problem in a basis with norm l1 linear restrictions. Extending it to the complex plane ℂ, we obtain:
m a x y Ω   y T x m i n given   that   A x = b ,
where Ω = ω N ω i 1 , i 1 , 2 , , N N , b M represents the input data,
  • x = α γ 1 α γ 2 α γ N T N represents the basic components,
  • A = ϕ γ 1 , ϕ γ 2 , , ϕ γ N M × N represents the overcomplete dictionary matrix.
  • D = ( ϕ γ )   formed dictionary, where ϕ γ M   is the component of the basis with indexes γ Γ   and   Γ L 2
The input data S can be represented as:
b = i = 1 N α γ i ϕ γ i + r N + 1 ,
where r N + 1 M —the residual vector or error vector—is the result of the approximation of data b by the basis functions ϕ γ 1 , ϕ γ 2 , , ϕ γ N from the dictionary D, and α γ i - i- is the weight coefficient.
The solution to the optimization problem can be obtained by minimizing the norm of the loss function:
J = r N + 1 p = b A x p ,
where   p is the p-norm; in general, p = 2 (norm l2), and p = 1 (norm l1).
There are two approaches to solving this problem for p = 1. The first approach is described as an MP algorithm. This approach operates with dictionaries obtained from wavelet dictionaries, orthogonal dictionaries, and cosine functions. The second approach is called BP (basis pursuit) and was proposed by Chen and Donoho [16]. It operates with a simple dictionary of basis functions and allows the finding of an estimate of x using the BCSV (Box–Cox stochastic volatility) method by minimizing the l1 norm of x.
An implementation option for the Algorithm 1 BP—IRLS (iteratively reweighted least squares) [20] is given below:
Algorithm 1: BP
Start:   Basis   matrix   A m × N ,   data   vector   y m ,   initial   value   of   the   scales
ω 0 N   ( by   default :   ω 0 = 1 ,   1 ,   ,   1 ).
ε 0 =
for   k = 0 ,   1 ,   2 ,   do
           x k + 1 = arg min z N z ,   diag   ω k z   subject   to   A z = y ,
           ε k + 1 = min ε k ,   σ s x k + 1 l 1 N ,
           ω k + 1 i = 1 max x i k + 1 , ε k + 1   for   each   i N  
end for
return Sequence   x k k 1
where σ s x l 1 = inf x z 1 : z N   is   s sparse .
However, the algorithm is still not fast enough, and it does not allow the use of complex data.
In [21,22,23], Algorithm 2 SL0 (smoothed l0 norm) is presented, which implements a fairly fast method of finding a sparse solution to an uncertain system of linear equations, including for complex data. The SL0 algorithm implements an approximate solution based on the gradient method, with step-by-step projection of the result onto the space of feasible vectors:
Algorithm 2: SL0
  • Start: b, A, [σ1, …, σK], μ
  • Initialization:   x 1 = x 0
  • Cycle in k with a given number of repetitions:
   x = x k exp x k 2 / 2 σ k 2
   x k = x k 1 μ x A H A A H 1 A x k 1 b
where b is the initial coefficient of the scales, A is the selected excess basis matrix, σ and μ are coefficients determining the speed of convergence, and x is the solution vector.
However, despite the high speed of convergence, this algorithm is sensitive to the choice of initial values of the coefficients and does not guarantee convergence to the optimal solution.
Summarizing the above, we can highlight three characteristic points regarding this algorithm:
(1)
Minimizing the loss function leads to minimizing the number of non-zero components in vector x.
(2)
Based on the perturbation theory, the l1 norm is more robust compared to the l2 norm, which affects the quality of processing of non-stationary signals that contain sections of transient processes.
(3)
From a computational point of view, the problem belongs to the LP problem (linear programming, the class of optimization problems).
In [18], approaches to solving systems of linear equations based on neural networks are presented. In [19], a solution to the fuel minimization problem is presented in the form of a neural network MFNN, an example of which, from [19] for M = 8 and N = 3, is shown in Figure 1, which is described by a system of differential equations of the form:
d x d t = A T A x y b P Ω x + A T y , d y d t = A x + A T y P Ω x + A T y + b ,
z = x + A T   y ,   z     N  
where P Ω z = P Ω z 1 , P Ω z 2 , , P Ω z n T is the activation function (8):
P Ω z i = 1 2 z i + 1 z i 1 = 1 , i f   z i > 1 1 , i f   z i < 1 z i , otherwise  
Cosine functions are used as a dictionary in [19]. Methods for forming a dictionary, in particular, are described in [16], which also shows the relationship between increasing the redundancy of the dictionary and the quality of data approximation when defining linear programming problems. In [19], it is shown that in the field of real numbers ℝ, the neural network described by the system Equation (7) gives a solution that is globally stable according to Lyapunov and approaches the exact detection of problem (4).

2. Minimum Fuel Neural Network for Complex Data

The described approaches were used in [18,19] for certain classes of real data. To process complex data, it is necessary to refine the presented approaches in terms of forming a dictionary and the logic of the activation function (8).
Let us consider, as an example, an algorithm for forming a dictionary for data representing complex signals, after transferring to zero frequency, received by the ULA:
b k , m = b ˙ k e j 2 π d m 1 sin θ k ,
where k = 1, 2, …, K, K is the number of point external sources, θ k is the angular direction to the k-th source,   b ˙ k is the amplitude of the k-th source, d is the lattice element pitch normalized to the wavelength, and j = 1 .
Then, the elements of the overcomplete basis dictionary matrix can be represented as:
A m , i = b k e j 2 π d m 1 sin θ i ,
where θ i = 180 ° · i N 90 ° ), i = 1, 2, …, N is the dimension of the angular direction grid of the overcomplete basis dictionary, and m = 1, 2, …, M. (M is the number of receiving elements of the antenna array.)
There are two possible implementation options for the minimum fuel neural network MFNN, based on the structure presented in Figure 1.
The first option for implementing the neural system under consideration is the option with the presentation of complex input data b = Re (b) + j Jm(b), specified as a double-length real data vector b = [Re(b1), …, Re(bM), Jm(b1), …, Jm(bM)]; then, the vector of the resulting basis components (solutions) X = Re(X) + j Jm(X) similarly can be represented in the form:
X = Re x 1   Re x 2 Re x N   Im x 1   Im x 2 Im ( x N ) T
The algorithm for forming a dictionary in this case will be as follows:
(1)
A grid of angular directions to the source can be formed with a uniform step both in wave numbers φ i = π N i 1 π 2  and in the angle sin φ i , where i = 1, 2, …, N is the total number of angular directions. In the first case, it turns out to be denser in the vicinity of the zero-angle value.
(2)
For an m-channel receiving system, two parts of the dictionary are formed in the following ways:
A 1 = sin ( π m 1 sin φ i ) , A 2 = cos ( π m 1 sin φ i )
(3)
The formed parts are combined into a common dictionary, according to the rule:
A = A 1   A 2 A 2               A 1 ,
In this case, all operations on data in the neural network will be carried out with real values.
The second option for implementing a neural network is the option of replacing real variables with complex ones and performing all operations with complex values.
The form of the system of Equation (7) in both the first and second cases will remain unchanged. The choice of one of the options will be determined only by the convenience of the specific hardware and software implementations. At the same time, since the absolute values of the real and complex variables are defined differently, the question arises of choosing the optimal form of the activation function PΩ for the case of using complex data.

3. Activation Function to Implement the Algorithm in Real Form

If we do not make changes to activation function (8), then for the implementation option with the representation of a vector of complex data by a vector of real data, twice the length of (11), we will obtain it in the form of a simple limiter:
  P Ω z ˙ = P Ω z ˙ 1   P Ω z ˙ 2 P Ω z ˙ N   P Ω z ˙ N + 1   P Ω z ˙ N + 2   P Ω z ˙ 2 N T
where z ˙ i = Re z i z ˙ i + N = Im z i ,  
  P Ω z ˙ i = 1 ,     if   z ˙ i > 1 1 ,     if   z ˙ i < 1   z ˙ i ,     otherwise     , i = 1 ,   2 , ,   2 N      
Another option for the activation function is to represent it in the form:
P Ω z ˙ i = z ˙ i w i , if   w ˙ i > 1 z ˙ i , if   w ˙ i 1   ,   P Ω z ˙ i + N = z ˙ i + N w ˙ i ,     if   w ˙ i > 1 z ˙ i + N ,     if   w ˙ i 1
where w ˙ i = z ˙ i 2 + z ˙ i + N 2 ,   i = 1 , 2 , , N , and w ˙ i + N = w ˙ i ,   i = 1 , 2 , , N .
One more possible option for the activation function is the following:
P Ω z ˙ i , P Ω z ˙ i + N = z ˙ i z ˙ i , z ˙ i + N z ˙ i , if   z ˙ i z ˙ i + N   Λ   z ˙ i > 1 ,   i = 1 , 2 , , N z ˙ i z ˙ i + N , z ˙ i + N z ˙ i + N , if   z ˙ i + N z ˙ i   Λ   z ˙ i + N > 1 ,   i = 1 , 2 , , N z ˙ i , z ˙ i + N , otherwise  

4. Activation Function to Implement the Algorithm in Complex Form

Similarly, for the implementation of a neural network with complex variables z i = Re z i + j Im ( z i ) ,   i = 1 , 2 , , N ,   j = 1 , we obtain the following options for the activation function:
Independent activation function for real and imaginary parts:
P Ω z i = P Ω Re z i + j P Ω Im ( z i ) ,
where P Ω Re z i = 1 , if   Re z i > 1 1 , if   Re z i < 1 Re z i , otherwise   , P Ω Im z i = 1 , if   Im z i > 1 1 , if   Im z i < 1 Im z i , otherwise  
Activation function that implements normalization to the module of a complex value:
P Ω z i = z i z i , if   z i > 1 z i , if   z i 1 ,
where z i = Re z i 2 + Im z i 2 ,   i = 1 , 2 , , N .
Activation function that implements normalization to the maximum of the values of the real and imaginary parts:
P Ω z i = z i Re z i , if     Re z i > 1 , Im z i Re z i z i Im z i , if     Im z i > 1 , Re z i Im z i z i , if     Im z i 1 , Re z i 1

5. Numerical Modeling

To analyze the efficiency of the neural network (Figure 1) with the specified activation functions, we considered the model problem of determining the DOA to estimate the position of point external sources in space, using complex signals received by the elements of a ULA. Typical examples of the DOA estimation method are the FFT-based beamforming algorithm and the MUSIC algorithm using the orthogonality of the noise part eigenvectors and the steering vectors.
Numerical modeling was carried out in MATLAB ver. R2020b using a computer on a processor Intel(R) Core (TM) i7-6700, with a frequency of 3.40 GHz; one step of the normalized calculation time was 0.00335 s. Values from (10) were selected as elements of the excess basis matrix with N = 256, M = 16, and d = 0.5. The number of point external sources varied from 1 to 16, and the amplitude and phase varied in accordance with the simulation conditions.

5.1. The Case of a Single External Source with a Varying Initial Phase

Below are the results of the numerical simulation of the operation of a neural network with various activation functions (18)–(20) under the influence of a single point source with an amplitude of b ˙ 1 = 1, an angle of arrival of θ = −22.5°, and different phases of φ = 0°, 72°, and −18.9°; on the left, vertically, are the values of the phases used in modeling the sources 0°, 72°, and −18.9°. For example, Figure 2d shows the plot obtained for the activation function (18) and the source phase φ = 72°.
In the graphs of Figure 2a–i, the angular directions in degrees are laid along the horizontal axis (in the grid of angular directions of an excessively complete basic dictionary); along the vertical axis are the values of the modules of the resulting basic components |X|, obtained as a result of processing using a neural network with the selected activation function. They correspond with the angular directions of the grid, the angles of the basic functions (blue lines), and the modules of the values of the response |F|, formed in the same angular grid using the standard FFT method (black lines). The position of the point source, considering its amplitude b ˙ 1 ,   is represented in red dots.
Figure 2. Single source b ˙ 1 =  1, θ 1 = −22.5°, with values of |X| and |F|, activation functions (18)—subfigures (a,d,g), (19)—subfigures (b,e,h), (20)—subfigure (c,f,i).
Figure 2. Single source b ˙ 1 =  1, θ 1 = −22.5°, with values of |X| and |F|, activation functions (18)—subfigures (a,d,g), (19)—subfigures (b,e,h), (20)—subfigure (c,f,i).
Applsci 14 01959 g002
From the results obtained, it is clear that when the phase value φ = 0°, that is, provided that the phase of the source and the phases of the basis functions (10) coincide, the use of all activation functions leads to a solution with only one non-zero basis coefficient, which corresponds to the amplitude and direction of the source, as shown in Figure 2a–c. However, if a discrepancy occurs between the phases of the source and the bases (18) and (20), it leads to errors in determining the direction and amplitude. In this case, the number of non-zero values of the basic coefficients significantly exceeds the number of sources, and among them, there are no corresponding directions to the source. Table 1 shows the number of basic coefficients with a value normalized to a maximum value max i x i   exceeding 10 3 .
In Figure 3a–i, the horizontal axis shows the time normalized to the duration of one internal calculation cycle t along the vertical axis (scale on the left), which represents the relative change in the coefficients of the basis functions in one normalized time step, ΔX(t) = max i x i t 1 x i t / max i x i t 1 ,   t = 1 , 2 , , T (blue), and the current error in representing the input data by basis components, which is defined as the change in the time of the loss function (6), as shown by the scale on the right (red).
From the results obtained, it follows that the time to establish the values of the coefficients of the basis functions when using the activation functions (18) and (19) weakly depends on the value of the source phase.
For the activation functions (18) and (20), it is clear that for some phase values, the process could not converge to a global minimum according to the l0 norm, that is, minimize the number of non-zero basis coefficients, although the loss function (6), for all options, is less than 10 16 and practically does not change over time (except for the option Figure 3i. However, even in this case, with an increase in time by 10 times, the loss function stabilizes at a level less than 10 14 without changing |X|), which indicates that the minimum has been reached, according to norm l1. Thus, like the case of using activation functions (18) and (20) and phase φ = 72°, it is clear that the value of the loss function (3) turns out to be very small, but the parameters of the non-zero basis coefficients are very different from the parameters of the external source. In addition, the convergence process slows down. The activation function (19) made it possible to achieve a global minimum according to the l0 norm, with an increase in time, in the cases considered, by 2–3 times.
Thus, all three options of the activation functions (18)–(20) allow, with a zero initial phase of the signal, the accurate representation of the external source signal with basis functions. However, if the initial phase is not zero, then options (18) and (20) may give an erroneous result.

5.2. The Case of Two Sources with Different Initial Signal Phases

Next, Figure 4 and Figure 5 (similar to Figure 2 and Figure 3) show the results of a numerical simulation of the operation of a neural network with various activation functions (18)–(20) under the influence of two point sources with an amplitude of b ˙ 1 = b ˙ 2 = 1, angles of arrival of θ 1 = −22.5° and θ 2 = −18.98°, and two phase options for each of the two sources: φ 1 = 0° and φ 2 = 0°, and φ 1 = 72° and φ 2 = −72°.
In Figure 4 and Figure 5, the activation functions used from (18) to (20) are indicated vertically on the left, and at the top, horizontally, are the phase values used in modeling sources ( φ 1 ,   φ 2 ) = (0°, 0°) and ( φ 1 ,   φ 2 ) = (72°, −72°). For example, Figure 4c shows the plot obtained for the activation function (19) and the phases of sources φ 1 = 0° and φ 2 = 0°.
In Figure 4a–f, the angular directions in degrees are laid along the horizontal axis (in the grid of angular directions of an excessively complete basic dictionary); along the vertical axis are the values of the modules of the resulting basic components |X|, obtained as a result of processing using a neural network with the selected activation function. They correspond to the angular directions of the grid, the angles of the basic functions (blue), and the modules of the values of the radiation pattern |F|, formed in the same angular grid using the standard FFT method (black). The position of point sources, taking into account amplitudes b ˙ 1   and  b ˙ 2 ,   is represented in red.
Figure 4. Two sources b ˙ 1 = b ˙ 2 =  1, θ 1  = −22.5°, and θ 2 = −18.98°, with values of |X| and |F|, activation functions (18)—subfigures (a,b), (19)—subfigures (c,d), (20)—subfigures (e,f).
Figure 4. Two sources b ˙ 1 = b ˙ 2 =  1, θ 1  = −22.5°, and θ 2 = −18.98°, with values of |X| and |F|, activation functions (18)—subfigures (a,b), (19)—subfigures (c,d), (20)—subfigures (e,f).
Applsci 14 01959 g004
From the results obtained in all cases, the radiation pattern formed using FFT does not allow resolving sources at an angular distance between them equal to 3.52°. With a phase value of φ = (0°, 0°), all activation functions lead to an optimal solution with two non-zero basis coefficients, which correspond with the amplitude and direction of the sources (Figure 4a,c,e). However, like the case of one source, when a discrepancy occurs between the phases of the sources and the basis functions, the use of activation functions (18) and (20) leads to errors in determining the directions and amplitudes of the sources, as shown in Figure 4b,f. In this case, the number of non-zero values of the basic coefficients increases, and among them, there are no corresponding ones in the direction of the source. Table 2 shows the number of basic coefficients with a value normalized to the maximum value max i x i   exceeding 10 3 .
In Figure 5a–f, the horizontal axis shows the time normalized to the duration of one internal calculation cycle t, and the vertical axis (scale on the left) shows the relative change in the coefficients of the basis functions for one normalized time step, Δ X t = max i x i t 1 x i t / max i x i t 1 ,   t = 1 , 2 , , T (blue), and the current error in representing the input data by basis components, which was defined as the change in time of the loss function (6), as shown in the scale on the right (red).
Figure 5. Two sources b ˙ 1 = b ˙ 2 =  1, θ 1 = −22.5°, and θ 2 = −18.98°, with values of Δ|X(t)| and J(t), activation functions (18)—subfigures (a,b), (19)—subfigures (c,d), (20)—subfigures (e,f).
Figure 5. Two sources b ˙ 1 = b ˙ 2 =  1, θ 1 = −22.5°, and θ 2 = −18.98°, with values of Δ|X(t)| and J(t), activation functions (18)—subfigures (a,b), (19)—subfigures (c,d), (20)—subfigures (e,f).
Applsci 14 01959 g005
From the results obtained, it follows that the time to establish the values of the coefficients of the basis functions when using activation functions (18) and (19) weakly depends on the value of the source phase.
For activation functions (18) and (20), it is clear that with phase values ( φ 1 ,   φ 2 ) = (72°, −72°), the process was unable to converge to a global minimum according to the l0 norm, that is, to minimize the number of non-zero basic coefficients, although the loss function (6), for all options, was less than 10 9 and practically did not change over time, which indicates that the minimum in the l1 norm has been reached. Thus, like the case of using activation functions (18) and (20) and phase φ = 72°, the value of the loss function (6) turned out to be very small, but the non-zero basis coefficients were very different from the source. In addition, the convergence process slowed down. Activation function (19), unlike the previous ones, made it possible to achieve a global minimum according to the l0 norm, with an increase in time, in the cases considered, by 2–3 times.
An example of a change in the number of non-zero basis coefficients over time is presented in Figure 6. The number of basis coefficients is plotted along the vertical axis, and normalized time is plotted along the horizontal axis. The number of base coefficients is shown in different colors, the values of the modules of which, normalized to the maximum value U = X t / max i X t , exceed the following specified levels: [ 10 14 ,   10 9 , 10 4 ,   10 3 ,   10 2 , 10 1 ]. (The legend is shown in the figure.)
It can be seen that all the basic coefficients quickly tend to become zero, and, in the end, in addition to the two basic coefficients corresponding to the source data, there are only three more coefficients with a level exceeding 10 9 .
Thus, like the case with a single source, all three versions of the activation functions (18)–(20) allow, at zero initial phases, the accurate representation of the data with basis functions; however, if the initial phases of the sources are not zero, then options (18) and (20) may give a result that does not correspond to their real position and amplitude. The variant of activation function (19) for the considered case, as well as for a single source, demonstrated resistance to phase changes.

5.3. A Large Number of Sources

The following is the result of an analysis of the influence of the number of sources, which varied from 7 to 16, on the results of the MFNN neural network for complex data with the activation function (19), which showed the best result.
In Figure 7a,c,e, on the graphs, the horizontal axis shows the angular direction in degrees, and the vertical axis shows the values of the modules of the coefficients of the basis components X obtained at different times. They were obtained as a result of processing using a neural network with activation function (19), corresponding to the angular directions of the grid of angles of basis functions, as shown along the horizontal axis (blue lines), and the modules of the radiation pattern formed using FFT (black lines). The position of point sources, considering amplitudes b ˙ i   and angles  θ i , is represented by red dots.
In Figure 7a–f, the results are shown for the case of seven external sources, with levels b ˙ 1 7 = [1.0, 1.0, 0.5, 2.0, 0.7, 1.0, 1.0] and angular directions θ 1 7   = [−42.19°, −33.75°, −22.50°, −18.98°, −11.25°, −5.63°, −2.11°] for the next three normalized time intervals T = [0–1000 ((a), (b)), 500–5000 ((c), (d)), 140,500–148,000 ((e), (f))], which show how the X coefficients and loss functions change (6). It can be seen that the X coefficients group quickly enough around the directions of the sources, and the values of the coefficients outside these zones drop below 10 8 (Figure 7a). If attention is paid to the graphs (Figure 7b,d,e), you can see that loss function (6) changes unevenly in jumps, while the changes in the basic coefficients remain generally monotonic. Subsequently, the process of changing the coefficients X slows down significantly, but their grouping continues in such a way that each direction to the source corresponds to its own area of significant coefficients, separated by zero coefficients from the rest (Figure 7c,e). By the time T = 148,000, there are only 16 significant basis coefficients of X, and only 10 of them have a value greater than 10 1 . It turns out that in fact, the sums of the coefficients in these areas are equal to the levels of the corresponding sources, and the weighted average values of the angular directions in these areas are very close to the directions of the sources. The directions and levels determined in this way are displayed in Figure 7a,c,e with red dotted lines, which show the coincidence with the directions and levels of external sources. This feature makes it possible to significantly reduce the time required to find a solution to the system of differential Equation (14). Almost upon completion of the interval T = [0–1000] (Figure 7a), a result similar to that obtained in the interval T = [140,500–148,000] (Figure 7e) is achieved.
The graphs shown in Figure 8a,b are similar to those shown in Figure 7a, except the number of external sources for Figure 8a was 15: b ˙ 1 15 = [1.0, 1.0, 0.5, 2.0, 0.7, 1.0, 1.0, 1.0, 0.5, 1.0, 1.0, 0.7, 0.8, 0.5, 1.0], and θ 1 15   = [−42.19°, −33.75°, −22.50°, −18.98°, −11.25°, −5.63°, −2.11°, 5.63°, 9.14°, 16.88°, 19.69°, 28.13°, 31.64°, 39.36°, 47.81°]; for Figure 8b—16, another external source was added: b ˙ 16 = 1.0, θ 16   = 53.44°.
When exposed to 15 external sources (Figure 8a), the basis coefficients are concentrated around the directions to the sources, as noted above. Using the method of summing the coefficients in these areas and determining the angular directions from the weighted average values, similar to the previous one, it is possible to separate all external sources, except for the pair with coordinates at 16.88° and 19.69°, and determine their angular directions, in contrast to the case of using the FFT-based diagram formation. At the same time, when the number of external sources reaches 16, that is, it becomes equal to the number of ULA channels (Figure 8b), ten external sources cannot be resolved, although the tendency of grouping around the directions to the sources remains.

6. Conclusions

The article discusses the proposed implementation of a sparse representation of complex data based on an overcomplete basis, l0/l1 norms, and a neural-like MFNN [19,20,21] for the case of complex data. It is shown that the modification of the initializing function of a neural network proposed in the article allows one to effectively solve the problem of representing complex data by elements of a complex basis. The influence of the choice of the type of initializing function on the characteristics of the neural network is analyzed. It is shown that the initializing function (19) is optimal if the phases of these sources and the selected complex overcomplete basis are different. In relation to the problem of determining the direction to the sources of signals received by a linear antenna array, a significant increase in the resolution in angular direction was demonstrated in comparison with the classical processing method based on the FFT algorithm. Unlike SL0 [21], this is guaranteed to converge with the optimal solution but shows a slower rate of convergence.
In subsequent studies, there are plans to conduct a detailed analysis of methods for accelerating the convergence of the method under consideration, as well as an experimental study of its characteristics for processing multidimensional data.

Author Contributions

Methodology, N.V.P.; software, N.V.P., A.V.A. and I.A.K.; writing—original draft preparation, N.V.P. and A.Y.N.; writing—review and editing, N.V.P., A.V.K. and D.I.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Higher Education of the Russian Federation within the framework of state assignment No. FZRR-2023-0008.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Schmidt, R.O. Multiple Emitter Location and Signal Parameter Estimation. IEEE Trans. Antennas Propag. 1986, 34, 276–280. [Google Scholar] [CrossRef]
  2. Rao, B.D.; Hari, K.V.S. Performance analysis of root-MUSIC. IEEE Trans. Acoust. Speech Signal Process. 1989, 37, 1939–1949. [Google Scholar] [CrossRef]
  3. Zala, C.A.; Barrodale, I.; Kennedy, J.S. High-resolution signal and noise field estimation using the L1 (least absolute values) norm. IEEE J. Ocean. Eng. 1987, 12, 253–264. [Google Scholar] [CrossRef]
  4. Bandler, J.W.; Kellerman, W.; Madsen, K. A nonlinear L1 optimization algorithm for design, modeling, and diagnosis of networks. IEEE Trans. Circuits Syst. 1987, 34, 174–181. [Google Scholar] [CrossRef]
  5. Abdelmalek, N.N. Solutions of minimum time problem and minimum fuel problem for discrete linear admissible control systems. Int. J. Syst. Sci. 1978, 8, 849–859. [Google Scholar] [CrossRef]
  6. Levy, S.; Walker, C.; Ulrych, T.J.; Fullagar, P.K. A linear programming approach to the estimation of the power spectra of harmonic processes. IEEE Trans. Acoust. Speech Signal Process. 1992, 30, 675–679. [Google Scholar] [CrossRef]
  7. Stanković, L.; Sejdić, E.; Stanković, S.; Daković, M.; Orović, I. A Tutorial on Sparse Signal Reconstruction and Its Applications in Signal Processing. Circuits Syst. Signal Process. 2019, 38, 1206–1263. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Xiao, S.; Huang, D.; Sun, D.; Liu, L.; Cui, H. L0-norm penalized shrinkage linear and widely linear LMS algorithms for sparse system identification. IET Signal Process. 2017, 11, 86–94. [Google Scholar] [CrossRef]
  9. Ishii, Y.; Koide, S.; Hayakawa, K. L0-norm Constrained Autoencoders for Unsupervised Outlier Detection. In Advances in Knowledge Discovery and Data Mining. PAKDD 2020; Lauw, H., Wong, R.W., Ntoulas, A., Lim, E.P., Ng, S.K., Pan, S., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; Volume 12085. [Google Scholar] [CrossRef]
  10. Marquardt, D.W.; Snee, R.D. Ridge regression in practice. Am. Stat. 1975, 29, 3–20. [Google Scholar] [CrossRef]
  11. Rajko, R. Studies on the adaptability of different Borgen norms applied in selfmodeling curve resolution (SMCR) method. J. Chemom. 2009, 23, 265–274. [Google Scholar] [CrossRef]
  12. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  13. Schmidt, R.O. A Signal Subspace Approach to Multiple Emitter Location and Spectral Estimation. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 1981. [Google Scholar]
  14. Capon, J. High resolution frequency-wavenumber spectrum analysis. Proc. IEEE 1969, 57, 1408–1418. [Google Scholar] [CrossRef]
  15. Sun, K.; Liu, Y.; Meng, H.; Wang, X. Adaptive Sparse Representation for Source Localization with Gain/Phase Errors. Sensors 2011, 11, 4780–4793. [Google Scholar] [CrossRef] [PubMed]
  16. Donoho, D.L.; Huo, X. Uncertainty Principles and Ideal Atomic Decomposition. IEEE Trans. Inf. Theory 2001, 47, 2845–2862. [Google Scholar] [CrossRef]
  17. Malioutov, D.M.; Cetin, M.; Willsky, A.S. Optimal sparse representations in general overcomplete bases. In Proceedings of the 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing, Montreal, QC, Canada, 17–21 May 2004. [Google Scholar] [CrossRef]
  18. Cichocki, A.; Unbehauen, R. Neural Networks for Solving Systems of Linear Equations and Related Problems. IEEE Trans. Circuits Syst. I Fund. Theory Appl. 1992, 39, 124–138. [Google Scholar] [CrossRef]
  19. Wang, Z.S.; Cheung, J.Y.; Xia, Y.S.; Chen, J.D.Z. Minimum fuel neural networks and their applications to overcomplete signal representations. IEEE Trans. Circuits Syst. I Fund. Theory Appl. 2000, 47, 1146–1159. [Google Scholar] [CrossRef]
  20. Kümmerle, C.; Verdun, C.M.; Stöger, D. Iteratively Reweighted Least Squares for Basis Pursuit with Global Linear Convergence Rate. In Proceedings of the Advances in Neural Information Processing Systems 34 (NeurIPS 2021), Online, 6–14 December 2021; Volume 34, pp. 2873–2886. [Google Scholar]
  21. Mohimani, G.H.; Babaie-Zadeh, M.; Jutten, C. Complex-valued sparse representation based on smoothed l0 norm. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Las Vegas, NV, USA, 31 March–4 April 2008; pp. 3881–3884. [Google Scholar] [CrossRef]
  22. Mohimani, H.; Babaie-Zadeh, M.; Jutten, C. A fast approach for overcomplete sparse decomposition based on smoothed l0 norm. IEE Trans. Signal Process. 2009, 57, 289–301. [Google Scholar] [CrossRef]
  23. Wang, L.; Yin, X.; Yue, H.; Xiang, J. A regularized weighted smoothed L0 norm minimization method for underdetermined blind source separation. Sensors 2018, 12, 4260. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Neural network architecture for M = 8 and N = 3 [19].
Figure 1. Neural network architecture for M = 8 and N = 3 [19].
Applsci 14 01959 g001
Figure 3. Single external source b ˙ 1 =  1, θ 1 = −22.5°, with values of Δ|X(t)| and J(t), activation functions (18)—subfigures (a,d,g), (19)—subfigures (b,e,h), (20)—subfigures (c,f,i).
Figure 3. Single external source b ˙ 1 =  1, θ 1 = −22.5°, with values of Δ|X(t)| and J(t), activation functions (18)—subfigures (a,d,g), (19)—subfigures (b,e,h), (20)—subfigures (c,f,i).
Applsci 14 01959 g003
Figure 6. Two sources b ˙ 1 = b ˙ 2 =  1, θ 1 = −22.5°, and θ 2 = −18.98°; number of basis coefficients for U > [ 10 14 ,   10 9 , 10 4 ,   10 3 ,   10 2 , 10 1 ].
Figure 6. Two sources b ˙ 1 = b ˙ 2 =  1, θ 1 = −22.5°, and θ 2 = −18.98°; number of basis coefficients for U > [ 10 14 ,   10 9 , 10 4 ,   10 3 ,   10 2 , 10 1 ].
Applsci 14 01959 g006
Figure 7. Seven sources b ˙ 1 7 = [1.0, 1.0, 0.5, 2.0, 0.7, 1.0, 1.0], and θ 1 7   = [−42.19°, −33.75°, −22.50°, −18.98°, −11.25°, −5.63°, −2.11°], with values of |X|, Δ|X(t)|, and J(t), the interval T = [0–1000] subfigures (a,b), T = [1300–5000] subfigures (c,d), T = [140,500–148,000] subfigure (e,f).
Figure 7. Seven sources b ˙ 1 7 = [1.0, 1.0, 0.5, 2.0, 0.7, 1.0, 1.0], and θ 1 7   = [−42.19°, −33.75°, −22.50°, −18.98°, −11.25°, −5.63°, −2.11°], with values of |X|, Δ|X(t)|, and J(t), the interval T = [0–1000] subfigures (a,b), T = [1300–5000] subfigures (c,d), T = [140,500–148,000] subfigure (e,f).
Applsci 14 01959 g007
Figure 8. Fifteen (a) and sixteen (b) external sources, with values of |X|.
Figure 8. Fifteen (a) and sixteen (b) external sources, with values of |X|.
Applsci 14 01959 g008
Table 1. Number of basis coefficients with a level exceeding 10 3 ,  single source.
Table 1. Number of basis coefficients with a level exceeding 10 3 ,  single source.
Activation Function External Source PhaseCoefficients Quantity
(18)1
(18)72°42
(18)−18.9°42
(19)1
(19)72°1
(19)−18.9°1
(20)1
(20)72°1
(20)−18.9°15
Table 2. Number of basis coefficients with a level exceeding 10 3 , two sources.
Table 2. Number of basis coefficients with a level exceeding 10 3 , two sources.
Activation FunctionSource PhaseCoefficients Quantity
(18)(0°, 0°)2
(18)(72°, −72°)5
(19)(0°, 0°)2
(19)(72°, −72°)2
(20)(0°, 0°)4
(20)(72°, −72°)1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Panokin, N.V.; Averin, A.V.; Kostin, I.A.; Karlovskiy, A.V.; Orelkina, D.I.; Nalivaiko, A.Y. Method for Sparse Representation of Complex Data Based on Overcomplete Basis, l1 Norm, and Neural MFNN-like Network. Appl. Sci. 2024, 14, 1959. https://doi.org/10.3390/app14051959

AMA Style

Panokin NV, Averin AV, Kostin IA, Karlovskiy AV, Orelkina DI, Nalivaiko AY. Method for Sparse Representation of Complex Data Based on Overcomplete Basis, l1 Norm, and Neural MFNN-like Network. Applied Sciences. 2024; 14(5):1959. https://doi.org/10.3390/app14051959

Chicago/Turabian Style

Panokin, Nikolay V., Artem V. Averin, Ivan A. Kostin, Alexander V. Karlovskiy, Daria I. Orelkina, and Anton Yu. Nalivaiko. 2024. "Method for Sparse Representation of Complex Data Based on Overcomplete Basis, l1 Norm, and Neural MFNN-like Network" Applied Sciences 14, no. 5: 1959. https://doi.org/10.3390/app14051959

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop