Next Article in Journal
Intelligent Congestion Control in Wireless Sensor Networks (WSN) Based on Generative Adversarial Networks (GANs) and Optimization Algorithms
Previous Article in Journal
Optimizing Water Use in Maize Irrigation with Reinforcement Learning
Previous Article in Special Issue
End-to-End Power Models for 5G Radio Access Network Architectures with a Perspective on 6G
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Signal Detection for Enhanced Spatial Modulation-Based Communication: A Block Deep Neural Network Approach

1
The School of Computer Science and Engineering, Macau University of Science and Technology, Macau 999078, China
2
Department of Computer and Networking Engineering, University of Jeddah, Jeddah 21959, Saudi Arabia
3
Department of Computer Science, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(4), 596; https://doi.org/10.3390/math13040596
Submission received: 25 November 2024 / Revised: 30 January 2025 / Accepted: 8 February 2025 / Published: 11 February 2025

Abstract

:
As a novel variant of spatial modulation (SM), enhanced SM (ESM) provides higher spectral efficiency and improved bit error rate (BER) performance compared to SM. In ESM, traditional signal detection methods such as maximum likelihood (ML) have the drawback of high complexity. Therefore, in this paper, we try to solve this problem using a deep neural network (DNN). Specifically, we propose a block DNN (B-DNN) structure, in which smaller B-DNNs are utilized to identify the active antennas along with the constellation symbols they transmit. Simulation outcomes indicate that the BER performance related to the introduced B-DNN method outperforms both the minimum mean-square error (MMSE) and the zero-forcing (ZF) methods, approaching that of the ML method. Furthermore, the proposed method requires less computation time compared to the ML method.

1. Introduction

Multiple-input multiple-output (MIMO) is a widely used technology in current wireless communication systems, which can improve data transmission rate and increase channel capacity. But this technology also faces some practical issues such as inter-channel interference (ICI) and inter-antenna synchronization (IAS). Consequently, to solve these problems, spatial modulation (SM) was proposed in [1]. In terms of SM, at any given time slot, only one antenna is active, and a specific number of information bits are mapped to the index part and the constellation part, where the index part is used to select the activated antenna while the constellation part decides the constellation symbol.
However, SM only has one activated antenna at each time slot, which results in a low data rate. Hence, to solve this issue, enhanced SM (ESM) was proposed in [2]. ESM takes into account more than one signal constellation to increase the combination of antennas and symbols. When only one antenna is activated, that single antenna transmits symbols from the primary constellation. When two antennas are activated, both antennas simultaneously transmit symbols from the secondary constellation. In comparison to SM, it enhances spectral efficiency and achieves a lower bit error rate (BER), showing significant potential for future applications.
Deep neural network (DNN) has made profound inroads in the wireless communications, manifesting its prowess in diverse applications such as channel coding [3], antenna selection [4] and modulation classification [5] as well as other applications [6,7,8,9]. Recently, several extensions of DNN research achieved notable results. Several researchers have improved the structure of neural networks, leading to the development of more powerful networks, such as Kolmogorov–Arnold networks (KAN) [10]. Other researchers focus on integrating DNN with other technologies to enhance their performance in complex tasks, such as cognitive computing [11]. These efforts can refine and expand the applications of DNN in wireless communications, representing a promising direction for future research. At the same time, KAN and cognitive computing also have some applications in wireless communications. For instance, cognitive computing can be employed to predict the flow status of rectifiers, allowing for preemptive adjustments of network parameters and optimized power management. Moreover, cognitive computing can be utilized for real-time monitoring and analysis of communication signal characteristics, thereby optimizing signal processing algorithms. In terms of KAN, as a recent development in deep learning, it can be used to predict the performance of flexible electrohydrodynamic (EHD) pumps, which in turn can optimize the allocation of network resources. In terms of signal detection, several investigations have been conducted to improve the detection accuracy and complexity problem. In these works, some researchers attempt to improve the structure of neural network in order to enhance the performance of such neural networks in applications related to wireless communications [12,13,14]. If we consider the use of SM, in [15], the author used a block deep neural network (B-DNN) for signal detection in generalized SM (GSM). Experimental results showed that the B-DNN could approach the performance of the maximum likelihood (ML) method in terms of BER while exhibiting lower complexity. In [16], the author considered the use of neural network-assisted partial detection in GSM, with one neural network module for detecting the antenna index and another for detecting the received constellation symbol. This approach further reduced complexity. Authors in [17,18] explored the possibility of using DNN for signal detection in quadrature SM (QSM) and the results demonstrated the feasibility of DNN. Previous work has demonstrated that employing DNN for signal detection in GSM and QSM can reduce complexity while maintaining a low BER. In ESM, as a new variant of SM, to the best of our knowledge, the DNN method for signal detection in ESM has not yet been utilized.
In order to fill this gap, we propose a B-DNN method for signal detection in ESM. As an improved version of DNN, the B-DNN has already demonstrated its superior performance in signal detection for GSM. The B-DNN detector is capable of achieving a lower BER compared to the two linear detectors, minimum mean-square error (MMSE) and zero-forcing (ZF). In addition, it approaches the optimal BER performance of the ML detector while maintaining lower complexity and computational time. Moreover, compared to conventional DNN, the B-DNN shows improvements in both training time and accuracy. Furthermore, when compared to other well-known neural networks, such as convolutional neural network (CNN), the B-DNN exhibits better BER performance, with specific numerical results provided in [15]. Building on these findings, we apply this B-DNN to ESM to demonstrate whether this neural network architecture continues to perform well in another form of SM. In the data pre-processing phase, we utilize a feature vector generator (FVG) to transform complex in-phase and quadrature (IQ) data into a pure format, enhancing the performance in the symbol classification phase. During the training phase, we design a feed-forward neural network, which is meticulously trained on input vectors from the data pre-processing phase. The output of the neural network is classified using the softmax function, renowned for its efficacy in differentiating between multiple classes. The simulation results from our experiments clearly demonstrate that the B-DNN-based detector not only accurately approximates the ML detector but also significantly outperforms the traditional MMSE and ZF detectors in terms of BER. Moreover, B-DNN based detector can also reduce the classification time compared to ML detector, which greatly enhances the efficiency of the detection process, highlighting its potential for practical applications in ESM systems.

2. ESM System Model and Conventional Detectors

2.1. MIMO System Model

In MIMO system operating over quasi-static flat fading channel, the received signal is expressed as follows:
y = H x + n ,
where H C N r × N t denotes a channel matrix that follows a complex Gaussian distribution CN ( 0 , 1 ) with N r representing the number of antennas at the receiver and N t representing the number of antennas at the transmitter. x C N p × 1 is the transmitted signal vector. In ESM, when N p equals to 1, the primary constellation is used, and when N p equals to 2, the secondary constellation is used. The details of this mechanism are introduced in the next section. And the Gaussian noise n follows the distribution of CN ( 0 , σ 2 I ) , where I means identity matrix and σ is the variance of the noise and the size of this parameter is C N r × 1 .

2.2. ESM System Framework

In this paper, we consider the case of two transmit antennas (2-TX) in the ESM system, and the primary modulation is quadrature phase shift keying (QPSK). Based on the above, the transmitted vector set X can be represented as
X { [ C 4 0 ] , [ 0 C 4 ] , [ B 2 0 B 2 0 ] , [ B 2 1 B 2 1 ] } ,
where C 4 denotes the QPSK signal constellation used as the primary constellation and B 2 0 , B 2 1 represent two secondary constellations given by B 2 0 = { + 1 , 1 } , B 2 1 = { + i , i } . Clearly, the first two components of the transmission vector correspond to the scenario where a single antenna is activated, transmitting a symbol from the primary constellation. The last two components correspond to a scenario where two antennas are activated, and both transmission antennas can simultaneously transmit symbols from the secondary constellation. However, each of these symbols carries only 1 bit of information, which is half the amount carried by the symbols from the primary constellation. The goal of this approach is to match the bit quantity conveyed by the first and second components. Additionally, the derivation of the secondary constellation is primarily guided by the criterion of maximizing the minimum Euclidean distance between the primary and secondary constellations without increasing the transmitted signal power. We name B 2 0 as Binary Space Shift Keying 0 (BPSK0) and B 2 1 as Binary Space Shift Keying 1 (BPSK1). An example of the signal constellation implemented for this design can be observed in Figure 1, where four signal space elements are described in Table 1. In each time slot, the number of activated antennas is N p and this scheme transmits 4 bits per channel use (bpcu). The number of bits of information is divided into two parts, where one part is used for the selection of transmit antenna combinations (TACs), and the other part is used for carrying symbol bits. Based on this, the number of bits for TAC ( B T A C ) and symbol ( B S Y M ) can be expressed as
B T A C = log 2 T A C ,
B S Y M = N p log 2 M ,
where the values of N p and M depend on primary constellation and secondary constellation. Since we have N p activated antennas, we can represent the transmitted signal vectors x by
x = [ s 1 , s 2 , s 3 , , s N p 1 , 0 , s N p , , 0 ] T ,
with symbols s 1 , s 2 , s 3 , , s N p 1 , s N p S , and S is the constellation collection which includes all the symbols in Figure 1.

2.3. Conventional Detection

2.3.1. ML Detector Scheme

ML detection has the best BER performance, and it can be formulated as
( s ^ ) = a r g m i n s S | | y H s | | F 2 ,
which performs a comprehensive search of potential transmitted signal vectors to concurrently identify the activated antennas and constellation points, resulting in a substantial increase in decoding complexity of the receiver. This exhaustive methodology, while ensuring optimal detection accuracy, requires considerable computational resources.

2.3.2. Conventional Detect Method

As the number of transmit antennas and modulation levels increases, the complexity of the ML algorithm escalates exponentially. Therefore, several computational efficient algorithms such as ZF and MMSE are also employed and tested in this paper. The formulation for the ZF detection approach is provided in the subsequent equation:
x ^ Z F = H H y ( H H H ) 1 ,
where the conjugate transpose of matrix H can be represented as H H . This method is sensitive to channel noise, which limits its performance in the presence of strong noise effects. From this perspective, the MMSE method, which takes noise into account, yields better BER performance than ZF. The MMSE expression is as follows:
x ^ M M S E = ( H H H + σ 2 I ) 1 H H y .
MMSE detection takes into account the impact of noise during the equalization process, but in doing so, the noise may be amplified, especially when the channel conditions are poor. Therefore, MMSE detection requires a higher degree of accuracy in channel state information (CSI), and from this perspective, the difficulty of implementing MMSE is greater.

3. Proposed ESM Block-DNN Detector

In this section, we elucidate the working principle of the proposed detection mechanism. Figure 2 illustrates the framework of the ESM detector when one antenna is activated and sends QPSK signal. Regarding the secondary constellation, the constellation part only needs to be adjusted to two antennas sending the same BPSK0 or BPSK1 signal simultaneously. In order to enhance the precision of the B-DNN detector, we train separate neural networks for each case of ESM-2TX presented in Table 1. The main parameters are given in Table 2. Comparing with traditional DNN, we use a smaller B-DNN to detect the number of antennas activated by each case and the corresponding constellation symbols. In order to build the B-DNN, three distinct steps are required: data pre-processing, training and prediction.

3.1. Data Pre-Processing

According to [19], in the realm of neural networks, the efficacy of data pre-processing is paramount in ensuring the excellence of outcomes. Data pre-processing is an essential preliminary step that involves the meticulous refinement of raw data before they are imported to the neural network for training. This process includes data purification.

3.1.1. Raw Data

According to Equation (1), signal vector y is the sum of signals received from N r antennas. Concurrently, we hold four distinct combinations. In the scenario where only one antenna is activated, we use column vectors from the channel matrix H . In the other scenario, the entire channel matrix H is used as the input vector.

3.1.2. Feature Vector Generator

Given that the raw data consist of multiple matrices containing complex-valued IQ components, it is crucial to convert them into vector form. In this pivotal phase, we employ a sophisticated technique called separated FVG (SFVG) to handle raw data. The quintessence of SFVG lies in its capability to decompose the complex-valued IQ data into real-valued vector components. For instance, we input channel matrix H into SFVG and the output is given by the following expression:
f S F V G ( H ) = [ | R ( h 1 , 1 ) | , | I ( h 1 , 1 ) | , | R ( h 1 , 2 ) | , | I ( h 1 , 2 ) | , , | R ( h u , o ) | , | I ( h u , o ) | , , | R ( h N r , N t ) | , | I ( h N r , N t ) | ] T .

3.1.3. Expression of the Training Input

Upon obtaining the output of SFVG, we denote D j = [ d 1 , , d i , , d N ] as the jth element of the N-dimensional training dataset input, where j ranges over { 1 , 2 , , N s } , signifying the time slot index of the data. Given the original datasets that consist of N combinations of active antennas from the column of the channel matrix H I and the received signal vector y, d i can be expressed as
d i = [ f S F V G ( y j ) T , f S F V G ( H I i ( j ) ) T ] T ,
where d i R ( 2 N r + 2 N r N p ) × 1 .

3.2. B-DNN Parameters and Training

An array of L fully connected neural layers, including L–1 hidden layers, is designed to decode the signals emanating from each active transmit antenna at the transmitter. Table 2 delineates the parameters of the proposed DNN, including the count of layers, where the lth layer is endowed with δ l nodes.
We use λ to represent the comprehensive collection of parameters within the DNN, where λ is defined as λ = { λ 1 , λ 2 , , λ L } . The parameters specific to the lth layer are signified by λ l = { W l , b l } . Consequently, the lth layer is delineated as
Z l = σ ( W l Z l 1 + b l ) ,
where the weight matrix can be expressed as W l R δ l 1 × δ l and the deviation vector b l R δ l × 1 , σ ( · ) is the activation function. Apart from the final layer, the end of each layer uses a famous method, called Rectified linear unit (ReLU), and the expression is σ ( x ) = m a x ( 0 , x ) . The derivative of this method consistently yields a singular value, oscillating between the binary states of zero and one, thereby preventing the exponential diminution of gradient magnitude as it traverses the depth of numerous layers. The ReLU function facilitates rapid learning within the DNN. This in turn renders the efficient training of a profound supervised network feasible, obviating the need for unsupervised pre-training, as elucidated in [20]. In the penultimate layer, we employ the renowned softmax function to map the output onto a continuum that spans the interval 0 , 1 , and the mapping of the input and output of the DNN can be represented by the following equation:
Z L = σ ( W ( L ) ( σ ( W ( L 1 ) ( σ ( W ( 1 ) Z 0 + b 1 ) ) + b L 1 ) ) + b L ) ) .
We utilize categorical cross-entropy to ascertain the discrepancy in the cost function between the actual data and the predicted outcomes, thereby facilitating the identification of parameters that enhance our network’s optimization. We let Z T represent the one-hot encoded vector signifying the labels for the supervised training phase, which contrasts with the output of the predictions, denoted as Z L . The expression for this method is given as follows:
L ^ ( Z T , Z L ) = n = 1 M Z T n l o g ( Z L n ) ,
where the difference between Z T n and Z L n should be minimized. The objective is to effectuate its reduction to the greatest extent possible. To achieve this, we employ the stochastic gradient descent (SGD) optimization technique. This process involves the strategic adjustment of the existing weight values by either incrementing or decrementing them in accordance with the gradient’s magnitude, modulated by the learning rate, thus facilitating an efficient traversal towards the minimum of the optimization target. The parameter η , known as the learning rate, is a crucial hyper-parameter that is constrained to the interval from zero to one. It plays a significant role in the SGD algorithm, which incrementally refines the parameters over iterations through the application of gradient information. The following equation illustrates this process of updating the parameters:
W t + 1 ( l ) = W t ( l ) η L ^ ( Z T , Z L ) W t ( l ) .
The iterative refinement of weights persists in a cyclical fashion until the cost function’s value indicates a state of equilibrium over successive iterations. Upon attaining the optimal set of weights, the capability to make predictions is enabled through the utilization of the trained feed-forward DNN.
Throughout the training phase, it is essential to have the input vectors for training and their corresponding tags. The input vectors for training are derived from the channel matrix H as well as the received signal vector y , which could be formulated as d i = [ f S F V G ( y j ) T , f S F V G ( H I i j ) T ] T . For the tags, a one-hot encoding scheme is implemented, creating a vector of length M that corresponds to the transmitted symbol. This vector represents the M-PSK symbol constellation, thereby providing a distinct label for each point in the constellation.
At the same time, during the training phase, certain issues often arise, such as overfitting. To mitigate the impact of this problem, we implement several strategies. Firstly, we employ L2 regularization, which imposes a penalty on the sum of the squares of the weights, preventing them from becoming excessively large. This helps to reduce the complexity of the model and prevent overfitting. The regularization coefficient is set to 0.001 during the training phase. Additionally, batch normalization is incorporated after the hidden layers; this technique normalizes the features of each mini-batch to have a mean of zero and a variance of one. Furthermore, dropout is introduced at the end of the hidden layers, which randomly sets a proportion of neurons and their connections to zero with a certain probability. This prevents neurons within the network from becoming overly reliant on specific features, thereby encouraging the network to learn more robust and generalized feature representations. These strategies can significantly reduce the effects of overfitting.

3.3. Prediction Phase

After completing the aforementioned tasks, we begin to use this well-trained model for predictions. We possess N input vectors d n within the jth time slot of predictive phase and each of these vectors engenders N p output vectors. These output vectors, denoted as Z L , belong to the positive real space R + M × 1 and correspond to the k-th active antenna. Subsequently, the prediction of the transmitted symbol s ^ i k = S n and n ^ can be secured and expressed as the following formula:
( n ^ ) = arg max n { 1 , , M } ( Z L n ) .
Identifying the result through calculating the Euclidean distance between the received symbol and the estimated symbol is given in the expression of
( i ^ ) = arg min i { 1 , , N } y ( j ) E s H I i ( j ) s ^ i ( j ) F 2 ,
Through utilizing the smallest index i ^ , we proceed to perform the bit-mapping operation on the symbol s ^ i ^ . Lastly, the demodulation of the vector b ^ is carried out, yielding the output for the proposed B-DNN ESM system.

4. Simulation Results

In this section, we conduct numerical simulations to evaluate the performance of the proposed B-DNN-based detector for ESM systems (ESM-DNN). We mainly compare the DNN-based signal detection method with traditional methods to evaluate detection performance in terms of BER and time complexity. Note that in all simulations, we assume CSI is perfectly known. It is also worth noting that we train multiple DNNs to accommodate the different scenarios outlined in Table 1. Regarding BER, our criterion is that for the four cases, the number of simulations at a given SNR is consistent, and the BER is calculated as their average value. This approach helps to avoid inaccuracies in BER estimation due to excessive randomness in occurrence frequency.
Figure 3 compares the BER performance of the proposed ESM-DNN-based detection method with other traditional detection methods, namely ML, ZF and MMSE, in a 2 × 2 ESM communication system with QPSK modulation. As shown in Figure 3, it is clear that the ESM-ML detection approach excels in terms of BER performance. The proposed ESM-DNN detection method closely matches the BER performance of the ESM-ML method and outperforms the ESM-ZF and ESM-MMSE detection techniques.
In Figure 4, it is shown that the proposed ESM-DNN method also performs better under the 2 × 4 ESM system configuration with QPSK modulation. Figure 4 illustrates that the ESM-ML detection method consistently achieves the optimal BER performance. Within the SNR range of 0 to 4 dB, the BER of ESM-DNN performance is as same as that of ESM-MMSE and approaches that of ESM-ML. However, as the SNR increases, the ESM-DNN method demonstrates a distinct advantage, with its BER performance starting to significantly exceed those of ESM-MMSE and ESM-ZF while remaining close to ML.
Figure 5 illustrates that ESM-DNN under the 2 × 2 setting shows a significant advancement in runtime performance. When compared with the ML detection method, ESM-DNN reduces computational time and enhances operational efficiency. At the same time, it achieves near-optimal BER performance, thus striking a commendable balance between efficiency and accuracy. In Figure 6, when the number of antennas at the receiver increases to four, the computational time increases, but the trend of overall performance remains consistent.
The computational complexity analysis findings are summarized in Table 3. It becomes evident that the computational complexity associated with a DNN depends on the network architecture rather than the number of antennas used for transmission or reception. It is worth noting that to achieve an optimal balance between BER and computational complexity, the ESM-DNN stands out as the most suitable option.

5. Conclusions

In this paper, a B-DNN based detection approach for the ESM system is proposed. This scheme demonstrates significant performance improvements, in both BER and computational efficiency. The numerical results confirm that the B-DNN based scheme outperforms the ZF and MMSE detection methods in terms of BER. Moreover, although the B-DNN does not surpass ML detection method in terms of BER, the proposed B-DNN method clearly excels over the ML scheme in computational demand, offering a more expeditious processing time. In the future, the advanced technology like aggregation will be considered to further improve the system performance of the proposed method.

Author Contributions

Conceptualization, S.J. and Y.P.; methodology, S.J. and Y.P.; software, S.J.; validation, S.J. and Y.P.; formal analysis, S.J. and Y.P.; investigation, S.J.; resources, Y.P., F.A.-H. and M.M.M.; writing—original draft preparation, S.J.; writing—review and editing, Y.P., F.A.-H. and M.M.M.; visualization, S.J.; supervision, Y.P.; project administration, Y.P.; funding acquisition, Y.P., F.A.-H. and M.M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Taif University, Saudi Arabia, Project No. (TU-DSPP-2025-71).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable.

Acknowledgments

The authors extend their appreciation to Taif University, Saudi Arabia, for supporting this work through project number (TU-DSPP-2025-71).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Mesleh, R.; Haas, H.; Sinanovic, S.; Ahn, C.; Yun, S. Spatial modulation. IEEE Trans. Veh. Technol. 2008, 4, 2228–2241. [Google Scholar] [CrossRef]
  2. Cheng, C.; Sari, H.; Sezginer, S.; Su, Y.T. Enhanced Spatial Modulation With Multiple Signal Constellations. IEEE Trans. Commun. 2015, 6, 2237–2248. [Google Scholar] [CrossRef]
  3. Huang, H.; Guo, S.; Gui, G.; Yang, Z.; Zhang, J.; Sari, H. Deep learning for physical-layer 5G wireless techniques: Opportunities, challenges and solutions. IEEE Wirel. Commun. 2020, 1, 214–222. [Google Scholar] [CrossRef]
  4. Zhu, F.; Hai, H.; Jiang, X. Efficient Antenna Selection for Adaptive Enhanced Spatial Modulation: A Deep Neural Network Approach. IEEE Commun. Lett. 2023, 5, 1352–1356. [Google Scholar] [CrossRef]
  5. O’Shea, T.; Hoydis, J. An introduction to deep learning for the physical layer. IEEE Trans. Cognit. Commun. Netw. 2017, 4, 563–575. [Google Scholar] [CrossRef]
  6. Jiang, C.; Zhang, H.; Ren, Y.; Han, Z.; Chen, C.; Han, Z. Machine learning paradigms for next-generation wireless networks. IEEE Wirel. Commun. 2017, 4, 98–105. [Google Scholar] [CrossRef]
  7. Dorner, S.; Cammerer, J.; Hoydis, J.; Brink, S. Deep learning based communication over the air. IEEE J. Sel. Top. Signal Process. 2018, 2, 132–143. [Google Scholar] [CrossRef]
  8. Aoudia, A.; Hoydis, J. End-to-end learning of communications systems without a channel model. In Proceedings of the 52nd Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 28–31 October 2018; pp. 298–303. [Google Scholar]
  9. Kim, H.; Jiang, Y.; Rana, R.; Kannan, S.; Viswanath, P. Communication algorithms via deep learning. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–17. [Google Scholar]
  10. Pinkus, A. Kolmogorov-Arnold neural network (KAN): A deep learning approach for function approximation. IEEE Trans. Neural Netw. Learn. Syst. 2017, 9, 2134–2145. [Google Scholar]
  11. Pineau, J.; Paquet, S.; Precup, D. Cognitive Computing: A Roadmap for the Future. IBM J. Res. Dev. 2016, 60, 1–10. [Google Scholar]
  12. Alawad, A.; Hamdi, A. A Deep Learning-Based Detector for IM-MIMO-OFDM. In Proceedings of the 2021 IEEE 94th Vehicular Technology Conference, Norman, OK, USA, 27–30 September 2021; pp. 2465–2477. [Google Scholar]
  13. Zhang, J.; Li, Y.; Huang, W.; Guo, H.; Yuen, C.; Guan, L. DNN-Aided Message Passing Based Block Sparse Bayesian Learning for Joint User Activity Detection and Channel Estimation. In Proceedings of the 2019 IEEE VTS Asia Pacific Wireless Communications Symposium, Singapore, 28–30 August 2019; pp. 1178–1203. [Google Scholar]
  14. Liao, Y.; Deng, H.; Xie, Y.; Liu, J.; Yuan, B. Low-complexity Neural Network-based MIMO Detector using Permuted Diagonal Matrix. In Proceedings of the 2020 54th Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 1–4 November 2020; pp. 1058–1065. [Google Scholar]
  15. Albinsaid, H.; Singh, K.; Biswas, S.; Li, C.P.; Alouini, M.S. Block deep neural network-based signal detector for generalized spatial modulation. IEEE Commun. Lett. 2020, 12, 2775–2779. [Google Scholar] [CrossRef]
  16. Shamasundar, B.; Chockalingam, A. A DNN Architecture for the Detection of Generalized Spatial Modulation Signals. IEEE Commun. Lett. 2020, 12, 2770–2773. [Google Scholar] [CrossRef]
  17. Zhu, J.; Deng, Y. Graph Neural Network Based Signal Detection in Quadrature Spatial Modulation System. In Proceedings of the 2023 IEEE the 23rd International Conference on Communication Technology (ICCT), NJ, CN, Wuxi, China, 20–22 October 2023; pp. 216–220. [Google Scholar]
  18. Shu, Y.; Peng, Y.; AL-Hazemi, F. Deep Learning Based Signal Detection for Quadrature Spatial Modulation System. China Commun. 2021, 5, 1–7. [Google Scholar]
  19. Mao, Q.; Hu, F.; Hao, Q. Deep learning for intelligent wireless networks: A comprehensive survey. IEEE Commun. Surv. Tutor. 2018, 4, 2595–2621. [Google Scholar] [CrossRef]
  20. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the 2011 14th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
Figure 1. An illustration of the constellations used. The crosses represent QPSK, the circles represent BPSK0 and the squares represent BPSK1 signal constellation.
Figure 1. An illustration of the constellations used. The crosses represent QPSK, the circles represent BPSK0 and the squares represent BPSK1 signal constellation.
Mathematics 13 00596 g001
Figure 2. Diagram of ESM communication system with DNN signal detection.
Figure 2. Diagram of ESM communication system with DNN signal detection.
Mathematics 13 00596 g002
Figure 3. The BER comparison for various detectors under N t = 2 and N r = 2.
Figure 3. The BER comparison for various detectors under N t = 2 and N r = 2.
Mathematics 13 00596 g003
Figure 4. The BER comparison for various detectors under N t = 2 and N r = 4.
Figure 4. The BER comparison for various detectors under N t = 2 and N r = 4.
Mathematics 13 00596 g004
Figure 5. The runtime complexity comparison for various detectors under N t = 2 and N r = 2.
Figure 5. The runtime complexity comparison for various detectors under N t = 2 and N r = 2.
Mathematics 13 00596 g005
Figure 6. The runtime complexity comparison for various detectors under N t = 2 and N r = 4.
Figure 6. The runtime complexity comparison for various detectors under N t = 2 and N r = 4.
Mathematics 13 00596 g006
Table 1. Enhanced SM, 2TX-4 bpcu.
Table 1. Enhanced SM, 2TX-4 bpcu.
Tx1Tx2
C1QPSK0
C20QPSK
C3BPSK0BPSK0
C4BPSK1BPSK1
Table 2. Significant Parameters in DNN.
Table 2. Significant Parameters in DNN.
Tx1
Input nodes2( N r N p + N r )
Learning rate0.005
Hidden layer3
Number of training set15,000,000
Output nodesM
Number of validation set5,000,000
Hidden layer activationReLu
Epoch50
Output layer activationSoftmax
Loss functionCross-entropy
QPSK hidden nodes256-128-64
BPSK0 hidden nodes256-128-64
BPSK1 hidden nodes256-128-64
OptimizerSGD
Table 3. Computational complexity.
Table 3. Computational complexity.
DetectorReal-Valued Multiply-Accumulate Operations
ESM-ML ( 2 × N r × N t × M N p + 1 ) × S × C
ESM-ZF N r × N t × log 2 ( M ) + O ( N r 3 )
ESM-MMSE 2 × ( N r × N t ) + O ( N r 3 ) + S × log 2 ( M ) + S
ESM-DNN N N p ( ( 4 N r N p + 4 N r 1 ) δ 1 + k = 1 k = L 1 δ k + 1 ( 2 δ k 1 ) )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, S.; Peng, Y.; AL-Hazemi, F.; Mirza, M.M. Signal Detection for Enhanced Spatial Modulation-Based Communication: A Block Deep Neural Network Approach. Mathematics 2025, 13, 596. https://doi.org/10.3390/math13040596

AMA Style

Jin S, Peng Y, AL-Hazemi F, Mirza MM. Signal Detection for Enhanced Spatial Modulation-Based Communication: A Block Deep Neural Network Approach. Mathematics. 2025; 13(4):596. https://doi.org/10.3390/math13040596

Chicago/Turabian Style

Jin, Shaopeng, Yuyang Peng, Fawaz AL-Hazemi, and Mohammad Meraj Mirza. 2025. "Signal Detection for Enhanced Spatial Modulation-Based Communication: A Block Deep Neural Network Approach" Mathematics 13, no. 4: 596. https://doi.org/10.3390/math13040596

APA Style

Jin, S., Peng, Y., AL-Hazemi, F., & Mirza, M. M. (2025). Signal Detection for Enhanced Spatial Modulation-Based Communication: A Block Deep Neural Network Approach. Mathematics, 13(4), 596. https://doi.org/10.3390/math13040596

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop