Next Article in Journal
Remaining Useful Life Prediction of Bearings via Semi-Supervised Transfer Learning Based on an Anti-Self-Healing Health Indicator
Previous Article in Journal
Simple and Cost-Effective Design of a THz-Metamaterial-Based Hybrid Sensor on a Single Substrate
Previous Article in Special Issue
Intelligent Health Monitoring in 6G Networks: Machine Learning-Enhanced VLC-Based Medical Body Sensor Networks
 
 
Due to scheduled maintenance work on our database systems, there may be short service disruptions on this website between 10:00 and 11:00 CEST on June 14th.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Demonstration of 50 Gbps Long-Haul D-Band Radio-over-Fiber System with 2D-Convolutional Neural Network Equalizer for Joint Phase Noise and Nonlinearity Mitigation

Key Laboratory for Information Science of Electromagnetic Waves (MoE), Shanghai Institute for Advanced Communication and Data Science, Fudan University, Shanghai 200433, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(12), 3661; https://doi.org/10.3390/s25123661
Submission received: 30 April 2025 / Revised: 3 June 2025 / Accepted: 10 June 2025 / Published: 11 June 2025
(This article belongs to the Special Issue Recent Advances in Optical Wireless Communications)

Abstract

:
High demand for 6G wireless has made photonics-aided D-band (110–170 GHz) communication a research priority. Photonics-aided technology integrates optical and wireless communications to boost spectral efficiency and transmission distance. This study presents a Radio-over-Fiber (RoF) communication system utilizing photonics-aided technology for 4600 m long-distance D-band transmission. We successfully show the transmission of a 50 Gbps (25 Gbaud) QPSK signal utilizing a 128.75 GHz carrier frequency. Notwithstanding these encouraging outcomes, RoF systems encounter considerable obstacles, including pronounced nonlinear distortions and phase noise related to laser linewidth. Numerous factors can induce nonlinear impairments, including high-power amplifiers (PAs) in wireless channels, the operational mechanisms of optoelectronic devices (such as electrical amplifiers, modulators, and photodiodes), and elevated optical power levels during fiber transmission. Phase noise (PN) is generated by laser linewidth. Despite the notable advantages of classical Volterra series and deep neural network (DNN) methods in alleviating nonlinear distortion, they display considerable performance limitations in adjusting for phase noise. To address these problems, we propose a novel post-processing approach utilizing a two-dimensional convolutional neural network (2D-CNN). This methodology allows for the extraction of intricate features from data preprocessed using traditional Digital Signal Processing (DSP) techniques, enabling concurrent compensation for phase noise and nonlinear distortions. The 4600 m long-distance D-band transmission experiment demonstrated that the proposed 2D-CNN post-processing method achieved a Bit Error Rate (BER) of 5.3 × 10−3 at 8 dBm optical power, satisfying the soft-decision forward error correction (SD-FEC) criterion of 1.56 × 10−2 with a 15% overhead. The 2D-CNN outperformed Volterra series and deep neural network approaches in long-haul D-band RoF systems by compensating for phase noise and nonlinear distortions via spatiotemporal feature integration, hierarchical feature extraction, and nonlinear modelling.

1. Introduction

The fast progression of worldwide informatization is instigating significant modifications in mobile communication technology. With the advent of the 6G communication era, burgeoning applications like the metaverse, autonomous driving, and holographic communication necessitate wireless networks characterized by ultra-high data speeds (about 100 Gbps), ultra-low latency (sub-millisecond), and ultra-large connection density [1]. The conventional microwave frequency ranges are becoming progressively insufficient for these demands. In this context, the D-band (110–170 GHz) has emerged as a highly attractive contender for 6G wireless communication. The enormous bandwidth of the D-band facilitates substantial data transmission, offering the capability for high-rate communications [2].
Photonics-aided millimeter-wave generation uses optical heterodyne beating and optical frequency comb generation to overcome bandwidth, transmission loss, and phase stability issues in conventional electronic devices at the D-band (110–170 GHz) [3]. This method uses photonic techniques to achieve high spectral purity, low phase noise, and wide bandwidth scalability, enabling the reliable generation of high-frequency millimeter-wave signals in the demanding D-band frequency range [4].
RoF technology integrates the high-bandwidth, low-loss attributes of optical communication with the adaptability of wireless communication [5,6,7]. In a RoF system, radio-frequency (RF) signals are modulated onto optical carriers, transmitted over optical fibers, and subsequently converted back to RF signals at the receiver. This method capitalizes on the low-loss characteristics of optical fibers to efficiently increase signal transmission distance and employs the high-bandwidth benefits of optical communication to improve system transmission capacity [8,9,10,11].
The nonlinear properties of optoelectronic devices, such as modulators and photodetectors, generate supplementary harmonic components during signal modulation and detection, disrupting the original signal. The Kerr effect in fiber transmission generates nonlinear phase noise by rendering the optical signal’s phase contingent upon its intensity. These effects collectively result in amplitude distortion and phase noise, significantly impairing system performance. The restricted linewidth of lasers presents another challenge: while laser output is not a perfect monochromatic signal but encompasses a specific frequency range (i.e., linewidth), it results in random phase drift. As transmission distance and signal rate escalate, the accumulation of phase drift causes rotation and diffusion of the constellation diagram of received signals, degrading signal quality and elevating the bit error rate [12].
In D-band RoF systems, addressing nonlinear distortion and phase noise compensation is a significant research problem. Classical nonlinear approaches like the Volterra series are based on linear time-invariant (LTI) system theory. These methods estimate nonlinear distortion using polynomial models to demonstrate how power amplifiers (PAs) link input and output as polynomial terms and delayed input signals. Limitations include the fast growth of model complexity as nonlinear order rises, which requires more processing power and hardware. In real-world application, high-order model parameters are hard to estimate, fit too tightly to training data, and perform poorly in varied conditions, resulting in inaccurate compensation performance.
Due to their robust nonlinear fitting characteristics, deep neural networks (DNNs) have seen extensive use in nonlinear compensation in recent years. They can autonomously acquire complex signal feature patterns to mitigate nonlinear distortion [13,14]. Nevertheless, DNNs have significant constraints in phase noise compensation. Phase noise is a random process that changes over time, making it hard for traditional statistical learning methods to effectively understand its changing characteristics. Furthermore, phase noise and nonlinear distortion demonstrate a coupling effect; nonetheless, current methods frequently treat them separately, overlooking their interrelations. This oversight results in diminished compensation efficiency and inadequate enhancement of system performance.
Two-dimensional convolutional neural networks (2D-CNNs) have been very successful in areas like image processing and computer vision, showing strong abilities to identify features over time and space [15]. Their basic structure includes convolutional layers, pooling layers, and fully connected layers, which help in effectively handling grid-like data and finding patterns and relationships in the information. Research teams have successfully adapted 2D-CNNs for wireless communication, leading to better results in improving signal quality, lowering bit error rates (BER), and increasing data capacity [16,17]. This work offers novel insights for concurrently tackling nonlinear distortion and phase noise in D-band RoF systems. The 2D-CNN can learn the spatial correlations and distribution characteristics of phase noise in signals by adaptively updating weight parameters, dynamically adjusting convolutional kernel weights during backpropagation optimization to effectively model and suppress phase noise. However, putting this into practice faces challenges, such as creating the right network designs and settings that fit the system’s specific issues, as well as choosing the right training data and methods to improve performance and reliability when channel conditions change and device parameters vary. In recent years, a series of research achievements have been made in the field of terahertz signal transmission, and the relevant progress is shown in Table 1 [18,19,20,21,22,23].
This research presents a 2D-CNN post-processing technique grounded in spatiotemporal feature fusion. In a 4600 m experiment using a single-channel photonics-assisted terahertz wireless transmission, the traditional Digital Signal Processing (DSP) technique was first used, and then the new 2D-CNN algorithm was applied to improve the signal. The experiments showed that this new method reduced nonlinear noise better and lowered the bit error rate (BER) compared to traditional DSP algorithms, along with a detailed look at the results. A D-band RoF system using photonics-assisted technology was created, allowing for stable transmission of QPSK signals at a speed of 25 Gbaud (which is 50 Gbps) over a distance of 4600 m at a frequency of 128.75 GHz. The system used a method to create millimeter-wave signals: it takes electrical signals, changes them into optical signals with a Mach–Zehnder modulator (MZM), boosts them with a PA, converts them into high-frequency millimeter-wave signals with a photodetector (PD), and then sends them wirelessly using strong directional antennas. Experimental findings indicate that our photonics-aided millimeter-wave generating method enabled steady transmission of a 50 Gbps (25 Gbaud) QPSK signal across a distance of 4600 m at 128.75 GHz, confirming its viability for long-distance communications in the D-band. In a transmission scenario with an optical power of 8 dBm, the proposed method diminished the system BER to 5.3 × 10−3, markedly surpassing the traditional Volterra series (BER = 1.4 × 10−2) and DNN approaches (BER = 1.3 × 10−2) and remaining below the SD-FEC threshold of 1.56 × 10−2.

2. Principle

2.1. Phase Recovery Theory

In RoF systems, phase noise caused by laser linewidth generates intricate interference. Conventional phase recovery algorithms encounter difficulties in accurately extracting and compensating for phase fluctuations [24]. Neural networks characterize phase noise in the channel by training on extensive datasets and exhibiting distinct benefits [25,26].
The QPSK symbol sequence produced by the transmitter constitutes a time-domain signal. The width of a laser’s light beam causes the signal’s phase to shift over time, resulting in random changes in the phase that can be considered a sequence of offsets θ n . Mathematically, we represent the received signal as complex exponential modulation:
y ( n ) = x ( n ) e j θ n h n + n n
where θ n represents the phase noise of the n-th symbol, denotes the convolution operation, h n is the channel impulse response, and n n is additive Gaussian white noise.
By decomposing the time-domain received signal into its real and imaginary components, we formulate a two-dimensional input tensor:
I n p u t = a n b n R M × 2
where a n and b n denote the real and imaginary components, M signifies the duration of the time frame, and the two channels correspond to the real-part and imaginary-part channels. Figure 1 demonstrates the two-dimensional structure of the input data. Continuous symbol segments are retrieved using a sliding window that encompasses the current symbol and M 1 2 neighboring symbols on either side.
Each row of the input tensor shows the real and imaginary parts of a single symbol, and each column relates to the real or imaginary parts of nearby symbols, forming a two-dimensional feature matrix that helps with 2D convolution extraction. The temporal direction refers to the phase continuity of neighboring symbols, whereas the channel direction pertains to the amplitude-phase mapping connection between the real and imaginary components.
Under optimal circumstances, the real and imaginary components of the channel fulfil a 2 + b 2 = 1 . In the event of a phase deflection θ , according to Euler’s formula e j θ = cos θ + j sin θ , the signal is expressed as:
  a + j b cos θ + j sin θ = a cos θ b sin θ + j a sin θ + b cos θ
The new real component a = a cos θ b sin θ and the imaginary component b = a sin θ + b cos θ are derived. Post-normalization, the total number of squares remains equal to 1; however, the phase has altered. Through extensive data training, the convolution kernel can recognize the “abnormal” coupling pattern induced by phase rotation. The convolution kernel weights will enhance the response to this pattern, thereby extracting the attributes of phase rotation.
Phase correction is analogous to doing a complex conjugate rotation on the incoming signal:
y ^ n = y n e j θ ^ n = a n + j b n cos θ ^ n j sin θ ^ n
Upon expansion, the real component is a ^ n = a n cos θ ^ n + b n sin θ ^ n , while the imaginary component is b ^ n = a n sin θ ^ n + b n cos θ ^ n . The 2D-CNN does the inverse transformation of phase deflection by acquiring the weights of the convolution kernel.
The mathematical representation of the convolution operation is:
Y l = Re L U X l K l + B l
Here, X l denotes the output of the l-th convolutional layer, K l signifies the convolution kernel, and B l indicates the bias. ReLU is an activation function that imparts nonlinear properties, allowing the network to learn intricate phase-mapping correlations. The mean squared error (MSE) serves as the loss function, defined as the average squared discrepancy between the real and imaginary components of the recovered signal and the original signal:
L = 1 2 M m = 1 M | | s ^ m s m | | 2 2 + λ θ | | θ | | 1
Let M denote the batch size, s ^ m = a ^ m , b ^ m represent the reconstructed signal, s m = a m , b m signify the original signal, and λ indicate the L 1 regularization coefficient. Figure 2 shows the flowchart of 2D-CNN. The network has convolutional, batch normalization, activation, pooling, and fully connected layers. The initial and subsequent convolutional layers are succeeded by a batch normalization layer (BN) to standardize feature maps, mitigate internal covariate shift, and accelerate training convergence. RELU introduces non-linearity, enabling the network to develop intricate patterns. POOL layers downsample features by max-pooling, decreasing dimensionality while preserving essential information.
Subsequent to the convolutional and pooling stages, the feature maps are transformed into a one-dimensional vector and input into a fully connected layer (FCL), which integrates the accumulated spatial-temporal characteristics to reconstruct the signal. The FCL transforms high-dimensional feature space into output space, enabling the network to generate final predictions or indicate adjustments. This architecture enables the 2D-CNN to integrate spatiotemporal data, therefore diminishing phase noise and nonlinear distortions.
In backpropagation, the gradient of the loss function concerning the convolution kernel L Κ l is computed using the following steps:
The gradient of the loss concerning the output feature map is:
L Y ( l ) = δ ( l + 1 ) K ( l + 1 )
Here, δ ( l ) denotes the error value of the l-th layer. δ ( l + 1 ) = L Y ( l + 1 ) = s ^ s , representing the error value of the subsequent layer. The derivative of the activation function concerning the linear transformation is:
Y ( l ) ( X ( l ) K ( l ) ) = ReLU ( Y ( l ) ) = I ( Y ( l ) > 0 )
Here, I ( ) denotes an indicator function that equals 1 when the input exceeds 0 and 0 otherwise. The gradient of the feature map concerning the convolution kernel is:
( X ( l ) K ( l ) ) K ( l ) = X ( l )
Consolidated using the chain rule:
L K ( l ) = L Y ( l ) Y ( l ) ( X ( l ) K ( l ) ) ( X ( l ) K ( l ) ) K ( l )
Here, ⊙ denotes the element-wise product (Hadamard product), where corresponding elements of two matrices are multiplied together. Deriving the formula for error term propagation:
δ ( l ) = δ ( l + 1 ) K ( l + 1 ) σ Y ( l )
L K ( l ) = i = 1 N δ i ( l ) X i ( l ) + λ sign ( K ( l ) )
Here, s i g n indicates the sign function. The forward propagation of the fully linked layer is as follows:
s ^ = W f c flatten ( Y ( L ) ) + b f c
where s ^ = a ^ ( n ) b ^ ( n ) . The gradient of the loss concerning the output is:
L s ^ = 1 N ( s ^ s )
The gradient of the output concerning the input of the fully connected layer is:
s ^ flatten ( Y ( L ) ) = W f c T
The gradient of the input to the fully connected layer concerning the weight is:
flatten ( Y ( L ) ) W f c = flatten ( Y ( L ) ) T
Consolidated using the chain rule:
L W f c = L s ^ s ^ flatten ( Y ( L ) ) flatten ( Y ( L ) ) W f c
Ultimately, we derive:
L W f c = 1 N ( s ^ s ) flatten ( Y ( L ) ) T + λ sign ( W f c )
The gradient of the bias b f c is expressed as:
L b f c = L s ^ s ^ b f c = 1 N ( s ^ s )
The parameter update formula using the AdamW optimizer, which separates weight decay from gradient computation, is as follows:
Estimation of gradient momentum and variance:
m t = β 1 m t 1 + ( 1 β 1 ) g t
v t = β 2 v t 1 + ( 1 β 2 ) g t 2
m ^ t = m t 1 β 1 t , v ^ t = v t 1 β 2 t
Here, g t represents the current gradient, whereas β 1 = 0.9 and β 2 = 0.999 denote the momentum coefficients. The parameter update including weight decay is expressed as:
θ t = θ t 1 η m ^ t v ^ t + ϵ + λ θ t 1
Here, η denotes the learning rate, λ signifies the weight-decay coefficient, ε = 10 8 represents the smoothing term, and θ encompasses parameters such as K l , W f c , and b f c , among others. Through the optimization of weights via backpropagation and the minimization of the MSE, a comprehensive mapping from the noisy signal to the denoised output post-neural network processing is accomplished.
f C N N a n b n ; W = cos θ ^ n sin θ ^ n sin θ ^ n cos θ ^ n a n b n = a ^ n b ^ n
The network outputs a ^ n and b ^ n constitute the reconstructed complex signal y ^ n = a ^ n + j b ^ n .

2.2. Principle of Two-Dimensional Convolutional Neural Networks

A 2D-CNN is a deep learning model, mainly made up of convolutional layers, pooling layers, and fully connected layers [27]. The convolutional layer utilizes the feature maps from the preceding layer as input for convolution processes, subsequently generating the output feature map via an activation function [28,29]. The output feature map can be represented as:
x n l = f ( k = 1 K x k l 1 W n k l + b n l )
Here, W nk l denotes the kernel connecting the n-th and k-th features of the l-th layer, while x n l and x k l 1 signify the n-th and k-th feature maps of the l-th and (l − 1)-th layers, respectively. Additionally, b n l represents the associated bias.
In executing intricate convolution processes, we typically split them into real-valued operations. The 2D-CNN works on the real and imaginary parts of the complex signal separately before combining them, which allows it to perform the convolution operation on the complex signal. Activation functions are crucial in neural networks because they introduce nonlinear properties to the network. After a linear transformation in the convolutional layer, the results are mixed with a nonlinear activation function before moving on to the next layer. In 2D-CNN, there are two common types of activation functions: traditional S-type functions like sigmoid and tanh, and non-saturated functions like ReLU, Leaky ReLU, ELU, PReLU, and RReLU [30]. However, traditional S-type functions can cause the problem of vanishing gradients as the model gets deeper. However, with traditional S-type functions, the problem of vanishing gradients starts to occur as the model gets deeper. Non-saturated activation functions have been proposed to address this issue [31,32].
The ReLU function is predominantly utilized in convolutional layers. The mathematical representation of ReLU is as follows:
Re L U x = { x , x < 0 x , x 0 = max 0 , x
When x is non-negative, ReLU outputs the input, whereas if x is negative, it outputs 0. This simple process gives the ReLU function nonlinear features and remarkable computing efficiency. Sigmoid functional representation:
S i g m o i d ( x ) = 1 1 + e x
The range is [0, 1], making it ideal for model output functions that generate probability values inside (0, 1). This method works for models that produce projected probabilities, such as binary classification categories or confidence levels. However, above a certain range, this function’s derivative approaches 0. This will cause neurons to become saturated during training, slowing weight updates during backpropagation and complicating model training. We call this the vanishing gradient issue. The tanh function formula is:
tanh ( x ) = e x e x e x + e x
The hidden layer uses the tanh function, whereas the output layer uses the sigmoid function in normal binary classification tasks. In 2019, Diganta Misra announced the Mish activation function [33]. This research used it:
M i s h x = x tanh ln 1 + e x
Mish accepts negative inputs and prevents ReLU from rejecting negative gradients using ReLU’s straight component and tanh’s curves. Mish is smoother and has fewer vanishing gradients than ReLU. Dropout [34], a regularization technique for neural network training, mitigated this issue in the 2D-CNN. Dropout randomly deactivates neurons with a probability p, resulting in an output of 0. Neuron output is maintained throughout training and testing by retaining and scaling all neurons.
A loss function quantifies the discrepancy between expected and real labels. The loss function modifies the parameters of network training. Changing model parameters lowers training loss function. The primary metric is MSE:
M S E = 1 n i = 1 n ( y i y ^ i ) 2
In this formula, y i is the actual value, y ^ i is the predicted value, and n is the total number of samples. Cross-entropy loss measures how different two probability distributions are and is often used in classification tasks, especially when using the softmax activation function to create probability distributions. In the context of binary classification problems:
L = 1 n i = 1 n y i log y ^ i + 1 y i log 1 y ^ i
For multi-classification problems:
L = 1 n i = 1 n j = 1 C y i j log y ^ i j
where C denotes the number of categories. We assessed mean squared error and cross-entropy loss during the training process. The MSE quantifies the average squared deviation between predicted and actual values, thereby assessing the model’s predictive capability for signal recovery and continuous signal values. If signal recovery, continuous signal value forecasting, or constellation diagram production is unnecessary, using cross-entropy as the loss function may enhance classification outcomes and reduce bit-error rates.
We evaluated three optimizers throughout the training process: stochastic gradient descent (SGD), Adam, and AdamW [35]. Optimizers adjust parameters autonomously according to rules derived from gradients generated by backpropagation, which reflect changes in loss associated with the parameters. SGD, the fundamental optimizer, iteratively adjusts model parameters with mini-batch gradients. The formula is:
θ t + 1 = θ t η θ t L θ t
The advantages include straightforward implementation, little processing burden, limited hyperparameters, and effective performance with small datasets or uncomplicated models. The learning rate dictates the speed of learning and may result in entrapment inside local minima, rendering the process unstable.
Adam modifies the learning rate via momentum by averaging gradients and their squares. The approach accelerates convergence in complex data distributions, exhibits insensitivity to the initial learning rate, and requires few parameter adjustments. Adam employs more memory and may encounter convergence issues. AdamW, an enhanced version of Adam, updates parameters without gradient computation and incorporates weight decay for regularization. This technique enhances model generalization and stability by removing the effects of momentum and adaptive learning rates on classic Adam’s L2 regularization. The automatic modification of the learning rate and weight-decay coefficient mitigates overfitting.
This study carefully looks at different neural network algorithms and traditional DSP algorithms through organized experiments; measures how things like model complexity, optimizers, and activation functions affect phase-recovery performance; and finds the best model design and settings for the D-band RoF system in these situations.

3. Experimental Setup

In this experiment, we built a single-channel experimental setup for long-distance wireless communication in the D-band using photonics-assisted technologies. The transmitter and receiver were situated on two campuses of Fudan University. The wireless transmitter was positioned atop Guanghua Building in Handan Campus, which stands at 142 m, and the wireless receiver was situated on the top floor of the Physics Building, at a 4600 m distance from the transmitter and at a height of 24 m. The system architecture is depicted in Figure 3a.
Figure 3b illustrates the distinct procedures pertaining to the transmitter DSP (Tx DSP) and receiver DSP (Rx DSP). The baseband signal was created offline in MATLAB (R2023b) on the transmitter side. The bit sequence was first mapped to QPSK signals. Subsequent to up-sampling, the signals underwent root-raised cosine (RRC) pulse shaping with a roll-off value of 0.01. Next, we produced the digital signal by resampling and input it into a 120 GSa/s arbitrary waveform generator (AWG), resulting in a 25 Gbaud QPSK signal. The signals were enhanced by two parallel electrical amplifiers (EAs), each possessing a gain of 25 dB.
Two free-running external-cavity lasers (ECLs) were utilized in the experiment. ECL-1 possessed a wavelength of 1550.00 nm, a linewidth under 100 kHz, and an output power of 12 dBm; ECL-2 had a wavelength of 1551.08 nm, with identical linewidth and output power as ECL-1. The steady light from ECL-1 was adjusted by an I/Q modulator that had a 3 dB bandwidth of 30 GHz, using the 12 dBm optical carrier from ECL-1 to complete the optical modulation. The modulated signal was boosted to 12 dBm by polarization maintaining erbium-doped fiber amplifier (PM-EDFA), and then it was combined with optical carrier from ECL-2 by polarization maintaining optical combiner (PM-OC). The combined signal travelled through 100 m of Single-Mode Fiber-28, and then its optical power was adjusted by a variable optical attenuator (VOA) before it went into a uni-traveling carrier photodiode (UTC-PD). In the UTC-PD, a D-band millimeter-wave signal was produced using optical heterodyne beating. Due to the quick attenuation of the signal during free-space transmission, amplification was necessary. Cascaded LNA-1, possessing a gain of 18 dB, and PA, exhibiting a saturated output power of 13 dBm, were employed to facilitate 4600 m wireless transmission. A transmitting horn antenna (HA-1) with a gain of 25 dB emitted the enhanced signal into free space. To improve the directivity and transmission range of the signal, a pair of lenses was positioned in front of the transmitting and receiving antennas, respectively. Lens 1 possessed a diameter of 10 cm, whereas Lens 2 had a diameter of 60 cm.
At the receiver end, the signal was initially amplified by LNA-2 with a gain of 33 dB. The 12.51 GHz local oscillator (LO) was then increased in strength by 12 times in the mixer, mixed with the terahertz signal, and changed into an intermediate-frequency (IF) signal. The down-converted IF signal was then boosted by EA-3, which increased its strength by 26 dB, and was finally recorded by a digital storage oscilloscope (DSO) that sampled at 100 GSa/s and had a 3 dB bandwidth of 33 GHz.
Figure 3b illustrates that the IF signal acquired at the receiver was processed offline using DSP in MATLAB. Initially, down-conversion and resampling were executed to transform the IF signal into a baseband signal. A 31-tap T/2 constant modulus algorithm (CMA) was used to improve the signal quality and reduce interference caused by the way the channel transmits data. Next, frequency-offset estimation (FOE) was done to reduce the effects of frequency drift caused by the laser’s frequency offset, and carrier-phase recovery (CPR) was performed to improve the phase noise caused by the laser’s linewidth. A 51-tap direct-detection least-mean-square (DD-LMS) technique was employed for equalization.
After using traditional DSP algorithms, the resulting signal was further equalized with a 2D-CNN, a 1D-CNN, a 1D-DNN, and a Volterra nonlinear equalizer (VNE). The VNE used signal-processing methods to reduce distortion in the signal caused by device nonlinearity. The 1D-DNN and 1D-CNN used the learning abilities of neural networks to identify by stacking fully connected layers and using one-dimensional convolution, which helped reduce inter-symbol interference and minor nonlinear issues. The 2D-CNN changed the complex signal into a two-dimensional format that included both real and imaginary parts, allowing it to handle the amplitude and phase information of I/Q components at the same time.

4. Results and Discussions

To evaluate and examine the performance of the 2D-CNN, 1D-CNN, 1D-DNN, and VNE, the three neural network models in the experiment utilized identical parameter configurations and employed the same training and testing datasets. Table 2 presents the pertinent parameter values.
The training approach used 100 epochs and a batch size of 512. These parameters were established through numerous trials, ensuring both training efficacy and enhanced training efficiency. The network used a two-layer design: the first layer had a 3 × 2 convolutional kernel (time span × real-imaginary channels), where the three-symbol coverage helped track the phase consistency between nearby symbols, and the two channels referred to the in-phase (I) and quadrature-phase (Q) parts of the QPSK signal. This setup allowed for simultaneous modeling of how the signal changed over time and the complex relationships between I/Q components. The second layer had a smaller 3 × 1 kernel, which kept the focus on capturing important time-related features while also looking for more complex patterns. This design allowed the network to simultaneously understand how the signal changes over time and how the I/Q components were related to each other in a complex way. The second-layer kernel was reduced to 3 × 1, which kept the ability to extract time-related features while focusing on capturing more complex abstract features. The pooling layer implemented (2, 1) subsampling to reduce feature dimensions along the time axis, hence augmenting translation invariance while maintaining complete information in the I/Q channels. An L1 regularization term ( λ = 10 6 ) was added to the loss function to help prevent overfitting by encouraging the weight matrix to have fewer non-zero values. We included a Dropout layer at the network output with a probability of 0.2, which randomly deactivated neurons, thereby enhancing the ability to generalize. A dynamic early stopping mechanism (patience = 25) monitored validation loss and halted training when performance remained unchanged for 25 consecutive epochs, therefore preventing overfitting and optimizing computing resource use. We empirically confirmed that a batch size of 512 optimized CPU memory limitations and gradient update stability, facilitating reliable training convergence in intricate channel settings.
The design of the VNE detailed in this work specifies the tap count of the LMS linear filter as 15 and the order of the Volterra nonlinear filter as 13. The settings were adjusted through testing: the 13th-order Volterra structure found a good balance between handling nonlinearity and being efficient in calculations, as improvements levelled off and the risk of overfitting rose if we went beyond this order. The 15-tap version of the LMS filter adequately addressed the principal inter-symbol interference (ISI) range and reduced unnecessary calculations, as established by the autocorrelation analysis of the received signals. Dynamic step-size adjustment ( u L M S _ i n i t = 2 × 10 5 ,   u v o l t e r r a = 2 × 10 6 ) and a regularization parameter ( λ i n i t = 10 6 ) were used to make the model more stable and adaptable, helping to effectively correct for nonlinear distortions and phase noise in long-distance transmissions.
Figure 4 illustrates the BER curves with optical power into PD. The figure illustrates that, under varying optical power settings, the 2D-CNN exhibited the lowest overall BER and optimal performance. At an optical power of 8 dBm, the BER was 5.3 × 10−3, which was below the threshold for soft-decision forward error correction of 1.56 × 10−2 with a 15% overhead. It could precisely process communication signals and diminish the error rate. The BER of the 1D-CNN was marginally superior to that of the 2D-CNN, varying between 2.2 × 10−2 and 6.3 × 10−3. Due to its processing of one-dimensional data, it inadequately leveraged the correlation of two-dimensional spatial signals, leading to marginally diminished performance. The BER of the 1D-DNN ranged from 1.3 × 10−2 to 4.6 × 10−2. The structure was susceptible to overfitting, resulting in inadequate generalization capability and a comparatively elevated BER. The conventional VNE approach exhibited the greatest BER, ranging from 1.4 × 10−2 to 4.9 × 10−2. It possessed constraints in addressing intricate issues, and its efficacy was subpar compared to neural networks.
We looked at the constellation diagrams of the VNE, 1D-DNN, 1D-CNN, and 2D-CNN to better understand how well they communicate and process signals. The constellation diagram illustrates signal quality and interference resilience. Figure 5 shows that the traditional VNE’s constellation diagram had a lot of spread and unclear points, meaning it struggled to reduce nonlinear distortion and noise interference, resulting in poor signal quality and a higher BER. The 1D-DNN constellation diagram exhibited dispersion and indistinct points, indicating overfitting and insufficient adaptability. The 1D-CNN constellation points were spread out and not grouped together well, which shows that it struggled to pick out important features and can’t effectively handle interference from nonlinearity and phase noise, possibly leading to signal distortion and a higher BER. In contrast, the 2D-CNN constellation diagram exhibited closely clustered points with little dispersion, facilitating separation. The figure shows that the 2D-CNN could adjust phase settings using convolutional network weights to rebuild the phase with very few errors. The constellations focused on standard QPSK constellation points, minimizing nonlinear and phase noise interference. The more concentrated and prominent constellation points exhibited superior signal quality and resilience to interference.
To see how the number of trainable parameters affects the performance of 2D-CNN, we ran tests with four different amounts of trainable parameters: 100,000, 150,000, 200,000, and 350,000. During the tests, we kept everything the same, including the network design, activation function (ReLU), loss function (MSE), optimizer (AdamW), convolution kernel size, pooling method, and the datasets used for training and testing, with the only change being the number of trainable parameters. As shown in Figure 6, increasing the number of trainable parameters from 100,000 to 150,000 lowered the system’s BER from 5.9 × 10−3 to 5.6 × 10−3, which means that adding a moderate number of parameters helped the model learn complex signal details better and improved its ability to handle nonlinearity and phase noise. As the parameters were elevated to 200,000 and 350,000, the BER climbed marginally to 6.0 × 10−3 and 6.3 × 10−3, respectively, indicating a tendency of minor performance deterioration.
The training and validation loss curves at 8 dBm optical power for various parameter configurations in Figure 7a–d indicate that the losses for all four groups consistently decreased with the increase in epochs, while the disparity between training and validation losses remained minimal. The training and validation loss graphs for the four sets of parameters show that there was no sign of overfitting; all graphs show a steady decrease during training, with the final values staying close together and no major splits observed. This result was mostly due to the weight-decay regularization technique used by the AdamW optimizer, which substantially reduced the danger of overfitting as the parameter scale expanded. When the quantity of trainable parameters surpassed 150,000, the decline in BER optimization performance was not due to overfitting; rather, it was likely a consequence of redundant feature representation or the introduction of extraneous noise from an excess of trainable parameters, which increased the model’s sensitivity to dynamic channel fluctuations. Experimental results showed that the 150,000-parameter setup struck the best balance between being complex enough and being able to generalize well: it reduced the chance of underfitting by effectively extracting features (which was shown by the higher BER in the 100,000-parameter setup) while avoiding issues of wasted computing power and increased noise that come with having too many parameters.
Figure 8 compares CNN BER with different activation functions. The figure shows that the 2D-CNN with ReLU activation function had the lowest BER of 5.5 × 10−3, suggesting higher performance. The main reasons ReLU works are ReLU has an easy mathematical expression. This function saves computational power and speeds model training. Second, the ReLU function handles gradient vanishing well. Vanishing gradients hamper deep neural network model training. When x > 0, the ReLU function gradient was 1, avoiding gradient vanishing and improving model convergence. The Sigmoid function outputted 0–1. The gradient approached zero when the input value was extremely high or low, increasing BER and preventing parameter updates. The Tanh function reduced gradient vanishing, but its output was limited to −1 and 1; thus, minor gradients still affected training. Mish had a greater BER than ReLU since its properties did not match the 2D-CNN model’s tasks, leading to lower performance. Therefore, in this experiment, the ReLU function outperformed Mish as an activation function.
Figure 9 illustrates the BER curves of the 2D-CNN using MSE and cross-entropy (CE) loss functions. We regulated variables and maintained other factors at a steady level. Utilizing MSE as the loss function, the 2D-CNN achieved a minimal BER of 5.9 × 10−3, signifying outstanding performance. Regression problems were addressed with MSE. The signal-processing environment of this experiment rendered signal prediction a regression problem. The MSE quantified deviation by averaging the squared discrepancies between expected and actual values. BER was significantly reduced during training since it emphasized absolute error and systematically adjusted parameters to align predicted values with actual values. The cross-entropy loss function measured the discrepancies across probability distributions in classification tasks. The regression of communication signals employed an error assessment method that failed to satisfy operational requirements, resulting in an increased BER. In short, the MSE loss function improved how well the model worked and reduced the BER in 2D-CNN communication signal processing.
We used the widely used optimizers AdamW, Adam, and SGD to evaluate their impact on the performance of 2D-CNNs. Figure 10 illustrates the BER comparisons across several optimizers. The AdamW optimizer trained the 2D-CNN, achieving a minimum BER of 5.9 × 10−3, signifying optimal performance. AdamW mitigated model overfitting with a weight-decay method derived from Adam. It may adjust the learning rate for each parameter and dynamically modify the update step size based on the gradient and historical gradients to expedite convergence to the optimal solution while maintaining stability. Adam, without a weight-decay mechanism, exhibited overfitting with intricate data and had a high BER. SGD employed a fixed learning rate and updated step magnitude. Large datasets and complex models made it hard to reach the best solution and could become stuck in less optimal points, leading to a higher BER.

5. Conclusions

This research has successfully demonstrated a long-distance wireless communication system using photonics-aided technology, allowing for a stable transmission of 50 Gbps (25 Gbaud) QPSK signals over 4600 m at a carrier frequency of 128.75 GHz. At an optical power of 8 dBm, the BER reached 5.3 × 10−3, which is lower than the threshold of 1.56 × 10−2 for SD-FEC with a 15% overhead. A method using a 2D-CNN which allows for the simultaneous correction of phase noise and nonlinear distortion was proposed to equalize nonlinear impairments and phase noise interference in the RoF system. Compared to traditional Volterra series and DNN methods, it has shown significant improvements, providing a new and effective way to fix signal problems in high-speed wireless communication systems. It offers essential technological assistance for the development of future 6G and higher-speed, high-capacity wireless communication networks.

Author Contributions

Conceptualization, Y.J.; methodology, S.X. and Q.W.; software, Y.J.; validation, Y.J. and S.X.; formal analysis, Y.J. and J.Z.; investigation, Y.J. and J.G.; resources, J.L. and Z.O.; data curation, Y.M. and S.W.; writing—original draft preparation, Y.J.; writing—review and editing, S.X.; visualization, Y.J.; supervision, S.X.; project administration, S.X. and W.Z.; funding acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

National Key R&D Program of China (2023YFB2905600), National Natural Science Foundation of China (No. 62127802, 62331004, 62305067, U24B20142, U24B20168, and 62427815), and Key Project of Jiangsu Province of China (No. BE2023001-4).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw/processed data required to reproduce these findings cannot be shared at this time as the data also form part of an ongoing study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hamada, H.; Nosaka, H.; Tsutsumi, T.; Sugiyama, H.; Matsuzaki, H.; Song, H.-J.; Itami, G.; Fujimura, T.; Abdo, I.; Okada, K. Millimeter-wave InP device technologies for ultra-high speed wireless communications toward beyond 5G. In Proceedings of the 2019 IEEE International Electron Devices Meeting (IEDM), San Francisco, CA, USA, 7–11 December 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  2. Strinati, E.C.; Barbarossa, S.; Gonzalez-Jimenez, J.L.; Ktenas, D.; Cassiau, N.; Maret, L.; Dehos, C. 6G: The next frontier: From holographic messaging to artificial intelligence using subterahertz and visible light communication. IEEE Veh. Technol. Mag. 2019, 14, 42–50. [Google Scholar] [CrossRef]
  3. Barrera, F.; Siligaris, A.; Blampey, B.; Gonzalez-Jimenez, J.L. A D-band 4-ways power splitter/combiner implemented on a 28nm bulk CMOS process. In Proceedings of the 2019 49th European Microwave Conference (EuMC), Paris, France, 1–3 October 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  4. Song, H.J.; Lee, N. Terahertz communications: Challenges in the next decade. IEEE Trans. Terahertz Sci. Technol. 2021, 12, 105–117. [Google Scholar] [CrossRef]
  5. Alavi, S.E.; Soltanian, M.R.K.; Amiri, I.S.; Khalily, M.; Supa’at, A.S.M.; Ahmad, H. Towards 5G: A photonic based millimeter wave signal generation for applying in 5G access fronthaul. Sci. Rep. 2016, 6, 19891. [Google Scholar] [CrossRef]
  6. Yang, Y.; Lim, C.; Nirmalathas, A. Investigation on transport schemes for efficient high-frequency broadband OFDM transmission in fibre-wireless links. J. Light. Technol. 2013, 32, 267–274. [Google Scholar] [CrossRef]
  7. Zhou, W.; Qin, C. Simultaneous generation of 40, 80 and 120 GHz optical millimeter-wave from one Mach-Zehnder modulator and demonstration of millimeter-wave transmission and down-conversion. Opt. Commun. 2017, 398, 101–106. [Google Scholar] [CrossRef]
  8. Xiao, J.; Yu, J.; Li, X.; Xu, Y.; Zhang, Z.; Chen, L. 40-Gb/s PDM-QPSK signal transmission over 160-m wireless distance at W-band. Opt. Lett. 2015, 40, 998–1001. [Google Scholar] [CrossRef]
  9. Zhou, W.; Zhao, L.; Zhang, J.; Wang, K. Four sub-channel single sideband generation of vector mm-wave based on an I/Q modulator. IEEE Photonics J. 2019, 11, 1–9. [Google Scholar] [CrossRef]
  10. Dat, P.T.; Kanno, A.; Yamamoto, N.; Kawanishi, T. 190-Gb/s CPRI-equivalent rate fiber-wireless mobile fronthaul for simultaneous transmission of LTE-A and F-OFDM signals. In Proceedings of the ECOC 2016; 42nd European Conference on Optical Communication, Dusseldorf, Germany, 18–22 September 2016; VDE: Berlin, Germany, 2016. [Google Scholar]
  11. Kanno, A.; Dat, P.T.; Kuri, T.; Hosako, I.; Kawanishi, T.; Yoshida, Y.; Yasumura, Y.; Kitayama, K. Coherent radio-over-fiber and millimeter-wave radio seamless transmission system for resilient access networks. IEEE Photonics J. 2012, 4, 2196–2204. [Google Scholar] [CrossRef]
  12. Asif Khan, M.K.; Ali, F.; Irfan, M.; Muhammad, F.; Althobiani, F.; Ali, A.; Khan, S.; Rahman, S.; Perun, G.; Glowacz, A. Mitigation of phase noise and nonlinearities for high capacity radio-over-fiber links. Electronics 2021, 10, 345. [Google Scholar] [CrossRef]
  13. Fan, Q.; Lu, C.; Lau, A.P.T. Combined Neural Network and Adaptive DSP Training for Long-Haul Optical Communications. J. Light. Technol. 2021, 39, 7083–7091. [Google Scholar] [CrossRef]
  14. Zhang, S.; Yaman, F.; Nakamura, K.; Inoue, T.; Kamalov, V.; Jovanovski, L.; Vusirikala, V.; Mateo, E.; Inada, Y.; Wang, T. Field and lab experimental demonstration of nonlinear impairment compensation using neural networks. Nat. Commun. 2019, 10, 3033. [Google Scholar] [CrossRef]
  15. Ghose, S.; Singh, N.; Singh, P. Image denoising using deep learning: Convolutional neural network. In Proceedings of the 2020 10th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, 29–31 January 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  16. Shi, J.; Huang, A.P.; Tao, L.W. Deep Learning Aided Channel Estimation and Signal Detection for Underwater Optical Communication. Chin. J. Lasers 2022, 49, 1706004. [Google Scholar]
  17. Zhixuan, C.; Hongbo, Z.; Min, Z.; Ju, C.; Jiao, L.; Jie, D.; Qianwu, Z. Fiber Nonlinear Impairments Compensation Algorithm Based on CNN-BiLSTM-Attention Model. Electr. Power Inf. Commun. Technol. 2023, 21, 7–12. [Google Scholar]
  18. Wang, K.; Wang, C.; Li, W.; Wang, Y.; Ding, J.; Liu, C.; Kong, M.; Wang, F.; Zhou, W.; Zhao, F.; et al. Complex-valued 2D-CNN equalization for OFDM signals in a photonics-aided MMW communication system at the D-band. J. Light. Technol. 2022, 40, 2791–2798. [Google Scholar] [CrossRef]
  19. Wei, Y.; Yu, J.; Zhao, X.; Yang, X.; Wang, M.; Li, W.; Tian, P.; Han, Y.; Zhang, Q.; Tan, J.; et al. Demonstration of a photonics-aided 4,600-m wireless transmission system in the sub-THz band. J. Light. Technol. 2024, 42, 8564–8576. [Google Scholar] [CrossRef]
  20. Lu, X.; Yu, J.; Li, W.; Zhao, X.; Zhu, M.; Zhang, J.; Yang, W.; Wang, M.; Wei, Y.; Zhang, Q.; et al. Realizing 10 Gbps Frequency Continuously Tunable D-Band Signal Transmission With 20 Kilometers Wireless Distance Based on Photon Assistance. J. Light. Technol. 2025, 43, 5709–5717. [Google Scholar] [CrossRef]
  21. Wang, C.; Wang, K.; Tan, Y.; Wang, F.; Sang, B.; Li, W.; Zhou, W.; Yu, J. High-speed terahertz band radio-over-fiber system using hybrid time-frequency domain equalization. IEEE Photonics Technol. Lett. 2022, 34, 559–562. [Google Scholar] [CrossRef]
  22. Torkaman, P.; Latifi, S.M.; Feng, K.-M.; Yang, S.-H. Nonlinearity-robust im/dd thz communication system via two-stage deep learning equalizer. IEEE Commun. Lett. 2024, 28, 1805–1809. [Google Scholar] [CrossRef]
  23. Zhang, H.; Zhang, L.; Yang, Z.; Yang, H.; Lü, Z.; Pang, X.; Ozolins, O.; Yu, X. Single-lane 200 Gbit/s photonic wireless transmission of multicarrier 64-QAM signals at 300 GHz over 30 m. Chin. Opt. Lett. 2023, 21, 023901. [Google Scholar] [CrossRef]
  24. Leshem, A.; Yemini, M. Phase noise compensation for OFDM systems. IEEE Trans. Signal Process. 2017, 65, 5675–5686. [Google Scholar] [CrossRef]
  25. Chung, M.; Liu, L.; Edfors, O.; Sheikh, F. Phase-noise compensation for OFDM systems exploiting coherence bandwidth: Modeling, algorithms, and analysis. IEEE Trans. Wirel. Commun. 2021, 21, 3040–3056. [Google Scholar] [CrossRef]
  26. Mei, R.; Wang, Z.; Hu, W. Robust blind equalization algorithm using convolutional neural network. IEEE Signal Process. Lett. 2022, 29, 1569–1573. [Google Scholar] [CrossRef]
  27. Chang, Z.; Wang, Y.; Li, H.; Wang, Z. Complex CNN-based equalization for communication signal. In Proceedings of the 2019 IEEE 4th International Conference on Signal and Image Processing (ICSIP), Wuxi, China, 19–21 July 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  28. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  29. Yongshi, W.; Jie, G.; Hao, L.; Li, L.; Zhigang, W.; Houjun, W. CNN-based modulation classification in the complicated communication channel. In Proceedings of the 2017 13th IEEE International Conference on Electronic Measurement & Instruments (ICEMI), Yangzhou, China, 20–22 October 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
  30. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 6999–7019. [Google Scholar] [CrossRef] [PubMed]
  31. Chatterjee, A.; Gupta, U.; Chinnakotla, M.K.; Srikanth, R.; Galley, M.; Agrawal, P. Understanding emotions in text using deep learning and big data. Comput. Hum. Behav. 2019, 93, 309–317. [Google Scholar] [CrossRef]
  32. Rusk, N. Deep learning. Nat. Methods 2016, 13, 35. [Google Scholar] [CrossRef]
  33. Misra, D. Mish: A Self Regularized Non-Monotonic Activation Function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  34. Hinton, G.E.; Srivastava, N.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R.R. Improving neural networks by preventing co-adaptation of feature detectors. arXiv 2012, arXiv:1207.0580. [Google Scholar]
  35. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
Figure 1. Two-dimensional structure of the input data.
Figure 1. Two-dimensional structure of the input data.
Sensors 25 03661 g001
Figure 2. Schematic flowchart of 2D-CNN.
Figure 2. Schematic flowchart of 2D-CNN.
Sensors 25 03661 g002
Figure 3. The experimental setup: (a) single-channel 4600 m wireless testing configuration, (b) transmitter-side and receiver-side DSP.
Figure 3. The experimental setup: (a) single-channel 4600 m wireless testing configuration, (b) transmitter-side and receiver-side DSP.
Sensors 25 03661 g003
Figure 4. Curve of the BER performance for 2D-CNN, 1D-CNN, 1D-DNN, and VNE.
Figure 4. Curve of the BER performance for 2D-CNN, 1D-CNN, 1D-DNN, and VNE.
Sensors 25 03661 g004
Figure 5. Comparison of signal constellation diagrams for various neural networks and VNE (optical power: 8 dBm).
Figure 5. Comparison of signal constellation diagrams for various neural networks and VNE (optical power: 8 dBm).
Sensors 25 03661 g005
Figure 6. BER performance curve of 2D-CNN with varying quantities of trainable parameters.
Figure 6. BER performance curve of 2D-CNN with varying quantities of trainable parameters.
Sensors 25 03661 g006
Figure 7. Training and validation loss curves for different numbers of trainable parameters at 8 dBm. (a) 100,000 parameters; (b) 150,000 parameters; (c) 200,000 parameters; (d) 350,000 parameters.
Figure 7. Training and validation loss curves for different numbers of trainable parameters at 8 dBm. (a) 100,000 parameters; (b) 150,000 parameters; (c) 200,000 parameters; (d) 350,000 parameters.
Sensors 25 03661 g007
Figure 8. Impact of various activation functions on the BER of 2D-CNN, with insets of recovered constellation diagrams for (i) Sigmoid, (ii) Tanh, (iii) Mish, and (iv) ReLU with an optical power of 8 dBm.
Figure 8. Impact of various activation functions on the BER of 2D-CNN, with insets of recovered constellation diagrams for (i) Sigmoid, (ii) Tanh, (iii) Mish, and (iv) ReLU with an optical power of 8 dBm.
Sensors 25 03661 g008
Figure 9. The BER performance of the 2D-CNN utilizing various loss functions.
Figure 9. The BER performance of the 2D-CNN utilizing various loss functions.
Sensors 25 03661 g009
Figure 10. The impact of various optimizers on the BER of 2D-CNN.
Figure 10. The impact of various optimizers on the BER of 2D-CNN.
Sensors 25 03661 g010
Table 1. Terahertz signal transmission performance with various equalizers and modulations.
Table 1. Terahertz signal transmission performance with various equalizers and modulations.
NumberFrequencyEqualizerModulation FormatDistanceBER
[18]120 GHz2D-CNN16-QAM200 m≤3.8 × 10−3
[19]125/135/145 GHzVolterra equalizerPDM-16-QAM4600 m≤2.4 × 10−2
[20]110–170 GHzLSTM + RFQPSK20 km<3.8 × 10−3
[21]340 GHz2D-CNN16-QAM54.6 m≤10−2
[22]125 GHzLSTM + RF16-QAM4.5 m5.64 × 10−4
[23]286/299/312 GHzMMA + Viterbi64-QAM30 m<4.5 × 10−3
Table 2. Parameter values in 2D-CNN, 1D-CNN, and 1D-DNN.
Table 2. Parameter values in 2D-CNN, 1D-CNN, and 1D-DNN.
Term1D-DNN1D-CNN2D-CNN
Layers3 FCL2 Conv + 2 FCL2 Conv (3 × 2, 3 × 1) + 2 FCL
Kernel sizeNot used(3), (3)(3, 2), (3, 1)
ActivationReLUReLUReLU
Epoch100100100
Batch Size512512512
Loss functionMSELossMSELoss + L1MSELoss + L1
OptimizerAdamWAdamWAdamW
Pooling sizeNot used(2)(2, 1)
Total Parameters239,042139,802139,702
MACC236,160229,440229,440
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, Y.; Xu, S.; Wang, Q.; Zhang, J.; Ge, J.; Lin, J.; Ma, Y.; Wang, S.; Ou, Z.; Zhou, W. Demonstration of 50 Gbps Long-Haul D-Band Radio-over-Fiber System with 2D-Convolutional Neural Network Equalizer for Joint Phase Noise and Nonlinearity Mitigation. Sensors 2025, 25, 3661. https://doi.org/10.3390/s25123661

AMA Style

Jiang Y, Xu S, Wang Q, Zhang J, Ge J, Lin J, Ma Y, Wang S, Ou Z, Zhou W. Demonstration of 50 Gbps Long-Haul D-Band Radio-over-Fiber System with 2D-Convolutional Neural Network Equalizer for Joint Phase Noise and Nonlinearity Mitigation. Sensors. 2025; 25(12):3661. https://doi.org/10.3390/s25123661

Chicago/Turabian Style

Jiang, Yachen, Sicong Xu, Qihang Wang, Jie Zhang, Jingtao Ge, Jingwen Lin, Yuan Ma, Siqi Wang, Zhihang Ou, and Wen Zhou. 2025. "Demonstration of 50 Gbps Long-Haul D-Band Radio-over-Fiber System with 2D-Convolutional Neural Network Equalizer for Joint Phase Noise and Nonlinearity Mitigation" Sensors 25, no. 12: 3661. https://doi.org/10.3390/s25123661

APA Style

Jiang, Y., Xu, S., Wang, Q., Zhang, J., Ge, J., Lin, J., Ma, Y., Wang, S., Ou, Z., & Zhou, W. (2025). Demonstration of 50 Gbps Long-Haul D-Band Radio-over-Fiber System with 2D-Convolutional Neural Network Equalizer for Joint Phase Noise and Nonlinearity Mitigation. Sensors, 25(12), 3661. https://doi.org/10.3390/s25123661

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop