Next Article in Journal
Design of Microstrip Antenna Integrating 24 GHz and 77 GHz Compact High-Gain Arrays
Previous Article in Journal
Height Measurement Method for Meter-Wave Multiple Input Multiple Output Radar Based on Transmitted Signals and Receive Filter Design
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Neural Architecture for Real-Time Deep Learning Wavefront Sensing

1
Free-Space Optical Communication Technology Research Center, Harbin Institute of Technology, Harbin 150001, China
2
National Key Laboratory of Laser Spatial Information, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(2), 480; https://doi.org/10.3390/s25020480
Submission received: 18 December 2024 / Revised: 9 January 2025 / Accepted: 14 January 2025 / Published: 16 January 2025
(This article belongs to the Section Environmental Sensing)

Abstract

:
To achieve real-time deep learning wavefront sensing (DLWFS) of dynamic random wavefront distortions induced by atmospheric turbulence, this study proposes an enhanced wavefront sensing neural network (WFSNet) based on convolutional neural networks (CNN). We introduce a novel multi-objective neural architecture search (MNAS) method designed to attain Pareto optimality in terms of error and floating-point operations (FLOPs) for the WFSNet. Utilizing EfficientNet-B0 prototypes, we propose a WFSNet with enhanced neural architecture which significantly reduces computational costs by 80% while improving wavefront sensing accuracy by 22%. Indoor experiments substantiate this effectiveness. This study offers a novel approach to real-time DLWFS and proposes a potential solution for high-speed, cost-effective wavefront sensing in the adaptive optical systems of satellite-to-ground laser communication (SGLC) terminals.

1. Introduction

Over recent decades, satellite laser communication (SLC) has emerged as a focal point of research due to its numerous advantages, including high data rates, extended communication distances, improved confidentiality, excellent beam directivity, and a license-free spectrum [1,2]. Among the various SLC systems, the satellite-to-ground laser communication (SGLC) link is particularly significant. However, the communication channel is significantly degraded by atmospheric turbulence [3,4], for which adaptive optics (AO) serves as an effective solution for wavefront compensation of distortions [5]. Wavefront sensors, such as Shack–Hartmann wavefront sensors, are integral components of AO systems. However, these sensors also pose challenges due to their large size, high cost, substantial measurement losses at sub-aperture levels, and poor performance under conditions of strong scintillation [6,7].
The rapid advancement of artificial intelligence and the enhancement in computational unit performance have sparked a surge in interest in deep learning wavefront sensing (DLWFS). This method has become prominent due to its simplicity, speed, and high precision [8,9]. Currently, the primary implementation strategies can be divided into two categories: The first leverages deep neural networks (DNN) for iterative optimization of wavefront compensation, adhering to the physical model of beam propagation [10,11]. However, this method is hindered by time-consuming iterations during application. The second strategy employs DNNs to directly learn the mapping function between light signal intensity and wavefront from data, thereby circumventing the time-intensive iterative optimization process during application [12,13]. In the second implementation scheme, many current works focus on classic image classification convolutional neural network (CNN) prototypes, which modify the model for fitting Zernike mode coefficients of the wavefront [14,15,16]. While these approaches have proven effective, the parameter amounts of CNN prototypes are enormous (in the millions), resulting in longer inference times (5–20 ms). This consumes substantial hardware resources and power, increases servo latency, and poses challenges in adapting to real-time DLWFS in complex and dynamic environments of atmospheric turbulence.
This paper addresses the real-time sensing problem of dynamic random wavefront distortions caused by atmospheric turbulence by proposing a real-time description method based on CNNs. We aim to achieve low-cost, miniaturized, and fast real-time DLWFS through improvements to the neural architecture of the CNN-based wavefront sensing neural network (WFSNet). By improving and applying the multi-objective neural architecture search (MANS) methodology [17], we identified significant potential for improvement in classic CNN prototypes for wavefront sensing applications. A real-time DLWFS scheme for SGLC communication systems is proposed, establishing a DLWFS physical model and presenting a straightforward and effective automated design and optimization scheme for WFSNet. A statistical model for wavefront distortion caused by sensor noise and atmospheric turbulence is developed, providing a theoretical foundation for the effectiveness of real-time DLWFS simulations. Through training and structural optimization based on the classic EfficientNet-B0 prototype [18], we introduce an optimized WFSNet model: Efficient-WFSNet. Simulation results indicate that for the first 35 Zernike modes of distortion, the accuracy of wavefront sensing improved by 22%, while FLOPs and inference time were reduced by 80% and 95%, respectively. The robustness of the method was studied and validated under varying intensities of atmospheric turbulence and sensor noise. Further validation through indoor experiments demonstrates that this model demonstrates superior performance. To our knowledge, the optimized WFSNet achieves an inference time significantly below 1 ms (approximately 0.2 ms), representing an improvement of at least one order of magnitude compared to the levels of other current works [10,11,12,13,14,15,16]. This work provides a novel approach for real-time DLWFS and offers a new method for developing low-loss, miniaturized, and low-cost adaptive wavefront compensation in SGLC terminals, with advantages such as reduced computational load, lower hardware dependency, and cost-effectiveness, thus holding significant application value.

2. Principle of the Method

2.1. Configuration

A typical SGLC terminal receiver consists of two main subsystems: the pointing, acquisition, and tracking (PAT) subsystem and the communication subsystem. The primary function of the PAT subsystem is to establish and maintain a stable optical link. Subsequently, the communication subsystem receives the incident light beam affected by turbulence and converts it into electrical signals. Figure 1 illustrates the wavefront sensing scheme used for AO correction within the communication subsystem. The incident light beam is split into two beams by a beam splitter (BS). One beam is directed for wavefront sensing and passes through a receiving lens into a camera. The CNN model processes the spot input and outputs wavefront distortion measurement signals, which are then converted into voltage control signals. The piezoelectric ceramic elements of the DM receive voltage control signals from the driver to correct the wavefront on the pupil plane. The other beam, intended for communication, is coupled into an optical fiber for demodulation.
Our approach employs a single-shot measurement at the defocused plane. Research has shown that DLWFS exhibits superior performance on scattered, overexposed, and defocused spots compared to focused spots [13]. In the context of SGLC systems, implementing defocusing is straightforward, causes little energy loss, and does not impact the speed of sensing.

2.2. Problem Modeling

The ideal light field emitted by the transmitter propagates through the atmospheric channel, resulting in wavefront distortion at the receiver plane. The complex amplitude of incident beam at the receiver plane can be expressed as U ( x , y ) = A ( x , y ) exp [ j W ( x , y ) ] , where A ( x , y ) is the aperture function, equal to 1 in a certain region and 0 elsewhere. W ( x , y ) represents the wavefront distortion, which can be expressed as a weighted sum of a series of Zernike polynomials Z i ( x , y ) in the case of a circular aperture, i.e., W ( r ) = i = 1 a i Z i ( x , y ) . Under the Fresnel approximation, the intensity pattern I ( x , y ) captured by the camera after propagating through a lens (focal length f) with defocus f d is expressed as follows:
I ( x , y ) = A U ( x , y ) · exp [ j β ( x 2 + y 2 ) ] · exp j 2 π x u + y v d x d y 2 = A F U ( x , y ) · exp j β ( x 2 + y 2 ) 2
where β = π ( 1 / f + 1 / f d ) / λ is a factor related to defocus, which is equivalent to Zernike coefficient a 4 . A is a real constant, u , v = x / λ ( f + f d ) , y / λ ( f + f d ) are spatial frequencies, and λ is the wavelength. The principle of the DLWFS scheme can be described as follows: to find a target mapping h such that the intensity distribution I is mapped to the N-dimensional Zernike coefficient vector a N of the wavefront distortion. The error is denoted as E , and the optimization process can be formulated as follows:
min h E [ h ( I ) a N ]
In this scheme, a CNN is employed to represent the target mapping h, and the model parameters are optimized using the backpropagation algorithm.
Noll [19] has demonstrated, based on the Kolmogorov model, that atmospheric-induced wavefront distortions are concentrated in the lower-order Zernike modes. Research indicates that for single-mode fiber coupling, when the corrected Zernike polynomial terms exceed 35, the improvement in coupling efficiency becomes negligible [6]. The correction requirement can be further relaxed for multimode fibers with larger core diameters [20]. Therefore, in our study, we set N = 35 . However, in practical environments, the Zernike mode fitting of distorted wavefronts will include an infinite number of terms. Therefore, in the analysis of phase estimation interference, it is essential to consider the random distortions caused by Zernike modes greater than N.

2.3. Model Structure Design and Optimization Method

For the task of real-time sensing of dynamic random wavefront distortions caused by atmospheric turbulence, the architecture of the CNN model h is critical to its performance. Key factors include the number of convolutional layers, the kernel size, the number of filters, stride, activation functions, and other structural hyperparameters. These factors are interrelated and collectively determine the performance of the CNN model. Our optimization objective is to explore the structural hyperparameters to achieve an optimal balance between computational complexity and generalization performance, referred to as multi-objective neural architecture search (MNAS) [17]. Traditionally, three main implementation methods exist [21]: grid search, random search, and reinforcement learning approaches. However, these methods require substantial computational resources and time, making them difficult to adapt to different optical systems and noise environments. Therefore, we propose a streamlined yet effective method that leverages the inherent characteristics of CNNs to reduce the number of uncertain hyperparameters, while simultaneously utilizing local gradient information during the search process.
Research by Tan et al. [18] indicates that the connection between neurons in CNN is efficient when the resolution (size of input images and hidden layers), depth (number of hidden layers), and width (number of channels in each hidden layer) are scaled proportionally. Therefore, once the size of the input image is determined, the structural hyperparameters to be specified can be simplified to the depth d and the width scaling factor w. Additionally, when other important components are present within the CNN, corresponding structural hyperparameters can be incorporated. Once the hyperparameters are established, a grid search space will be defined, with each point Θ representing a structural choice, and the intervals δ determined by the user’s practical requirements. Typically, these hyperparameters are integers, making a discretized search space more meaningful. Based on this, the optimized value evaluation function L for the model can be expressed as follows:
min d , w , L ( d , w , ) = E v a l i d ( d , w , ) × C ( d , w , ) C 0 ξ
where E v a l i d represents the validation loss, C represents the complexity evaluation function, which calculates the model’s FLOPs, C 0 indicates the FLOPs of the starting point model, and ξ is an exponential weight factor. Given the substantial reduction in search space, we propose a gradient-assist grid search (GAGS) algorithm for multi-objects optimization, as presented in Algorithm 1.
In this algorithm, the starting point of the search can be determined empirically or from a small-scale initial search space. Once the model value evaluation function of all adjacent grid points at the search starting point is determined, the partial derivative at the starting point can be calculated, and the search will proceed along the direction of the partial derivative until a local minimum point is reached. The grid can be further refined based on practical requirements.
Algorithm 1: Gradient-assist grid search
Sensors 25 00480 i001

3. Simulation Modeling

3.1. Sensor Noise

Spot patterns obtained from the image sensor are contaminated by noise. To enhance the robustness of DLWFS, sensor noise is introduced into the simulated data. The total noise of the image sensor encompasses dark noise, photon noise, and readout noise, etc. When the optical intensity is weak, signal-independent noise (such as readout noise) takes domination with a small and constant noise standard deviation σ c . Conversely, at high optical intensities, signal-dependent noise (such as photon noise and dark noise) becomes predominant, exhibiting a proportionality to the γ power of the input pattern. Thus the total noise η of a typical image sensor pixel can be defined as a piecewise function of the pattern I and Gaussian noise N with zero mean and unit variance, modeled as [22]
η i j = σ c N i j , I i j B σ u I i j γ N i j , I i j > B
where B represents the threshold value of the normalized light intensity within the sensor’s dynamic range (approximately e 5 ), which depends on the image sensor. The incident light intensity distribution undergoes photoelectric conversion to obtain the spot image. During the linear phase of photoelectric conversion, the noise distribution can be considered consistent with this distribution. It is evident that in the stage where photoelectric conversion approaches saturation, the noise is suppressed.

3.2. Wavefront Distortion

In our simulation, a two-dimensional phase screen is utilized to emulate the wavefront aberration caused by turbulence along the propagation path. To ensure the inclusion of infinite-order Zernike modes, we employ the power spectrum inversion method [23], which entails random sampling of the power spectral density in the complex domain, subsequently followed by a Fourier inverse transformation to generate the phase screen. Typically, additional sampling of low-frequency components is performed to supplement components with periods exceeding the screen size. For the modified von Karman model, the power spectral density of vertical phase fluctuations is given by [24]
Φ w ( κ ) = 0.023 r 0 5 / 3 exp ( ( κ / κ m ) 2 ) ( κ 2 + κ 0 2 ) 11 / 6 , κ 0 1 L 0 , κ m = 5.92 l 0
where r 0 is the Fried parameter. In the plane wave approximation we have
r 0 = 0.42 k 2 sec ζ 0 L C n 2 ( z ) d z 3 / 5
where ζ is the zenith angle of optical link, C n 2 ( z ) represents the distribution of the refractive index structure constant along the propagation path z, k is the wave vector, and L is the path length. L 0 is the outer scale of turbulence, and l 0 is the inner scale of turbulence.

3.3. Simulation Setting

The simulated dataset is used to train and evaluate the CNN model. Turbulence induced wavefront distortion are simulated by the method described in Section 3.2. Because the receiving aperture D of the satellite and ground terminals is generally about 0.5 m [25,26], this value is used in the simulations. The ratio D / r 0 is used to characterize turbulence strength, and nine representative values are selected to describe weak to strong turbulence conditions: 1, 2, 3, 4, 5, 6, 8, 10, 15. In this context, the value of r 0 is chosen relatively uniformly from 3 cm to 50 cm. The corresponding Zernike coefficients of the wavefront are obtained by least squares (LS) fitting. The spot image reflects the intensity distribution of the optical field after propagating through the receiving lens, with a defocus factor set at β = 6 . The size of phase screen and spot image are set to 128 × 128 and 64 × 64, respectively. The total sensor noise is added according to the noise model described in Section 3.1 ( σ c = 0.002 , σ u = 0.02 , γ = 0.5 ). Finally, to obtain the input of CNN, spot data are normalized to [0, 1]. Python libraries NumPy and Aotools [27] are used for scientific calculations and turbulence simulations.The total number of datasets is 135,000, with an average signal-to-noise ratio (SNR) of 25.6 dB, which is divided into training, validation, and test sets in proportions of 8:1:1.

4. Validation of the Method

4.1. Model Optimization: EfficientNet Prototype

We select EfficientNet-B0 [18] as a starting point for our design. The whole WFSNet is structured into three distinct phases: channel expansion, depthwise convolution, and channel compression. Figure 2 presents a schematic representation of the mobile inverted bottleneck convolution (MBConv) block [28] within the model. Compared to traditional convolutions, the dense connections between channels are decomposed into pointwise convolutions and single-channel spatial convolutions, which significantly enhances connectivity efficiency. Additionally, the squeeze-and-excitation (SE) layer [29] incorporates a channel attention mechanism at an exceptionally low computational expense. During the squeeze phase, channel features are extracted by global average pooling, while output channel weights are determined through excitation. Under this circumstance, an additional hyperparameter depthwise convolution expansion ratio e is introduced, alongside depth d and width scaling ratio w.
Next, we determine the settings for the search grid based on our specific requirements. Note that EfficientNet-B0 is suitable for processing 224 × 224 image inputs, while empirically the range of CCD spot spread is often narrower. For an input image size of 64 × 64, after scaling the model as suggested in the literature, the starting grid point for the search is set to be h ( 3 , 2.35 , 3 ) , which results in a small model containing only three MBConv blocks. This scaled model is referred to as Base-WFSNet. It is crucial to emphasize that the scaling ratios proposed by the authors may not always represent the optimal selection, given that the optimization of hyperparameters is contingent upon the specific dataset. Therefore, we will further search for better model architectures.
For a single step of searching, it is preferable for the change in structural hyperparameters to be reasonable. Considering our prototype after scaling, the minimum number of channels is 16 / 2.35 7 . Therefore, setting the interval for w at 0.35 (which is 1/7 of 2.35) generally meets the requirements. The intervals for d and e can be set to integers of 1 based on their definitions. Thus, we have established the hyperparameter search intervals δ = ( 1 , 0.35 , 1 ) . To expand the search range, a basic search space is chosen before the search begins: (d: 3, 4); (w: 2.7, 2.35, 2); (e: 2, 3, 4). ξ is set to −0.7.
During the training procedure, we selected the Huber function as loss function defined as follows:
H u b e r ( x , ϵ ) = 0.5 · x 2 , | x | ϵ 0.5 · x 2 + ϵ · ( | x | ϵ ) , | x | > ϵ
where x denotes the error between the model output vector and the true Zernike coefficient vector. It can be seen that the Huber function is actually a combination of mean square error (MSE) and mean absolute error (MAE). Note that MSE loss function expedites the training speed, while the MAE loss function ensures stability in the presence of outliers. The Huber function tends toward MSE for minor errors and MAE for major errors, thereby combining the benefits of both. The constant h determines the range designated as outliers, which is set to 1 during our training procedure.
The introduction of severe noises exacerbates the instability of convergence. Following a series of trials, we opted for the Adam optimizer and set the batch size to a relatively large number as 80. Additionally, we introduced a learning rate schedule that decreases linearly from 0.001 to 0.00001 over a span of 300 epochs. The aforementioned measures effectively increase the stability and accuracy of weight updates. All networks were constructed using the TensorFlow 2.9 framework in Python 3.9 and trained on a personal computer equipped with an AMD Ryzen 7 5800H CPU, NVIDIA GeForce RTX 3060 GPU, and 16 GB RAM. TensorFlow’s Lite mobile library is utilized for inference acceleration.
Next, with the GAGS algorithm, the search begin in the initial grid space and will further descend to the stationary point under the guidance of the gradient direction. The start point, optimal point, and six partial derivatives computed throughout the entire search process are detailed in Table 1. The search terminates at h ( 6 , 2 , 4 ) , which has global minimum L value among the evaluated WFSNets, with confirmaton as the local minimum point. Figure 3a presents the L-function values of all evaluated architectures. To visualize the performance, a two-dimensional comparison of validation losses and FLOPs is depicted in Figure 3b. Remarkably, among the architectures with minimal validation loss, h ( 6 , 2 , 4 ) exhibits the lowest FLOPs. This observation strongly supports the rationality of the objective function L. Consequently, h ( 6 , 2 , 4 ) is designated as Efficient-WFSNet to highlight its balance between accuracy and speed. Its detailed structural information is illustrated in Figure 4. Two samples demonstrating the model inference results are shown in Figure 5.
The performance of Efficient-WFSNet, Base-WFSNet, and adapted EfficientNet-B0 is verified using the preserved test set, as shown in Table 2. In this context, the training loss, validation loss, and test loss refer to the Huber loss of the model’s inferred first 35 Zernike mode coefficients compared to the true values from the simulated training set, validation set, and test set. We present these losses to evaluate the model’s regression capability in mapping the captured light spots to the 35 Zernike modes of the wavefront at the pupil plane. Since the loss function only captures the errors of the first 35 Zernike coefficients, the root mean square (RMS) error of wavefront is provided to indicate total wavefront sensing errors. Based on the test outcomes, our proposed Efficient-WFSNet shows a 22% improvement in turbulence-induced wavefront sensing accuracy over 35 Zernike modes compared to the Adapted EfficientNet-B0, while it reduces FLOPs and inference time by approximately 80% and 95%, respectively.

4.2. Robustness of the Method

In this subsection, we investigate the robustness of real-time DLWFS from two perspectives: turbulence intensity and sensor noise intensity. Figure 6a illustrates the performance variations of Efficient-WFSNet, Base-WFSNet, and adapted EfficientNet-B0 under different turbulence intensities. The red dotted and dashed line indicates the average fitting error of the 35 Zernike modes. Since our model outputs only the Zernike coefficients, using the average fitting error as a benchmark is more meaningful for a reasonable assessment of model performance. As turbulence intensity increases, the model’s inference error gradually rises but does not significantly deviate from the benchmark. At strong turbulence conditions ( D / r 0 = 15 ), the average RMS wavefront sensing error of Efficient-WFSNet reaches 0.882 rad (approximately 0.14 λ ), which is slightly higher than the benchmark of 0.119 rad (approximately 0.019 λ ).
Figure 6b presents the performance of the three models as a function of sensor noise intensity (from the model in Equation (4)). Here, we maintained γ = 0.5 and scaled σ u and σ c by the same proportion to achieve a SNR variation from 15 dB to 40 dB. It is evident that when the SNR exceeds 20 dB, the model’s inference error remains stable. At this stage, the average RMS wavefront sensing error of Efficient-WFSNet over D / r 0 = 1 to 15 remains below the Marechal criterion ( 1 / 14 λ ). This indicates that, for moderate or worse image quality, our method can achieve stable wavefront sensing. However, in the presence of severe noise, the sensing performance deteriorates rapidly due to DLWFS’s reliance on the detailed information of the spot patterns. At SNR = 13.9 dB, the average RMS error of Efficient-WFSNet reaches 0.551 rad (approximately 0.88 λ ). Furthermore, it can be observed that when the SNR falls below 15 dB, the performance of Efficient-WFSNet becomes slightly inferior to that of adapted EfficientNet-B0. We attribute this to the fact that the parameter scale of Efficient-WFSNet is significantly smaller than that of adapted EfficientNet-B0 (approximately 1 / 5 ), resulting in relatively more unstable connections among hidden features that are more susceptible to noise.

4.3. Experiment Validation

To further validate the feasibility of the wavefront sensing scheme, we conducted an indoor experiment. The setup is depicted in Figure 7. The 638 nm monochromatic laser beam was generated by a single-mode fiber-coupled source (Src, SFOLT FC-638-100-SM), and emitted to free space by a fiber collimator (FC). To mitigate scattering and aberrations, the beam was confined by an iris diaphragm (D). Following lightpath adjustments with a reflective mirror (M), the beam was modulate by a spatial light modulator (SLM, HOLOEYE LC-2500). A polarizer (P) was employed to adjust the polarization direction to the optimal angle for the liquid crystal on silicon (LCoS). Lens L1 and L2 expanded the beam to encompass the LCoS surface. Phase screens were projected onto the LCoS, and the distorted wavefront was received by objective lens L3. A commercial charge coupled device (CCD) was positioned behind the lens to capture the defocused spot.
When applying a data-based wavefront sensing scheme to a new setup, calibration is imperative. We generated a dataset containing of 2000 digital phase screens and executed transfer learning on pre-training models with spot images captured from the CCD. The majority of the model parameters are frozen, with only the fully connected layers in the head block undergoing optimization. Consequently, this transfer learning process can be succinctly described as follows:
min ω , b | | f M ω + b a N | | 2
where f M denotes n samples of M-dimensional features, a N signifies n samples of N order Zernike coefficients, ω represents an M × N matrix, and b corresponds to an N-dimensional offset. Here, M depends on the model architecture while N remains a constant at 35. This optimization problem can be efficiently resolved by LS method, requiring only a brief period of seconds on our hardware platform.
An indoor wavefront sensing experiment is conducted under moderate turbulence simulated by a SLM. Here, we utilized an infinite phase screen [30] to represent wavefront distortion. The pseudo wind speed v w is set at 15 m/s, and D / r 0 = 2 . Since the modulation speed of the SLM is merely 60 Hz, for the adapted EfficientNet-B0 model, an inference time of around 5 ms introduced a one-frame lag. The experiment results is demonstrated in Figure 8, with detailed information listed in the Table 3.

4.4. Disscusion

Generally, our methodology substantially improves existing CNN prototypes for real-time DLWFS. However, being a data-driven method, it does come with certain limitations. Optical system aberrations, defocus deviations, camera exposure time, and pixel size can introduce input distribution deviations, thereby impacting the performance of the pre-trained CNN. As a result, additional calibration training is required. We propose a simplified calibration approach: since aberrations and defocus deviations can be represented by a fixed bias in Zernike coefficients, we freeze the parameters of the convolutional layers and optimize only the linear transformation from features to outputs (i.e., the fully connected layer in the head block). Our experiments have confirmed the effectiveness of this approach, although overcoming this drawback entirely remains challenging.
Additionally, image quality is critical for sensing performance in image-based wavefront sensing methods. When the SNR of the light spot is low, intense noise distorts the intensity distribution, leading the CNN to capture incorrect information. To mitigate this issue, we introduced significant noise (with an SNR of approximately 25 dB) during model training, allowing the model to adapt to distortions caused by image sensor noise. In Section 4.2, we provide a demonstration, through a series of simulations at SNRs ranging from 15 to 40 dB, that our sensing scheme maintains stable performance when image quality is moderate or worse (greater than 20 dB). However, as image quality continues to deteriorate, sensing accuracy will inevitably decline rapidly.

5. Conclusions

In this paper, a real-time DLWFS method for dynamic stochastic wavefront distortion was proposed, which demonstrates the significant potential for improvement of the CNN prototype in DLWFS. We present an effective automated model structure design and optimization algorithm, GAGS, which led to the development of a compact network, Efficient-WFSNet, based on the EfficientNet-B0 prototype. Our method significantly overcomes the inherent problem of the great computational demand of deep CNNs while improving accuracy. To our knowledge, as a CNN-based method, Efficient-WFSNet achieves an unprecedented inference time of only 0.2 ms, which greatly reduces servo lag and thus reduces measurement error. Both simulated and experimental results show the superior performance of Efficient-WFSNet, an outcome primarily attributed to its network architecture, which is suitable for turbulence-induced wavefront sensing tasks. This method significantly overcomes the high computational demands associated with deep CNNs while enhancing the accuracy of wavefront reconstruction. Our research provides valuable insights for wide applications of real-time DLWFS in SGLC systems, offering advantages such as low cost, high precision, and low power consumption. It represents an intelligent, algorithm-based approach to real-time adaptive wavefront compensation, opening new avenues for optical wavefront correction, which holds significant importance.

Author Contributions

Conceptualization, J.L.; methodology, J.L. and Q.L.; software, J.L.; validation, J.L. and N.C.; formal analysis, Q.L.; investigation, J.L.; resources, L.T., J.M., Q.L. and J.L.; data curation, Q.L.; writing—original draft preparation, J.L.; writing—review and editing, L.T. and N.C.; visualization, N.C.; supervision, L.T. and J.M.; project administration, L.T. and J.M.; funding acquisition, L.T. and J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Khalighi, M.A.; Uysal, M. Survey on Free Space Optical Communication: A Communication Theory Perspective. IEEE Commun. Surv. Tutor. 2014, 16, 2231–2258. [Google Scholar] [CrossRef]
  2. Flannigan, L.; Yoell, L.; Xu, C.-q. Mid-wave and long-wave infrared transmitters and detectors for optical satellite communications—A review. J. Opt. 2022, 24, 043002. [Google Scholar] [CrossRef]
  3. Tyson, R.K. Bit-error rate for free-space adaptive optics laser communications. J. Opt. Soc. Am. A 2002, 19, 753–758. [Google Scholar] [CrossRef]
  4. Huang, J.; Deng, K.; Liu, C.; Zhang, P.; Jiang, D.; Yao, Z. Effectiveness of adaptive optics system in satellite-to-ground coherent optical communication. Opt. Express 2014, 22, 16000–16007. [Google Scholar] [CrossRef]
  5. Levine, B.M.; Martinsen, E.A.; Wirth, A.; Jankevics, A.; Toledo-Quinones, M.; Landers, F.; Bruno, T.L. Horizontal line-of-sight turbulence over near-ground paths and implications for adaptive optics corrections in laser communications. Appl. Opt. 1998, 37, 4553–4560. [Google Scholar] [CrossRef]
  6. Wang, Y.; Xu, H.; Li, D.; Wang, R.; Jin, C.; Yin, X.; Gao, S.; Mu, Q.; Xuan, L.; Cao, Z. Performance analysis of an adaptive optics system for free-space optics communication through atmospheric turbulence. Sci. Rep. 2018, 8, 1124. [Google Scholar] [CrossRef]
  7. Zhang, S.; Wang, R.; Wang, Y.; Mao, H.; Xu, G.; Cao, Z.; Yang, X.; Hu, W.; Li, X.; Xuan, L. Extending the detection and correction abilities of an adaptive optics system for free-space optical communication. Opt. Commun. 2021, 482, 126571. [Google Scholar] [CrossRef]
  8. Li, Z.; Li, X.; Gao, Z.; Jia, Q. Review of wavefront sensing technology in adaptive optics based on deep learning. High Power Laser Part. Beams 2021, 33, 081001. [Google Scholar] [CrossRef]
  9. Guo, Y.; Zhong, L.; Min, L.; Wang, J.; Wu, Y.; Chen, K.; Wei, K.; Rao, C. Adaptive optics based on machine learning: A review. Opto-Electron. Adv. 2022, 5, 200082. [Google Scholar] [CrossRef]
  10. Zhou, Z.; Zhang, J.; Fu, Q.; Nie, Y. Phase-diversity wavefront sensing enhanced by a Fourier-based neural network. Opt. Express 2022, 30, 34396–34410. [Google Scholar] [CrossRef]
  11. Li, S.; Wang, B.; Wang, X. Untrained physics-driven aberration retrieval network. Opt. Lett. 2024, 49, 4545–4548. [Google Scholar] [CrossRef] [PubMed]
  12. Paine, S.W.; Fienup, J.R. Machine learning for improved image-based wavefront sensing. Opt. Lett. 2018, 43, 1235–1238. [Google Scholar] [CrossRef] [PubMed]
  13. Nishizaki, Y.; Valdivia, M.; Horisaki, R.; Kitaguchi, K.; Saito, M.; Tanida, J.; Vera, E. Deep learning wavefront sensing. Opt. Express 2019, 27, 240–251. [Google Scholar] [CrossRef]
  14. Guo, H.; Xu, Y.; Li, Q.; Du, S.; He, D.; Wang, Q.; Huang, Y. Improved Machine Learning Approach for Wavefront Sensing. Sensors 2019, 19, 3533. [Google Scholar] [CrossRef]
  15. Wang, M.; Guo, W.; Yuan, X. Single-shot wavefront sensing with deep neural networks for free-space optical communications. Opt. Express 2021, 29, 3465–3478. [Google Scholar] [CrossRef]
  16. You, J.; Gu, J.; Du, Y.; Wan, M.; Xie, C.; Xiang, Z. Atmospheric Turbulence Aberration Correction Based on Deep Learning Wavefront Sensing. Sensors 2023, 23, 9159. [Google Scholar] [CrossRef]
  17. Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. MnasNet: Platform-Aware Neural Architecture Search for Mobile. arXiv 2019, arXiv:1807.11626. [Google Scholar]
  18. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2020, arXiv:1905.11946. [Google Scholar]
  19. Noll, R.J. Zernike polynomials and atmospheric turbulence. J. Opt. Soc. Am. 1976, 66, 207–211. [Google Scholar] [CrossRef]
  20. Ke, X. Spatial Optical-Fiber Coupling Technology in Optical-Wireless Communication; Springer: Singapore, 2023. [Google Scholar] [CrossRef]
  21. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; pp. 259–262. [Google Scholar]
  22. Faraji, H.; MacLean, W. CCD noise removal in digital images. IEEE Trans. Image Process. 2006, 15, 2676–2685. [Google Scholar] [CrossRef]
  23. Lane, R.G.; Glindemann, A.; Dainty, J.C. Simulation of a Kolmogorov phase screen. Waves Random Media 1992, 2, 209–224. [Google Scholar] [CrossRef]
  24. Andrews, L.C.; Phillips, R.L. Laser Beam Propagation Through Random Media; SPIE Press: Bellingham, WA, USA, 2005. [Google Scholar]
  25. Comeron, A.; Rubio, J.A.; Belmonte, A.M.; Garcia, E.; Prud’homme, T.; Sodnik, Z.; Connor, C. Propagation experiments in the near infrared along a 150-km path and from stars in the Canarian archipelago. Phys. Eng. Environ. Sci. 2002, 4678, 78–90. [Google Scholar] [CrossRef]
  26. Israel, D.J.; Edwards, B.L.; Staren, J.W. Laser Communications Relay Demonstration (LCRD) update and the path towards optical relay operations. In Proceedings of the 2017 IEEE Aerospace Conference, Big Sky, MT, USA, 4–11 March 2017; pp. 1–6. [Google Scholar] [CrossRef]
  27. Townson, M.; Farley, O.; Orban de Xivry, G.; Osborn, J.; Reeves, A. AOtools: A Python package for adaptive optics modelling and analysis. Opt. Express 2019, 27, 31316. [Google Scholar] [CrossRef] [PubMed]
  28. Sandler, M.; Howard, A.G.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  29. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. arXiv 2019, arXiv:1709.0150. [Google Scholar]
  30. Assémat, F.; Wilson, R.W.; Gendron, E. Method for simulating infinitely long and non stationary phase screens with optimized memory storage. Opt. Express 2006, 14, 988–999. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the DLWFS scheme for AO correction in communication subsystem.
Figure 1. Schematic diagram of the DLWFS scheme for AO correction in communication subsystem.
Sensors 25 00480 g001
Figure 2. Schematic diagram of WFSNet and MBConv block.
Figure 2. Schematic diagram of WFSNet and MBConv block.
Sensors 25 00480 g002
Figure 3. Comparison of network architectures performance. (a) Objective function L (Equation (3)) of WFSNets, where marker style indicates d, color indicates w, and x coordinate indicates e. The selected architecture is h ( 6 , 2 , 4 ) . (b) FLOPs and validation loss comparison among WFSNets, in addition to Adapted EfficientNet-B0.
Figure 3. Comparison of network architectures performance. (a) Objective function L (Equation (3)) of WFSNets, where marker style indicates d, color indicates w, and x coordinate indicates e. The selected architecture is h ( 6 , 2 , 4 ) . (b) FLOPs and validation loss comparison among WFSNets, in addition to Adapted EfficientNet-B0.
Sensors 25 00480 g003
Figure 4. Structural diagram of Efficient-WFSNet. The architecture of MBConv is detailed in Figure 2, where the notation ‘4’ signifies that the 1 × 1 point convolution expands the number of channels by a factor of 4. All convolutional layers are followed by batch normalization, and the activation function utilized in both convolutional and fully connected layers is the Swish function.
Figure 4. Structural diagram of Efficient-WFSNet. The architecture of MBConv is detailed in Figure 2, where the notation ‘4’ signifies that the 1 × 1 point convolution expands the number of channels by a factor of 4. All convolutional layers are followed by batch normalization, and the activation function utilized in both convolutional and fully connected layers is the Swish function.
Sensors 25 00480 g004
Figure 5. Two samples demonstrating the 35 Zernike mode inference performance of the prototype, Base-WFSNet, and Efficient-WFSNet. Phasescreens (a,d), normalized spots (b,e), and performance comparisons (c,f) are presented. Top row: D / r 0 = 1 . Low row: D / r 0 = 15 .
Figure 5. Two samples demonstrating the 35 Zernike mode inference performance of the prototype, Base-WFSNet, and Efficient-WFSNet. Phasescreens (a,d), normalized spots (b,e), and performance comparisons (c,f) are presented. Top row: D / r 0 = 1 . Low row: D / r 0 = 15 .
Sensors 25 00480 g005
Figure 6. The performance of the three models under different (a) turbulence intensity and (b) sensor noise intensity. The red dotted and dashed line indicates the average fitting error of the 35 Zernike modes. The seven pairs of values ( σ u , σ c ) in (b) are as follows: (0.002, 0.0002), (0.005, 0.0005), (0.01, 0.001), (0.015, 0.0015), (0.02, 0.002), (0.05, 0.005), (0.08, 0.008). The corresponding SNR ranges from 35.9 dB to 13.9 dB.
Figure 6. The performance of the three models under different (a) turbulence intensity and (b) sensor noise intensity. The red dotted and dashed line indicates the average fitting error of the 35 Zernike modes. The seven pairs of values ( σ u , σ c ) in (b) are as follows: (0.002, 0.0002), (0.005, 0.0005), (0.01, 0.001), (0.015, 0.0015), (0.02, 0.002), (0.05, 0.005), (0.08, 0.008). The corresponding SNR ranges from 35.9 dB to 13.9 dB.
Sensors 25 00480 g006
Figure 7. Experimental setup diagram: Src, 638 nm laser source; FC, fiber collimator; D, iris diaphragm; M, reflective mirror; L1, lens (diameter 20 mm, focal length 10 mm); L2, lens (diameter 20 mm, focal length 50 mm); L3, objective lens (diameter 5 mm); P, polarizer; SLM, spatial light modulator; CCD, charge coupled device; PC, personal computer.
Figure 7. Experimental setup diagram: Src, 638 nm laser source; FC, fiber collimator; D, iris diaphragm; M, reflective mirror; L1, lens (diameter 20 mm, focal length 10 mm); L2, lens (diameter 20 mm, focal length 50 mm); L3, objective lens (diameter 5 mm); P, polarizer; SLM, spatial light modulator; CCD, charge coupled device; PC, personal computer.
Sensors 25 00480 g007
Figure 8. RMS loss of 35 Zernike modes in continuous wavefront sensing experiment under simulated moderate turbulence.
Figure 8. RMS loss of 35 Zernike modes in continuous wavefront sensing experiment under simulated moderate turbulence.
Sensors 25 00480 g008
Table 1. Partial derivatives of grid neighborhood in the search process.
Table 1. Partial derivatives of grid neighborhood in the search process.
Step01234
Start point ( 3 , 2.35 , 3 ) ( 4 , 2 , 4 ) ( 5 , 1.65 , 5 ) ( 5 , 2 , 4 ) ( 6 , 2 , 4 )
L / d + ( × 10 4 )\−2.91−1.323.493.33
L / d ( × 10 4 )\−33.62.17−2.91−3.10
L / w + ( × 10 4 )\4.68−0.6200.1570.869
L / w ( × 10 4 )\0.763−9.82−3.49−2.74
L / e + ( × 10 4 )\−0.9052.670.3281.87
L / e ( × 10 4 )\−3.560.410−1.09−0.435
Adjacent best point ( 4 , 2 , 4 ) ( 5 , 2 , 4 ) ( 4 , 1.65 , 5 ) ( 6 , 2 , 4 ) ( 6 , 2 , 4 )
Adjacent best L ( × 10 2 )1.2231.1931.2111.1621.162
Table 2. Comparison of model performance, parameters, and inference speed.
Table 2. Comparison of model performance, parameters, and inference speed.
ParamsFlopsInference
Time
(ms)
Training
Loss
(rad)
Validation
Loss
(rad)
Test
Loss
(rad)
Test
RMS
(rad)
Adapted EfficientNet-B0324,64426.75 m5.3150.22610.22780.23070.4207
Base-WFSNet13,8692.77 m0.1220.26810.27130.27330.4430
Efficient-WFSNet67,7665.83 m0.2040.17560.17740.17930.3906
Table 3. Comparison of model performance with experiment data.
Table 3. Comparison of model performance with experiment data.
Training Loss (rad)Test Loss (rad)Test RMS (rad)
Adapted EfficientNet-B00.05100.13250.2179
Base-WFSNet0.06510.09910.2032
Efficient-WFSNet0.03990.05800.1918
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Liu, Q.; Tan, L.; Ma, J.; Chen, N. Enhanced Neural Architecture for Real-Time Deep Learning Wavefront Sensing. Sensors 2025, 25, 480. https://doi.org/10.3390/s25020480

AMA Style

Li J, Liu Q, Tan L, Ma J, Chen N. Enhanced Neural Architecture for Real-Time Deep Learning Wavefront Sensing. Sensors. 2025; 25(2):480. https://doi.org/10.3390/s25020480

Chicago/Turabian Style

Li, Jianyi, Qingfeng Liu, Liying Tan, Jing Ma, and Nanxing Chen. 2025. "Enhanced Neural Architecture for Real-Time Deep Learning Wavefront Sensing" Sensors 25, no. 2: 480. https://doi.org/10.3390/s25020480

APA Style

Li, J., Liu, Q., Tan, L., Ma, J., & Chen, N. (2025). Enhanced Neural Architecture for Real-Time Deep Learning Wavefront Sensing. Sensors, 25(2), 480. https://doi.org/10.3390/s25020480

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop