Next Article in Journal
Optical Multi-Frequency Discrimination and Phase Identification System Based on On-Chip Dual MZM
Previous Article in Journal
From PSNR to Frequency Evidence: Evaluating Super-Resolution Reliability on Low-SNR Fluorescence Channels
Previous Article in Special Issue
Multi-Spectral and Single-Shot Wavefront Detection Technique Based on Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wavefront Prediction for Adaptive Optics Without Wavefront Sensing Based on EfficientNetV2-S

1
School of Electronic Engineering, Jiangsu Ocean University, Lianyungang 222005, China
2
Shanghai Astronomical Observatory, Chinese Academy of Sciences, Shanghai 200030, China
3
School of Network & Telecom Engineering, Jinling Institute of Technology, Nanjing 211169, China
*
Author to whom correspondence should be addressed.
Photonics 2026, 13(2), 144; https://doi.org/10.3390/photonics13020144
Submission received: 5 December 2025 / Revised: 22 January 2026 / Accepted: 29 January 2026 / Published: 2 February 2026
(This article belongs to the Special Issue Adaptive Optics: Recent Technological Breakthroughs and Applications)

Abstract

Adaptive optics (AO) aims to counteract wavefront distortions caused by atmospheric turbulence and inherent system errors. Aberration recovery accuracy and computational speed play crucial roles in its correction capability. To address the issues of slow wavefront aberration detection speed and low measurement accuracy in current wavefront sensorless adaptive optics, this paper proposes a wavefront correction method based on the EfficientNetV2-S model. The method utilizes paired focal plane and defocused plane intensity images to directly extract intensity features and reconstruct phase information in a non-iterative manner. This approach enables the direct prediction of wavefront Zernike coefficients from the measured intensity images, specifically for orders 3 to 35, significantly enhancing the real-time correction capability of the AO system. Simulation results show that the root mean square error (RMSE) of the predicted Zernike coefficients for D/r0 values of 5, 10, and 15 are 0.038λ, 0.071λ, and 0.111λ, respectively, outperforming conventional convolutional neural network (CNN), ResNet50/101 and ConvNeXt-T models. The experimental results demonstrate that the EfficientNetV2-S model maintains good wavefront reconstruction and prediction capabilities at D/r0 = 5 and 10, highlighting its high precision and robust wavefront prediction ability. Compared to traditional iterative algorithms, the proposed method offers advantages such as high precision, fast computation, no need for iteration, and avoidance of local minima in processing wavefront aberrations.

1. Introduction

Adaptive optics (AO) improves imaging quality by measuring and compensating wavefront errors introduced by the propagation medium in real time. It has been widely used in ground-based telescopes, microscopy, ocular aberration correction, and laser communication [1,2,3,4,5]. A conventional AO system typically consists of a wavefront sensor, a controller, and a wavefront corrector: the sensor measures the distorted wavefront, the controller computes the control signals, and the corrector generates a conjugate phase to cancel aberrations. Despite its effectiveness, this architecture suffers from two major limitations. First, the wavefront-sensing branch usually requires an additional measurement optical path, increasing system complexity, alignment difficulty, and cost. Second, as AO applications expand to scenarios such as free-space optical communication, intracavity aberration correction, and extended-target imaging, the constraints of traditional sensor-based AO increasingly hinder efficiency and practicality. These challenges have motivated the development of wavefront sensorless (WFS-less) AO as an effective alternative [6]. WFS-less AO reconstructs the pupil phase from intensity information using algorithmic inversion and can be broadly categorized into single-image and phase-diversity approaches. Single-image methods infer wavefront errors from one intensity image; however, due to the inherent many-to-one ambiguity in intensity-to-phase mapping, they often exhibit limited accuracy and robustness. Phase-diversity methods use focused and defocused images to uniquely constrain the phase, thereby improving reconstruction fidelity [7]. Nevertheless, conventional WFS-less AO typically relies on iterative optimization and repeated measurements to reach an optimal correction, which reduces correction efficiency and limits deployment in time-critical applications.
In recent years, machine learning—especially data-driven deep learning—has attracted growing interest in AO [8,9]. Early work by Angel et al. [1] employed a three-layer feedforward network to map intensity distributions to Zernike coefficients, enabling piston and tilt correction in multi-telescope systems. Sandler and Barrett [10,11] further demonstrated real-star observations using a backpropagation neural network with focused/defocused image pairs, achieving Zernike coefficient estimation up to the 11th order. However, learning-based prediction under stronger turbulence and higher-order aberration regimes remains challenging, where the intensity patterns become more complex and subtle, and model generalization can degrade. Convolutional neural networks (CNNs) have a strong capacity for learning nonlinear input–output relationships and have achieved outstanding performance in many vision tasks, such as medical imaging [12], street-scene recognition [13], video recognition [14], vehicle detection [15] and microscopy [16]. In WFS-less AO, Nishizaki et al. [17] proposed a CNN to estimate the first 32 Zernike coefficients from a single intensity image. Ma et al. [18] introduced an AlexNet-based approach trained with focused/defocused image pairs and the corresponding coefficients. Yuan et al. [19] fine-tuned EfficientNet using simulated turbulence phase screens and predicted 66 coefficients from a single intensity image, and Yue et al. [20] developed an EfficientNet-B0-based sensorless prediction method using a single degraded focal-plane image. To enhance estimation accuracy, Ma et al. [21] proposed a GoogLeNet-based dual-frame recovery method, and Kok et al. [22] presented a ResNet50-based aberration characterization framework that can further refine predictions with a small number of iterations using phase-diversity images.
Although existing studies have demonstrated the feasibility of deep learning-based wavefront prediction for WFS-less AO, several limitations remain. First, under medium-to-strong turbulence conditions, where high-order aberrations dominate the wavefront distortion, many existing network backbones lack sufficient representational capability to reliably capture fine-grained, high-frequency distortion features from intensity measurements, resulting in increased prediction errors. Second, a number of reported approaches rely on multi-frame inputs and/or iterative refinement strategies, which impose additional acquisition and computational burdens and significantly constrain their applicability in high-speed, real-time AO systems. To address these challenges, this paper adopts EfficientNetV2-S as the backbone network for wavefront prediction in a WFS-less AO framework. A focused–defocused image pair is used as the network input to alleviate the inherent ambiguity of single-intensity measurements, while the corresponding Zernike coefficients are directly regressed in a single forward pass, thereby avoiding iterative optimization and enabling high-speed operation.
The selection of EfficientNetV2-S is motivated by both architectural and practical considerations specific to WFS-less AO. Compared with conventional CNN backbones, the EfficientNet family achieves superior accuracy–efficiency trade-offs through compound scaling of network depth and width, which is particularly advantageous for modeling turbulence-induced intensity distortions across multiple spatial scales. Compared with EfficientNetV1, EfficientNetV2 introduces Fused Mobile Inverted Bottleneck Convolution (Fused-MBConv) blocks that enhance low-level feature extraction and significantly improve training stability and efficiency, which is critical for learning high-order aberration representations from large-scale simulated datasets. Furthermore, the incorporation of squeeze-and-excitation (SE) attention strengthens channel-wise feature discrimination, enabling the network to better distinguish subtle intensity variations caused by complex atmospheric turbulence. Among the EfficientNetV2 variants, the S version provides a balanced compromise between representational capacity and computational complexity. While larger variants such as EfficientNetV2-M and -L offer increased model capacity, they incur substantially higher computational and memory costs, which are unnecessary for the present task and detrimental to real-time deployment. In contrast, EfficientNetV2-S retains sufficient modeling capability for high-order aberration prediction while maintaining a compact parameter size and low inference latency, making it well suited for deployment on embedded platforms such as NVIDIA Jetson. These characteristics render EfficientNetV2-S particularly appropriate for medium-to-strong turbulence conditions in WFS-less AO systems, where both robust nonlinear fitting and real-time performance are essential.
In this work, EfficientNetV2-S is compared with several representative architectures, including a conventional CNN, ResNet50/101, and ConvNeXt-T (Tiny), to evaluate wavefront prediction performance. Simulation and experimental results demonstrate that EfficientNetV2-S consistently achieves superior wavefront reconstruction accuracy and improved image restoration quality across different turbulence conditions, indicating its effectiveness and efficiency for deep learning-based wavefront prediction in WFS-less AO systems.

2. Materials and Methods

The schematic diagram of the WFS-less AO system based on deep learning is shown in Figure 1. The system mainly comprises a wavefront correction element (deformable mirror, DM), optical lenses, high-speed scientific cameras (CCD1 and CCD2), and a wavefront control module. Collimated light emitted from the source is distorted after propagating through atmospheric turbulence. The aberrated beam is then reflected by the DM and divided into two optical paths by a beam splitter. One beam is focused by a lens onto CCD1, where the focal-plane intensity distribution is recorded. The other beam, after an artificially introduced defocus, is incident on CCD2 to capture the intensity image at the defocused plane. The intensity images from the two CCDs are sent to the “Image collection” unit and combined as input data, which are then processed through a neural network to obtain the corresponding Zernike coefficients. The wavefront controller calculates the driving electrical signal for the DM based on these coefficients, and the signal returns to the DM via the solid line path to achieve closed-loop compensation of the wavefront aberration. In this way, the system can perform efficient wavefront correction using both in-focus and out-of-focus information without the need for a dedicated wavefront sensor.
Atmospheric turbulence phase screens are typically generated using two main approaches. The first is the power spectrum inversion method, which reconstructs the phase distribution from the turbulence power spectral density. The second is the Zernike polynomial expansion method, in which the phase is represented as a weighted sum of orthogonal Zernike basis functions. The power spectrum inversion approach primarily provides accurate representation of high-frequency components, while the Zernike polynomial expansion method is more widely adopted in practical applications for constructing atmospheric turbulence phase screens. This paper employs the Zernike polynomial expansion method to generate turbulence phase screens and constructs a numerical simulation model for WFS-less AO, thereby producing degraded target images and the corresponding Zernike aberration coefficients to form the training dataset for the network. Under atmospheric turbulence with different strengths, a randomly distorted wavefront φ ( x , y ) can be expressed as:
φ ( x , y ) = i = 1 M a i z i ( x , y )
In Equation (1), φ ( x , y ) represents the generated phase; a i is the coefficient of the i -th Zernike polynomial; z i ( x , y ) is the i -th Zernike polynomial; and M is the Zernike order. When an aberrated plane wave passes through a lens, it generates an incident initial light field W 0 ( x , y ) , which is imaged on CCD1 and CCD2 in the far field, obtaining the intensity distribution at the focal and defocused planes. By performing a Fourier transform on the incident light field W 0 ( x , y ) , the complex amplitude of the light field at the focal plane can be reconstructed, that is, the complex amplitude W 1 f x , f y of the light field at the focal plane can be expressed as:
W 1 f x , f y = Q 1 F W 0 ( x , y )
In Equation (2), Q 1 represents a complex constant, and F represents the Fourier transform. The complex amplitude of the light field W 2 f x , f y at the defocused plane can be expressed as:
W 2 f x , f y = Q 2 F W 0 ( x , y ) exp i φ d ( x , y )
In Equation (3), Q 2 represents a complex constant, and φ d ( x , y ) represents the defocus-induced phase term. The intensity at the focal and defocused planes is the squared magnitude of the complex optical field, as given in Equation (4):
I 1 ( x , y ) = W 1 f x , f y 2
I 2 ( x , y ) = W 2 f x , f y 2
The intensity images acquired at the focal and defocused planes contain information about the wavefront aberrations of the optical system. To retrieve these aberrations, a neural network is employed to extract the corresponding Zernike coefficients from the recorded images, thereby enabling reconstruction of the system wavefront. To achieve this, the neural network is trained offline using a large number of samples, where the input consists of intensity images captured on the CCD under static distorted wavefronts, and the output corresponds to the associated Zernike coefficients. The wavefront reconstruction process based on a neural network can be viewed as a nonlinear mapping: by inputting two intensity images, the network outputs the Zernike coefficients of the corresponding distorted wavefront. After training the model, a 1 , a 2 , , a M can be expressed as Equation (5), in which f NN ( ) represents the neural network model.
a 1 , a 2 , , a M = f NN I 1 ( x , y ) , I 2 x , y y
This paper proposes a WFS-less AO correction method based on the EfficientNetV2-S model. EfficientNetV2 is a next-generation CNN family proposed by Google in 2021 [23]. Through the use of neural architecture search (NAS) and compound scaling, it significantly improves training speed while maintaining parameter efficiency. EfficientNetV2-S, a lightweight variant in the EfficientNetV2 family, adopts compound scaling to jointly scale the network depth, width, and input resolution, thereby achieving a balanced trade-off between computational efficiency and model performance. Its overall architecture is shown in Figure 2 and Table 1.
EfficientNetV2 is organized into multiple stages, with each stage built upon the Fused-MBConv block as the fundamental unit. Compared with the original EfficientNet, EfficientNetV2 introduces three key improvements to enhance training efficiency and feature representation. First, the Fused-MBConv module, as illustrated in Figure 3, replaces the depthwise 3 × 3 convolution and the 1 × 1 expansion convolution in Mobile Inverted Bottleneck Convolution (MBConv) with a single standard 3 × 3 convolution, which simplifies the computation graph and better exploits the parallelism of GPUs/TPUs, thereby significantly improving training throughput with only a slight increase in FLOPs. Second, the SE attention module adaptively recalibrates channel-wise feature responses by computing channel weights in the shortcut branch and applying them to the main-branch features through element-wise multiplication, effectively enhancing channel dependencies with negligible parameter overhead. Third, Drop-Connect is adopted instead of conventional Dropout by randomly dropping hidden layer connections rather than outputs, which further alleviates overfitting without introducing additional inference cost. Owing to these advantages, the lightweight variant EfficientNetV2-S is selected in this work as the backbone network to achieve a favorable trade-off between computational efficiency and representation capability. The detailed configurations of the EfficientNetV2 family are summarized in Table 2.
To verify the performance of EfficientNetV2-S for wavefront prediction, several representative network architectures are selected for comparison, including a conventional CNN, ResNet50, ResNet101, and ConvNeXt-T. The structural characteristics and key configurations of these counterpart networks are described as follows. The conventional CNN adopted in this study is composed of an eight-layer architecture, as illustrated in Figure 4. Within the network, a “3 × 3” notation denotes the convolution kernel size, while “64” specifies the number of kernels employed in the corresponding convolutional layer. The stride parameter defines the step length of each convolution operation. To improve nonlinear representation capability, the ReLU activation function is applied after every layer. The output layer consists of 33 neurons, corresponding to the prediction of 33 Zernike coefficients from the 3rd to the 35th modes, excluding the piston and tip/tilt components.
Residual networks (ResNet) have developed from early 18-layer configurations to very deep architectures containing up to 152 layers. In principle, increasing network depth enables the extraction of more complex and abstract features, potentially enhancing model performance. However, excessive nonlinear transformations in deep networks may lead to progressive loss of feature information, resulting in performance degradation as depth increases. To overcome this limitation, ResNet introduces residual blocks that employ shortcut connections to implement identity mappings. These connections allow the network to preserve feature information better, enhance sensitivity to data variations, and facilitate the construction of robust mappings from image inputs to numerical outputs. Moreover, residual learning effectively alleviates the vanishing gradient problem associated with deep architectures, thereby preventing network degradation and improving prediction accuracy. The detailed ResNet architecture is illustrated in Figure 5. As the target output of this study consists of 33 Zernike coefficients, only the final fully connected layer is adjusted to produce 33 output nodes. The residual module is shown in Figure 6. Through the “shortcut connection”, the input is directly passed to the output H ( x ) = F ( x ) + x , resulting in H ( x ) = x   when F ( x ) = 0 . F ( x ) denotes the residual function, which represents the residual mapping to be learned [24]. Therefore, the training objective is to make the residual result close to 0, thus avoiding the accuracy degradation caused by deepening the network. Figure 7 depicts the specific building block for ResNet50/101.
ConvNeXt [25] is a novel and efficient CNN model proposed by the Facebook AI Research (FAIR) team in 2022. Compared with traditional convolutional networks, ConvNeXt significantly improves training and inference performance while maintaining structural efficiency, by incorporating a series of advanced design choices such as larger convolution kernels and hierarchical architectures. The ConvNeXt model adopts a simplified “modular” design. Within the ConvNeXt family, ConvNeXt-T serves as a lightweight baseline model targeting high-efficiency application scenarios, and its architecture is illustrated in Figure 8.
The model stacks 3, 3, 9, and 3 ConvNeXt blocks in its four stages, respectively, and progressively reduces the spatial resolution while increasing the channel dimensionality, thereby achieving a favorable balance between computational cost and representational capacity. The architecture of a ConvNeXt block is illustrated in Figure 9. While maintaining a relatively small model size, ConvNeXt-T demonstrates excellent accuracy and inference efficiency across a variety of visual tasks. Compared with larger ConvNeXt-S, B, and L models, ConvNeXt-T significantly reduces computational complexity and memory overhead at the cost of only a minor drop in peak accuracy, resulting in faster inference speed and higher energy efficiency.
According to the Kolmogorov turbulence theory, the covariance matrix C of the Zernike polynomial coefficient vector A = a 1 , a 2 , , a M can be obtained, which is expressed as:
a i a j = c i j D r 0 5 / 3
In Equation (6), c i j represents the covariance coefficient; a i and a j are the i -th and j -th Zernike coefficients, respectively; D is the aperture diameter; r 0 is the atmospheric coherence length; r 0 = 0.423 k 2 0 L C n 2 ( z ) d z 3 / 5 ; k = 2 π λ ; λ is the wavelength; C n 2 ( z ) is the refractive-index structure constant along the propagation path; and L is the propagation distance. The Zernike polynomial coefficient vector A can be obtained using the Karhunen–Loève (KL) polynomials.
C = V S V T A = V B
In Equation (7), S is a diagonal matrix, V is the coefficient matrix of the KL polynomials with the Zernike basis, and B is the corresponding vector of KL coefficients. Using Equation (1) and the Zernike coefficient vector statistically generated according to the Kolmogorov turbulence spectrum, atmospheric turbulence-distorted wavefronts can be synthesized. In this study, three datasets with D/r0 = 5, 10, 15 were prepared for training and testing.
By simulating the optical system together with atmospheric turbulence effects, a large and reliable dataset is generated, consisting of degraded intensity images and their corresponding Zernike coefficients. Specifically, 15,000 randomly generated sets of Zernike coefficients a 1 , a 2 , , a M and the associated intensity images I 1 ( x , y ) and I 2 ( x , y ) are used for network training, while an additional 1000 independently generated samples are reserved for testing. The training and test datasets are strictly non-overlapping. The proposed model is trained using the generated datasets, and its learning performance is evaluated based on the loss function. In this study, the L1 loss between the predicted and ground-truth Zernike coefficients is adopted, which is defined as:
f loss   = i = 1 M a t r u e , i a p r e d i c t , i
In Equation (8), a t r u e , i is the true value of the i -th Zernike coefficient, and a p r e d i c t , i is the predicted value of the i -th Zernike coefficient. Subsequently, the network model is iteratively optimized using the backpropagation algorithm. After each training epoch, the updated model is evaluated on a validation set to monitor its generalization performance. Based on the validation results, the training process is further refined to obtain the final optimized prediction model. In this paper, the effectiveness of the proposed method is validated through comparative experiments with other neural network models under identical datasets and experimental conditions.

3. Results and Discussion

3.1. Numerical Simulation

All simulations and experiments were conducted using the PyTorch deep learning framework (PyTorch 2.6.0+cu124). The figures were generated using MATLAB R2024b. The computer used for the experiment was running Windows 11, with the following hardware configuration: Intel Core i9-10900K CPU (3.7 GHz), 32 GB DDR4 RAM, and an NVIDIA GeForce RTX 2080Ti GPU. We constructed conventional CNN, ResNet50/101, ConvNeXt-T and EfficientNetV2-S networks, and trained these networks based on the established dataset. Each network was trained for 400 epochs with a batch size of 64. The loss function used was the L1 loss of the Zernike coefficients. A custom cosine annealing learning rate decay strategy was used, and the Adam optimizer was employed. The specific parameters of the four convolutional neural networks are shown in Table 3.
To further investigate the contribution of individual modules in the EfficientNetV2-S architecture, ablation experiments were conducted by removing the SE attention module and by replacing Fused-MBConv blocks with standard MBConv blocks. The L1 loss and the Zernike Root Mean Square Error (ZRMSE) were employed as evaluation metrics to quantify the prediction accuracy of the Zernike coefficients. For a more intuitive comparison of algorithm performance, the L1 loss function was scaled by a factor of 5000. The ZRMSE is defined in Equation (9).
Z R M S E = 1 M i = 1 M a t r u e , i a p r e d i c t , i 2
The experiments were conducted under two different turbulence conditions, D/r0 = 5 and D/r0 = 10, to examine model performance across different turbulence strengths, and the corresponding results are summarized in Table 4 and Table 5. Specifically, the model without the SE attention module is denoted as “Variant 1”, while the model in which Fused-MBConv is replaced by standard MBConv is denoted as “Variant 2”. These ablation studies aim to elucidate the individual contributions of the SE attention and Fused-MBConv modules to the overall prediction performance, particularly under varying turbulence conditions.
Under D/r0 = 5, the experimental results indicate that there are differences in model performance, especially with the removal of the SE module or the replacement of Fused-MBConv, which have a certain impact on model accuracy. Specifically, removing the SE module leads to a slight increase in L1 loss and ZRMSE, indicating that this module plays a role in enhancing model accuracy. A more noticeable change occurs when Fused-MBConv is replaced with MBConv, resulting in a significant drop in performance. This is especially evident in weaker turbulence conditions, where Fused-MBConv plays a particularly important role in improving feature extraction and convolution efficiency.
When the turbulence intensity increases to D/r0 = 10, L1 loss and ZRMSE for all models show a noticeable increase, reflecting the more complex impact of wavefront aberrations on model prediction under stronger turbulence conditions. Under these conditions, the effect of removing the SE module becomes more pronounced, and performance further declines. The performance degradation becomes even more evident when Fused-MBConv is replaced by the standard MBConv. In strong turbulence environments, standard MBConv may not effectively handle high-order aberrations, leading to performance deterioration.
Overall, the increase in turbulence intensity significantly affects model performance, and removing or replacing key modules such as the SE module and Fused-MBConv leads to a notable decline in accuracy. This indicates that these modules are crucial for the stability and accuracy of the model under different turbulence conditions.
Building upon the ablation study, we further evaluated the proposed approach against several widely used deep learning models to comprehensively assess its effectiveness. We trained the conventional CNN, ResNet50, ResNet101, ConvNeXt-T, and EfficientNetV2-S using 15,000 training images and 1000 validation images. Figure 10a–c present the training and validation loss curves computed by Equation (8). For all models, the loss decreases with increasing epochs and gradually converges to a stable level. Across both weak and strong turbulence conditions in WFS-less AO, EfficientNetV2-S consistently converges faster and achieves the lowest final L1 loss, indicating its superior stability. Benefiting from the efficient feature representation of Fused-MBConv and SE modules and effective feature reuse, it maintains lower loss and more stable outputs, demonstrating stronger robustness and generalization.
Figure 11 provides a quantitative comparison of the Zernike coefficient prediction errors (in units of λ) for the five networks under different turbulence strengths D/r0 = 5, 10, and 15. When the atmospheric turbulence strength is specified as D/r0 = 5, the ZRMSE of the conventional CNN, ResNet50, ResNet101, ConvNeXt-T, and EfficientNetV2-S are 0.067λ, 0.051λ, 0.061λ, 0.054λ, and 0.038λ, respectively. When D/r0 = 10, the ZRMSE for the five respective networks are 0.168λ, 0.128λ, 0.148λ, 0.096λ, and 0.071λ. When D/r0 = 15, the ZRMSE for the five respective networks are 0.292λ, 0.239λ, 0.261λ, 0.131λ, and 0.111λ.
As the turbulence strength increases, EfficientNetV2-S shows a more pronounced advantage over the other models. This is mainly attributed to the efficient feature extraction of Fused-MBConv, the enhanced channel representation provided by SE attention, and the well-balanced depth–width configuration enabled by compound scaling. Meanwhile, the chart indicates that ResNet50 outperforms ResNet101 in both loss value and wavefront detection accuracy. Increasing network depth introduces more nonlinear transformations, which may cause feature information to degrade and lead to “network degradation”, making training more difficult. Although residual connections can mitigate this effect, their benefit diminishes when the network becomes excessively deep, resulting in poorer training performance.
The first column shows the true wavefront, the second column shows the wavefront predicted by EfficientNetV2-S, and the third column shows the residual wavefront aberration. As shown in the figures, across different turbulence intensities, the wavefront reconstructed by EfficientNetV2-S is highly consistent with the 3D morphology of the true wavefront: peak/valley locations and high-order fluctuations are captured with good accuracy, indicating strong nonlinear fitting ability and good generalization in predicting high-order Zernike coefficients.
However, as turbulence increases, the reconstruction accuracy decreases. Under strong turbulence (D/r0 = 15), a slight phase discontinuity ring appears near the pupil edge. With stronger turbulence, recovery errors for high-order Zernike distortion coefficients become more pronounced, which may be attributed to the limited resolution of the simulation grid and the greater complexity of recovering high-order distortions.
The evaluation of randomly selected samples at D/r0 = 5, 10, and 15 shows that the model maintains good performance under varying turbulence strengths. As illustrated in Figure 12 (final epoch), an accurate prediction of Zernike coefficients from the 3rd to the 35th order is achieved. Even under strong turbulence, despite increased errors, the model preserves reasonable amplitude distributions and trends, offering reliable prior information for subsequent wavefront compensation.

3.2. Experimental Validation

To further validate the feasibility of the proposed method, a real-world experiment has been conducted. The experimental optical setup is schematically illustrated in Figure 13. This section directly uses the data collected from the experimental system to train an EfficientNetV2-S model. It mainly consists of a light source (Laser), a linear polarizer (P), a beam splitter (BS), a spatial light modulator (SLM), a focusing lens (L), a CCD camera, and a PC. The polarizer aligns the laser’s polarization direction to meet the operational requirements of the SLM. The SLM is then used to modulate the optical path length of the beam, thereby emulating atmospheric turbulence disturbances. Key parameters of the SLM and the camera are listed in Table 6.
The SLM employed in the experimental platform is a liquid-crystal optical phased array manufactured by Meadowlark Optics Inc. (Frederick, CO, USA). Image acquisition is performed using a Teledyne BFS-U3-23S3C-C camera (Thousand Oaks, CA, USA). All experiments are conducted within a PyTorch-based deep learning framework on a computing platform running Windows 11. The hardware configuration includes an Intel Core i9-10900K CPU operating at 3.7 GHz, 32 GB of DDR4 memory, and an NVIDIA GeForce RTX 2080 Ti GPU. To simulate wavefront perturbations induced by atmospheric turbulence, Zernike mode coefficients following a Kolmogorov turbulence spectrum are first randomly generated. Based on these coefficients, grayscale images are synthesized under different turbulence strengths, specifically D/r0 = 5 and 10.
The synthesized grayscale images were sequentially loaded onto the SLM to generate the corresponding disturbed wavefronts. During this process, the coefficients of the first two Zernike modes (piston and tip–tilt) were set to zero. The camera was positioned at the focal plane of the lens to acquire focal-plane images. Rather than physically translating the camera to obtain defocused-plane images, a defocus aberration with a peak-to-valley (PV) value of one wavelength was introduced by applying an additional phase term to the SLM. In total, 13,000 data samples were collected. Each sample consists of one focal-plane image, one defocused-plane image, and the corresponding Zernike coefficients from the 3rd to the 35th order describing the disturbed wavefront. The complete dataset was divided into a training set and a test set containing 12,000 and 1000 samples, respectively.
The wavefront reconstruction results of the EfficientNetV2-S network under the conditions of D/r0 = 5 and D/r0 = 10 are shown in Figure 14. For D/r0 = 5, the PV values of the initial and predicted wavefronts are 1.1097λ and 0.9077λ, respectively. The model successfully recovers the overall wavefront morphology, and the predicted 3rd to 35th-order Zernike coefficients are highly consistent with the ground truth. Although the recovery error for higher-order aberrations is relatively larger, especially near the pupil boundary, the model still effectively reconstructs the wavefront morphology.
When D/r0 = 10, the PV values of the initial and predicted wavefronts are 2.7166λ and 2.5137λ, respectively. As the turbulence intensity increases, the wavefront reconstruction performance exhibits a slight degradation; however, the model remains capable of recovering wavefronts with similar spatial morphology. These results demonstrate that EfficientNetV2-S can provide reliable and effective prior information across varying turbulence conditions, thereby facilitating subsequent wavefront compensation.

4. Conclusions

This paper presented a deep learning-based wavefront prediction framework for WFS-less AO using EfficientNetV2-S as the backbone network. By leveraging focused–defocused image pairs and directly regressing Zernike coefficients in a single forward pass, the proposed method avoids iterative optimization and effectively addresses the ambiguity inherent in single-intensity measurements. The framework enables the efficient and accurate prediction of high-order wavefront aberrations under various turbulence conditions. Comprehensive simulation and experimental results demonstrate that EfficientNetV2-S consistently outperforms conventional CNNs as well as representative modern architectures, including ResNet50/101 and ConvNeXt-T, in terms of L1 loss and ZRMSE. Ablation studies further confirm that key architectural components—namely the Fused-MBConv blocks and the SE attention mechanism—play a critical role in improving feature extraction, training stability, and robustness, particularly under medium-to-strong turbulence where high-order aberrations dominate. Even as turbulence intensity increases and overall prediction errors rise, the proposed model remains capable of recovering wavefronts with reasonable spatial morphology, providing reliable prior information for subsequent wavefront compensation. Beyond prediction accuracy, EfficientNetV2-S offers a favorable balance between representational capability and computational efficiency. Its compact parameter size and low inference latency make it well suited for real-time wavefront prediction, which is essential for practical WFS-less AO systems. These characteristics are particularly advantageous for deployment on embedded and edge-computing platforms, such as NVIDIA Jetson devices, where computational resources and power budgets are limited.

Author Contributions

Conceptualization, Z.Z. and H.Y.; methodology, Z.H.; software, Z.H.; validation, X.L. and J.W.; resources, Z.H.; writing—original draft preparation, Z.H. and C.L.; writing—review and editing, H.Y. and Z.Y.; funding acquisition, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (grant number U2141255 and 12473081).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Angel, J.R.P.; Wizinowich, P.; Lloyd-Hart, M.; Sandler, D. Adaptive optics for array telescopes using neural-network techniques. Nature 1990, 348, 221–224. [Google Scholar] [CrossRef]
  2. Jin, Y.; Zhang, Y.; Hu, L.; Huang, H.; Xu, Q.; Zhu, X.; Si, K. Machine learning guided rapid focusing with sensor-less aberration corrections. Opt. Express 2018, 26, 30162–30171. [Google Scholar] [CrossRef] [PubMed]
  3. Alvarez Diez, C.; Shao, F.; Bille, J. Pyramid and Hartmann–Shack wavefront sensor with artificial neural network for adaptive optics. J. Mod. Opt. 2008, 55, 683–689. [Google Scholar] [CrossRef]
  4. Guo, H.; Korablinova, N.; Ren, Q.; Bille, J. Wavefront reconstruction with artificial neural networks. Opt. Express 2006, 14, 6456–6462. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, N.; Zhu, L.; Ma, S.; Yang, P.; Xie, W. Research on distortion wavefront prediction technique for laser communication system based on LSTM network. In Proceedings of the Eighth Symposium on Novel Photoelectronic Detection Technology and Applications, Kunming, China, 9–11 November 2021; SPIE: Bellingham, WT, USA, 2022; Volume 12169, pp. 821–827. [Google Scholar]
  6. Wong, A.P.; Norris, B.R.; Tuthill, P.G.; Scalzo, R.; Lozi, J.; Vievard, S.; Guyon, O. Predictive control for adaptive optics using neural networks. J. Astron. Telesc. Instrum. Syst. 2021, 7, 019001. [Google Scholar] [CrossRef]
  7. Hu, Y.; Liu, X.; Cui, F.; Liu, X.; Hao, X. Research progress and prospects of adaptive optics technology based on deep learning. Chin. J. Lasers 2023, 50, 142–154. [Google Scholar]
  8. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  9. Xu, Z.; Yang, P.; Hu, K.; Xu, B.; Li, H. Deep learning control model for adaptive optics systems. Appl. Opt. 2019, 58, 1998–2009. [Google Scholar] [CrossRef] [PubMed]
  10. Sandler, D.G.; Barrett, T.K.; Palmer, D.A.; Fugate, R.Q.; Wild, W.J. Use of a neural network to control an adaptive optics system for an astronomical telescope. Nature 1991, 351, 300–302. [Google Scholar] [CrossRef]
  11. Barrett, T.K.; Sandler, D.G. Artificial neural network for the determination of Hubble Space Telescope aberration from stellar images. Appl. Opt. 1993, 32, 1720–1727. [Google Scholar] [CrossRef]
  12. Suzuki, K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef] [PubMed]
  13. Hebbalaguppe, R.; Garg, G.; Hassan, E.; Ghosh, H.; Verma, A. Telecom Inventory management via object recognition and localisation on Google Street View Images. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; IEEE: New York, NY, USA, 2017; pp. 725–733. [Google Scholar]
  14. Ringeval, F.; Schuller, B.; Valstar, M.; Jaiswal, S.; Marchi, E.; Lalanne, D.; Cowie, R.; Pantic, M. AV + EC 2015: The first affect recognition challenge bridging across audio, video, and physiological data. In Proceedings of the 5th International Workshop on Audio/Visual Emotion Challenge, Brisbane, Australia, 26 October 2015; pp. 3–8. [Google Scholar]
  15. Manana, M.; Tu, C.; Owolawi, P.A. A survey on vehicle detection based on convolution neural networks. In Proceedings of the 2017 3rd IEEE International Conference on Computer and Communications (ICCC), Chengdu, China, 13–16 December 2017; IEEE: New York, NY, USA, 2017; pp. 1751–1755. [Google Scholar]
  16. Nguyen, T.; Bui, V.; Lam, V.; Raub, C.B.; Chang, L.-C.; Nehmetallah, G. Automatic phase aberration compensation for digital holographic microscopy based on deep learning background detection. Opt. Express 2017, 25, 15043–15057. [Google Scholar] [CrossRef] [PubMed]
  17. Nishizaki, Y.; Valdivia, M.; Horisaki, R.; Kitaguchi, K.; Saito, M.; Tanida, J.; Vera, E. Deep learning wavefront sensing. Opt. Express 2019, 27, 240–251. [Google Scholar] [CrossRef] [PubMed]
  18. Ma, H.; Liu, H.; Qiao, Y.; Li, X.; Zhang, W. Numerical study of adaptive optics compensation based on convolutional neural networks. Opt. Commun. 2019, 433, 283–289. [Google Scholar] [CrossRef]
  19. Wang, M.; Guo, W.; Yuan, X. Single-shot wavefront sensing with deep neural networks for free-space optical communications. Opt. Express 2021, 29, 3465–3478. [Google Scholar] [CrossRef] [PubMed]
  20. Li, Y.; Yue, D.; He, Y. Prediction of wavefront distortion for wavefront sensorless adaptive optics based on deep learning. Appl. Opt. 2022, 61, 4168–4176. [Google Scholar] [CrossRef] [PubMed]
  21. Ma, H.; Zhang, W.; Ning, X.; Liu, H.; Zhang, P.; Zhang, J. Turbulence aberration restoration based on light intensity image using GoogLeNet. Photonics 2023, 10, 265. [Google Scholar] [CrossRef]
  22. Kok, Y.E.; Bentley, A.; Parkes, A.; Wright, A.J.; Somekh, M.G.; Pound, M. Direct Zernike Coefficient Prediction from Point Spread Functions and Extended Images using Deep Learning. arXiv 2024, arXiv:2404.15231. [Google Scholar] [CrossRef]
  23. Tan, M.; Le, Q. EfficientNetV2: Smaller models and faster training. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10096–10106. [Google Scholar]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  25. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 27 September 2022; pp. 11976–11986. [Google Scholar]
Figure 1. Schematic of the WFS-less AO system.
Figure 1. Schematic of the WFS-less AO system.
Photonics 13 00144 g001
Figure 2. Structure diagram of the EfficientNetV2-S network.
Figure 2. Structure diagram of the EfficientNetV2-S network.
Photonics 13 00144 g002
Figure 3. Fused-MBConv module [23].
Figure 3. Fused-MBConv module [23].
Photonics 13 00144 g003
Figure 4. Structure diagram of the conventional CNN.
Figure 4. Structure diagram of the conventional CNN.
Photonics 13 00144 g004
Figure 5. Structure diagram of the ResNet network.
Figure 5. Structure diagram of the ResNet network.
Photonics 13 00144 g005
Figure 6. Residual module [20].
Figure 6. Residual module [20].
Photonics 13 00144 g006
Figure 7. ResNet50/101 series residual module [20].
Figure 7. ResNet50/101 series residual module [20].
Photonics 13 00144 g007
Figure 8. Architecture of the ConvNeXt-T network.
Figure 8. Architecture of the ConvNeXt-T network.
Photonics 13 00144 g008
Figure 9. Structure of the ConvNeXt block module [25].
Figure 9. Structure of the ConvNeXt block module [25].
Photonics 13 00144 g009
Figure 10. (a) D/r0 = 5, (b) D/r0 = 10, and (c) D/r0 = 15 training and testing results of the five models.
Figure 10. (a) D/r0 = 5, (b) D/r0 = 10, and (c) D/r0 = 15 training and testing results of the five models.
Photonics 13 00144 g010
Figure 11. Zernike coefficient prediction errors under D/r0 = 5, 10 and 15.
Figure 11. Zernike coefficient prediction errors under D/r0 = 5, 10 and 15.
Photonics 13 00144 g011
Figure 12. Comparison of the original wavefront (a1a3), predicted wavefront (b1b3), and Zernike coefficients of the true and predicted wavefronts (c1c3) at D/r0 = 5, 10, and 15.
Figure 12. Comparison of the original wavefront (a1a3), predicted wavefront (b1b3), and Zernike coefficients of the true and predicted wavefronts (c1c3) at D/r0 = 5, 10, and 15.
Photonics 13 00144 g012
Figure 13. Schematic illustration of the experimental system.
Figure 13. Schematic illustration of the experimental system.
Photonics 13 00144 g013
Figure 14. Comparison of the original wavefront (a1,a2), predicted wavefront (b1,b2), and Zernike coefficients of the true and predicted wavefronts (c1,c2) at D/r0 = 5, 10.
Figure 14. Comparison of the original wavefront (a1,a2), predicted wavefront (b1,b2), and Zernike coefficients of the true and predicted wavefronts (c1,c2) at D/r0 = 5, 10.
Photonics 13 00144 g014
Table 1. EfficientNetV2-S network model parameters table [23].
Table 1. EfficientNetV2-S network model parameters table [23].
StageOperatorStrideChannelsLayers
0Conv3 × 32241
1Fused-MBConv1, k3 × 31242
2Fused-MBConv4, k3 × 32484
3Fused-MBConv4, k3 × 32644
4MBConv4, k3 × 3, SE0.2521286
5MBConv6, k3 × 3, SE0.2511609
6MBConv6, k3 × 3, SE0.25225615
7Conv1 × 1 & Pooling & FC-12801
Table 2. EfficientNetV2 series network configuration [23].
Table 2. EfficientNetV2 series network configuration [23].
EfficientNetV2SML
MinMaxMinMaxMinMax
Image size128300128380128380
RandAugment515520525
Mixup alpha0000.200.4
Dropout0.10.30.10.40.10.5
Table 3. Specifications of the neural network.
Table 3. Specifications of the neural network.
Neural NetworkInputEpochsOptimizerBatch SizeLoss Function
CNN128 × 128400Adam64L1
ResNet50/101128 × 128400Adam64L1
ConvNeXt-T128 × 128400Adam64L1
EfficientNetV2-S128 × 128400Adam64L1
Table 4. Performance comparison of ablation variants under D/r0 = 5.
Table 4. Performance comparison of ablation variants under D/r0 = 5.
ModelL1 LossZRMSE
Baseline1090.038λ
Variant 11370.047λ
Variant 22080.068λ
Table 5. Performance comparison of ablation variants under D/r0 = 10.
Table 5. Performance comparison of ablation variants under D/r0 = 10.
ModelL1 LossZRMSE
Baseline2080.071λ
Variant 12280.077λ
Variant 22940.091λ
Table 6. Specific parameters of the SLM and the camera.
Table 6. Specific parameters of the SLM and the camera.
DevicesPixel SizeNumber of PixelsBit Depth
SLM9.2 µm × 9.2 µm1920 × 115212-bit
CCD3.45 µm × 3.45 µm1920 × 120012-bit
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Z.; Huang, Z.; Wu, J.; Yan, Z.; Li, X.; Liu, C.; Yang, H. Wavefront Prediction for Adaptive Optics Without Wavefront Sensing Based on EfficientNetV2-S. Photonics 2026, 13, 144. https://doi.org/10.3390/photonics13020144

AMA Style

Zhang Z, Huang Z, Wu J, Yan Z, Li X, Liu C, Yang H. Wavefront Prediction for Adaptive Optics Without Wavefront Sensing Based on EfficientNetV2-S. Photonics. 2026; 13(2):144. https://doi.org/10.3390/photonics13020144

Chicago/Turabian Style

Zhang, Zhiguang, Zelu Huang, Jiawei Wu, Zhaojun Yan, Xin Li, Chang Liu, and Huizhen Yang. 2026. "Wavefront Prediction for Adaptive Optics Without Wavefront Sensing Based on EfficientNetV2-S" Photonics 13, no. 2: 144. https://doi.org/10.3390/photonics13020144

APA Style

Zhang, Z., Huang, Z., Wu, J., Yan, Z., Li, X., Liu, C., & Yang, H. (2026). Wavefront Prediction for Adaptive Optics Without Wavefront Sensing Based on EfficientNetV2-S. Photonics, 13(2), 144. https://doi.org/10.3390/photonics13020144

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop