Next Article in Journal
Laplacian Scores-Based Feature Reduction in IoT Systems for Agricultural Monitoring and Decision-Making Support
Previous Article in Journal
A New Feature Descriptor for Multimodal Image Registration Using Phase Congruency
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Letter

A Single Far-Field Deep Learning Adaptive Optics System Based on Four-Quadrant Discrete Phase Modulation

1
Key Laboratory of Adaptive Optics, Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, Sichuan, China
2
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, Sichuan, China
3
University of Chinese Academy of Science, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5106; https://doi.org/10.3390/s20185106
Submission received: 5 August 2020 / Revised: 26 August 2020 / Accepted: 3 September 2020 / Published: 8 September 2020
(This article belongs to the Section Optical Sensors)

Abstract

:
In adaptive optics (AO), multiple different incident wavefronts correspond to a same far-field intensity distribution, which leads to a many-to-one mapping. To solve this problem, a single far-field deep learning adaptive optics system based on four-quadrant discrete phase modulation (FQDPM) is proposed. Our method performs FQDPM on an incident wavefront to overcome this many-to-one mapping, then convolutional neural network (CNN) is used to directly predict the wavefront. Numerical simulations indicate that the proposed method can achieve precise high-speed wavefront correction with a single far-field intensity distribution: it takes nearly 0.6ms to complete wavefront correction while the mean root mean square (RMS) of residual wavefronts is 6.3% of that of incident wavefronts, and the Strehl ratio of the far-field intensity distribution increases by 5.7 times after correction. In addition, the experiment results show that mean RMS of residual wavefronts is 6.5% of that of incident wavefronts and it takes nearly 0.5 ms to finish wavefront reconstruction, which verifies the correctness of our proposed method.

1. Introduction

Adaptive optics (AO) has played an important role in beam shaping [1], super-resolution microscopy [2], human eye imaging [3], and so on. AO is composed of three key elements: a wavefront sensor (WFS) to measure distorted wavefronts, a wavefront controller that outputs control signals to a wavefront corrector based on information acquired by WFS, and a wavefront corrector such as a deformable mirror (DM) to provide wavefront correction. Conventional WFSs include Shack–Hartmann wavefront sensors [4], curvature sensors [5], plenoptic sensors [6], pyramid wavefront sensors [7], and so on. Although these WFSs have achieved real-time closed-loop wavefront detection, they have delicate optical structures, and the implementation cost is high.
Phase retrieval (PR) can directly detect wavefront phases using far-field intensity distribution. PR can work without WFS, as it has no need for calibration and implementation cost is low. Classic PR algorithms include Gerchberg–Saxton algorithm (GS) [8], phase diversity (PD) algorithm [9,10], and so on. However, PR algorithms need multiple iterations to converge, and their real-time performance is poor. In addition, PD algorithms require a camera such as a charge coupled device (CCD) to make multiple measurements on focal and defocus planes, and algorithms have limited applications in wavefront detection.
Convolutional neural network (CNN), as a branch of deep learning, has excellent performance rates in image processing. In recent years, scholars have begun to use the nonlinear fitting characteristics of CNN to acquire wavefront aberration information based on a single far-field intensity distribution image [11], achieving high wavefront detection accuracy with a simple optical system structure. However, in AO, far-field intensity distribution corresponds to multiple different incident wavefronts. This many-to-one mapping relationship means that, in supervised leaning, a sample can be related to many labels, so CNN cannot handle this kind of many-to-one mapping relationship. For this reason, scholars can only achieve precise wavefront reconstruction under small aberrations [12,13] with a single far-field intensity distribution. Authors in [14,15] applied the idea that regarding PD, far-field intensity distributions on focal and defocus planes are acquired to overcome many-to-one mapping. CCD measuring on a defocus plane is equivalent to introducing a defocus aberration, and mapping from a far-field intensity distribution to the incident wavefront is no longer a many-to-one mapping. Nevertheless, these algorithms require multiple measurements on focal and defocus planes, their system structures are complex, and the algorithms have limited applications.
In this paper we analyze the origin of the many-to-one mapping relationship and propose a single far-field CNN-AO based on four-quadrant discrete phase modulation (FQDPM). Our method overcomes many-to-one mapping by modulating the incident wavefront with FQDPM, and it reduces system computing time by replacing multiple iterations with CNN. This paper is arranged as follows. In Section 2, we introduce the working principle of FQDPM-based CNN-AO. In Section 3, numerical simulations and corresponding analyses are given. In Section 4, relevant experiments are built to verify the correctness of our theory, while conclusions follow in Section 5.

2. Working Principle of FQDPM-Based CNN-AO

The relationship between the complex amplitude of far-field U f a r ( x 0 , y 0 ) and pupil plane U n e a r ( x , y ) can be expressed as
U f a r ( x 0 , y 0 ) =   U n e a r ( x , y ) e x p [ i 2 π ( u x + v y ) ] d x d y  
U n e a r ( x , y ) = A n e a r ( x , y ) e x p [ i φ ( x , y ) ]
where (x0, y0) and (x, y) are rectangular coordinates on imaging and pupil planes, respectively, u and v are spatial frequencies u = x 0 λ f and v = y 0 λ f , λ is wavelength, f is focal length, A n e a r ( x , y ) represents the amplitude of the incident wavefront, and φ ( x , y ) is the wavefront phase. Assuming that the system is under uniform irradiation, A n e a r ( x , y ) can be normalized to 1. By rotating wavefront φ ( x , y ) 180 degrees and then flipping it, we gain a new wavefront φ ( x , y ) = φ ( x , y ) . According to Euler’s formula, U f a r ( x 0 , y 0 ) and U f a r ( x 0 , y 0 ) can be newly expressed as
U f a r ( x 0 , y 0 ) =   { cos [ φ ( x , y ) 2 π ( u x + v y ) ] + i [ φ ( x , y ) 2 π ( u x + v y ) ] } d x d y
U f a r ( x 0 , y 0 ) =   { cos { φ ( x , y ) 2 π [ u ( x ) + v ( y ) ] } i sin { φ ( x , y ) 2 π [ u ( x ) + v ( y ) ] } } d ( x ) d ( y )
where U f a r ( x 0 , y 0 ) and U f a r ( x 0 , y 0 ) are corresponding complex amplitudes of φ ( x , y ) and φ ( x , y ) , respectively. According to Equations (3) and (4), U f a r ( x 0 , y 0 ) and U f a r ( x 0 , y 0 ) have the same real components and opposite imaginary components. In Fourier optics, the far-field intensity distribution is equivalent to the squared modular operation of the complex amplitude; thus, the corresponding far-field intensity distribution can be expressed as
| U f a r ( x 0 , y 0 ) | 2 = | U f a r ( x 0 , y 0 ) | 2
where φ ( x , y ) = φ ( x , y ) , U f a r ( x 0 , y 0 ) , and U f a r ( x 0 , y 0 ) correspond to the same far-field intensity distribution. In supervised learning, Equation (5) means that a sample corresponds to multiple labels, the mapping relationship is morbid, and supervised learning cannot decipher which label is true. Performing a phase modulation on the incident wavefront can fix this problem. Assuming that we perform phase modulation ϕ ( x , y ) on incident wavefronts φ ( x , y ) and φ ( x , y ) , the new complex amplitude U n e w f a r ( x 0 , y 0 ) and U n e w f a r ( x 0 , y 0 ) can be expressed as
U n e w f a r ( x 0 , y 0 ) =   { cos [ φ ( x , y ) + ϕ ( x , y ) 2 π ( u x + v y ) ] + i sin [ φ ( x , y ) + ϕ ( x , y ) 2 π ( u x + v y ) ] } d x d y
U n e w f a r ( x 0 , y 0 ) =   { cos [ φ ( x , y ) ϕ ( x , y ) 2 π ( u x + v y ) ] i sin [ φ ( x , y ) ϕ ( x , y ) 2 π ( u x + v y ) ] } d x d y
According to Equations (6) and (7), when ϕ ( x , y ) ϕ ( x , y ) , U n e w f a r ( x 0 , y 0 ) and U n e w f a r ( x 0 , y 0 ) no longer correspond to the same far-field intensity distribution, and a pair of rotating and flipped incident wavefronts will not be the same again.
The phase modulation ϕ ( x , y ) must be discrete because when ϕ ( x , y ) is continuous, incident wavefronts φ ( x , y ) and φ ( x , y ) = Φ ( x , y ) Φ ( x , y ) φ ( x , y ) still correspond to the same far-field intensity distribution, so continuous phase modulation cannot fundamentally destroy the many-to-one mapping relationship. Discrete phase modulation is not impacted by this problem because φ ( x , y ) does not exist when the incident wavefront is continuous.
FQDPM divides a unit circle into four quadrants: the first and third quadrants generate a phase step of π 4 , while the second and fourth quadrants generate a phase step of π 4 . FQDPM is one of the simplest discrete phase modulations that meet the requirements of phase modulation. In addition, FQDPM aberration plates are easy to manufacture. For these reasons, FQDPM is the chosen modulation. Assuming that the expression of FQDPM is δ ( x , y ) , δ ( x , y ) can be expressed as
δ ( x , y ) = { π 4 x y 0 π 4 x y < 0
When we perform FQDPM on an incident wavefront, according to Equations (6)–(8), FQDPM can break the many-to-one mapping that multiple incident wavefronts correspond to in a same far-field intensity distribution. Due to this, it is feasible to use CNN to fit the mapping from the far-field intensity distribution to the wavefront aberration information. A FQDPM-based CNN-AO system is shown in Figure 1, and a FQDPM aberration plate is added to the optical system between a DM and a positive lens. A FQDPM-based CNN-AO system is composed of three key elements: a FQDPM-based CNN-WFS to measure the far-field intensity distribution on a lens focal plane and calculate wavefront aberration information, a voltage reconstruction matrix that outputs voltage control signals to DM based on calculated aberration information, and a DM to provide wavefront correction on the basis of input voltage control signals.
The working principle of FQDPM-based CNN-AO is shown in Figure 2. CNN is used to fit the nonlinear mapping between modulated PSFs and Zernike coefficients. Errors between Zernike coefficients are predicted by CNN, and target coefficients are sent to the cost function, while CNN adjusts its parameters based on gradient information of the cost function; this process is called CNN training. After CNN training, CNN directly outputs Zernike coefficients according to input modulated PSFs, then the voltage reconstruction matrix converts predicted coefficients acquired by CNN into DM control voltages. DM generates corresponding surface shapes to compensate for the incident wavefront on the basis of control voltages; thus, the system achieves closed-loop correction. Numbers of CCD pixels used in FQDPM-based CNN-WFS are 120 × 120 pixels.

3. Numerical Simulations

Relevant parameters of the numerical simulations are as follows: wavelength was 1064 nm, radius of the optical aperture was 8 mm, focal length of the lens was 1000 mm, CCD pixel size was 10   μ m × 10   μ m , and number of pixels in CCD was 120   × 120 pixels. CNN architecture and its hyper parameters are shown in Figure 3. It has 13 layers, including 7 convolutional layers, 3 pooling layers, and 2 fully connected (FC) layers. A PSF image with a size of 120   × 120 pixels was processed in the input layer; the number of channels of each convolution layer was 16, 32, 32, 32, 64, 64, and 64, respectively; the size of each convolution kernel was 5 × 5, 5 × 5, 5 × 5, 5 × 5, 4 × 4, 4 × 4, and 4 × 4, respectively; the pooling layer used max pooling with a stride of 3; and the number of neurons in FC layers were 200 and 20, respectively. CNN selected ReLU as the activation function and was applied in all convolutional layers and the first FC layer. This task is a regression task, so we chose the mean square error (MSE) function as CNN loss function. During CNN training, the batch size was set to 100, the epoch was set to 200, and the Adam function with an initial learning rate 10−3 was chosen as the CNN gradient descent function. Workstation configurations were Intel Core i7 9700 K 3.6 GHz, Kingston 64 GB, and NVIDIA GeForce RTX2080Ti. Training and testing were finished on the GPU. It took nearly 14 min to complete CNN training and 0.6 ms for our method to complete one round of wavefront correction.

3.1. Generate Dataset

A series of 10,000 incident wavefronts were randomly generated based on the 4th to 23rd Zernike coefficients. After modulation by FQDPM, new incident wavefronts were measured by CCD to obtain PSFs. The range of each Zernike coefficients was within ±0.5 μm. Samples in training and test sets were independently and identically distributed, so we randomly selected 9000 PSFs and corresponding Zernike coefficients as samples and labels in the training set, respectively. The remaining 1000 PSFs and corresponding Zernike coefficients were used as samples and labels in the test set, respectively. Labels were 20 × 1 column vectors. A training set was used so CNN could learn the mapping relationship between the modulated PSFs and corresponding Zernike coefficients. A test set was used to evaluate the accuracy of our method.

3.2. Results and Analyses of Simulations

In order to evaluate the correction results of numerical simulations, the root mean square (RMS) of residual wavefronts and Strehl ratio (SR) of the far-field intensity distribution were selected as evaluation indexes. Figure 4a,b shows RMS and SR, respectively, before and after correction corresponding to 1000 samples in the test set. It is known from Figure 4a that the mean RMS of incident wavefronts before correction was 1.259 μ m , and mean RMS of residual wavefronts after correction dropped to 0.079 μ m ; thus, the mean RMS of the corrected residual wavefronts was 6.3% of that of incident wavefronts. In Figure 4b, the mean SR before correction was 0.161, and the mean SR after correction rose to 0.918; thus, the mean SR increased by 5.7 times after correction. Results in Figure 4 indicated that the well-trained CNN combined with the voltage reconstruction matrix could effectively and directly output DM control voltages to make AO closed-loop based on a single modulated PSF. Our method converged the algorithm after one calculation, ensuring that no more iterations were needed to optimize the performance index, which led to a greatly reduced computing time.
Two samples were randomly selected from the test set. Figure 5 exhibits two corresponding incident wavefronts, initial PSFs, residual wavefronts, and PSFs after correction. Fitting results between predicted Zernike coefficients and labels are also shown in Figure 5; the scales on the X-axis and Y-axis represent Zernike modes and Zernike coefficients, respectively. In each Zernike mode, the left pillars represent the labeled Zernike coefficients while the right pillars represent Zernike coefficients predicted by CNN. Figure 5a,b corresponds to the 494th and 932nd samples in the test set, respectively. Root-mean-square error (RMSE) was introduced to evaluate fitting results of Zernike coefficients predicted by the proposed method.
In Figure 5a, the initial RMS and SR corresponding to the 494th sample in the test set were 1.382 μm and 0.316 respectively. After correction, RMS and SR were 0.066 μm and 0.982, respectively, and RMSE between predicted Zernike coefficients and labels was 0.033 μm. In Figure 5b, the initial RMS and SR corresponding to the 932nd sample in the test set were 1.366 μm and 0.152, respectively. After correction, RMS and SR were 0.087 μm and 0.971, respectively, and RMSE between predicted Zernike coefficients and labels was 0.039 μm. CNN achieved good fitting and the method achieved precise aberration correction.

4. Experiments

4.1. Experimental Setup

It is known from Figure 2 in Section 2 that the voltage reconstruction matrix needs exact expressions of DM influence functions. In simulations, DM influence functions can be precisely defined, but in experiments they should be measured by a WFS such as Shack–Hartmann wavefront sensor or interferometer. To simplify procedures of experiments, in this section we built a wavefront detection system based on a liquid crystal spatial light modulator (LC-SLM) and ignored the wavefront correction. A schematic diagram of the optical system structure is shown in Figure 6. Our system contained three parts: a coherent point source, an aberration generator, and a FQDPM-based CNN-WFS. The coherent source was composed of a laser, a pinhole, and a positive lens. The pinhole was placed at the focal point of the positive lens to generate a collimated beam. The aberration generator was composed of a beam splitter and LC-SLM. When the incident beam is reflected by LC-SLM, LC-SLM adds additional aberrations to the incident beam to change its wavefront phase distribution. The FQDPM-based CNN-WFS contained a FQDPM aberration plate, a positive lens, and a CCD which was placed at the focal plane of the lens. A distorted beam was reflected by a beam splitter and reached FQDPM-based CNN-WFS, the FQDPM aberration plate then performed FQDPM on the incident beam, and the positive lens focused the beam, and CCD measured its corresponding far-field intensity distribution.
A schematic diagram of the real optical path is shown in Figure 7, and relevant experimental parameters are given as follows: wavelength was 1064 μ m , pinhole diameter was 10 μm, focal lengths of the lens in coherent point source and FQDPM-based CNN-WFS were 100 mm and 350 mm, respectively, CCD pixel size was 7.4 μm × 7.4 μm, and numbers of CCD pixels used in FQDPM-based CNN-WFS was 120 × 120 pixels.
According to Section 3.1, 10,000 series of samples were randomly generated according to the 4th to 23rd Zernike coefficients by LC-SLM. The range of each Zernike coefficients was within ±0.5 μm. A total of 9000 PSFs and corresponding Zernike coefficients were randomly selected as samples and labels in the training set, respectively, and the remaining 1000 PSFs and corresponding Zernike coefficients were used as samples and labels in the test set, respectively. Workstation and CNN setups were the same as mentioned in Section 3. It took nearly 14 min to complete CNN training and nearly 0.5ms for our method to reconstruct incident wavefront a single time.

4.2. Results and Analyses of Experiments

Figure 8 shows RMS of residual wavefronts and RMS of incident wavefronts corresponding to 1000 samples in the test set. It is known from Figure 8 that the mean RMS of incident wavefronts was 1.276 μm, mean RMS of residual wavefronts dropped to 0.084 μm, and mean RMS of reconstructed residual wavefronts was 6.5% of that of incident wavefronts. The well-trained CNN could directly output wavefront aberration information after one calculation with a single modulated PSF.
Two samples were randomly selected from the test set. Figure 9 exhibits two incident PSFs, incident wavefronts, and predicted wavefronts. Fitting results between predicted Zernike coefficients and labels are also shown in Figure 9; the scales on the X-axis and Y-axis represent Zernike modes and Zernike coefficients, respectively. In each Zernike mode, the left pillars represent labeled Zernike coefficients and the right pillars represent Zernike coefficients predicted by CNN. Figure 9a,b corresponds to the 54th and 580th samples in the test set, respectively.
In Figure 9a, RMS of incident wavefront corresponding to the 54th sample was 1.248 μm, RMS of the predicted wavefront was 1.224 μm, and RMSE between labels and predicted coefficients was 0.031 μm. In Figure 9b, the initial RMS corresponding to the 580th sample was 1.158 μm, RMS of predicted wavefront was 1.132 μm, and RMSE between labels and predicted coefficients was 0.046 μm. According to Figure 9, surface shapes of predicted wavefronts and incident wavefronts were nearly identical, and FQDPM-based CNN-WFS could reconstruct wavefronts quickly and perfectly with a single modulated PSF.

5. Conclusions

In this paper we propose a single far-field FQDPM-based CNN-AO system. Our method applies FQDPM to incident wavefronts to overcome the ill-conditioned many-to-one mapping that multiple incident wavefronts correspond to a same far-field intensity distribution. We also establish a new mapping relationship between single modulated PSF and DM control voltages. Numerical simulations indicated that the mean RMS of residual wavefronts corresponding to 1000 samples in the test set dropped from 1.259 μm to 0.079 μm after correction by our proposed method, the mean SR rose from 0.161 to 0.9176, and the computing time was 0.6 ms for one correction. In the experiments we built a FQDPM-based CNN-WFS. Results showed that our proposed method achieved precise and high-speed wavefront reconstruction, the mean RMS of residual wavefronts corresponding to 1000 samples in the test set dropped from 1.276 μ m to 0.084 μ m , and it took nearly 0.5 ms to finish a single wavefront reconstruction.

Author Contributions

Conceptualization, X.Q. and T.C.; methodology, X.Q.; software, T.C. and L.K.; validation, X.Q. and L.K.; formal analysis, X.Q.; investigation, X.Q. and T.C.; resources, B.X.; data curation, X.Q. and L.K.; writing—original draft preparation, X.Q.; writing—review and editing, X.Q.; visualization, X.Q.; supervision, S.W.; project administration, S.W.; funding acquisition, B.X. All authors have read and agreed to the published version of the manuscript.

Funding

National Natural Science Foundation of China (61875203).

Acknowledgments

Suggestions and comments from Wang Zhao, Licheng Zhu, Jingjing Yang, and Chao Yang are very much appreciated.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Y.Y.; Wu, C.X.; Song, Y.C.; Si, K.; Zheng, Y.; Hu, L.J.; Chen, J.J.; Tang, L.M.; Gong, W. Machine learning based adaptive optics for doughnut-shaped beam. Opt. Express 2019, 27, 16871–16881. [Google Scholar] [CrossRef] [PubMed]
  2. Ji, N. Adaptive optical fluorescence microscopy. Nat. Methods 2017, 14, 374–380. [Google Scholar] [CrossRef] [PubMed]
  3. Liang, J.; Williams, D.R.; Miller, D.T. Supernormal vision and high-resolution retinal imaging through adaptive optics. J. Opt. Soc. Am. A 1997, 14, 2884–2892. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Dubose, T.B.; Gardner, D.F.; Watnik, A. Intensity-enhanced deep network wavefront reconstruction in Shack-Hartmann sensors. Opt. Lett. 2020, 45, 1699–1702. [Google Scholar] [CrossRef] [PubMed]
  5. Roddier, F. Curvature sensing and compensation: A new concept in adaptive optics. Appl. Opt. 1988, 27, 1223–1225. [Google Scholar] [CrossRef] [PubMed]
  6. Wu, S.C.; Ko, J.; Davis, C.C. Determining the phase and amplitude distortion of a wavefront using a plenoptic sensor. J. Opt. Soc. Am. A 2015, 32, 964–978. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, S.Q.; Wei, K.; Zheng, W.J. Modulation-nonmodulation pyramid wavefront sensor with direct gradient reconstruction algorithm on the closed-loop adaptive optics system. Opt. Express 2018, 26, 20952–20964. [Google Scholar] [CrossRef] [PubMed]
  8. Misell, D.L. An examination of an iterative method for the solution of the phase problem in optics and electron optics: I. Test calculations. J. Phys. D Appl. Phys. 1973, 6, 2200–2216. [Google Scholar] [CrossRef]
  9. Gonsalves, R.A. Phase retrieval and diversity in adaptive optics. Opt. Eng. 1982, 21, 829–832. [Google Scholar] [CrossRef]
  10. Nicolas, V.; Mugnier, L.M.; Michau, V.; Marie, T.V.; Bierent, R. Laser beam complex amplitude measurement by phase diversity. Opt. Express 2014, 22, 4575–4589. [Google Scholar]
  11. Paine, S.W.; Fienup, J.R. Machine learning for improved image-based wavefront sensing. Opt. Lett. 2018, 43, 1235–1238. [Google Scholar] [CrossRef] [PubMed]
  12. Nishizaki, Y.; Valdivia, M.; Horisaki, R.; Kitaguchi, K.; Saito, M.; Tanida, J.; Vera, E. Deep learning wavefront sensing. Opt. Express 2019, 27, 240–251. [Google Scholar] [CrossRef] [PubMed]
  13. Tian, Q.H.; Lu, C.D.; Liu, B.; Zhu, L.; Pan, X.L.; Zhang, Q.; Yang, L.J.; Tian, F.; Xin, X.J. DNN-based aberration correction in a wavefront sensorless adaptive optics system. Opt. Express 2019, 27, 10765–10776. [Google Scholar] [CrossRef] [PubMed]
  14. Ju, G.H.; Qi, X.; Ma, H.C.; Yan, C.X. Feature-based phase retrieval wavefront sensing approach using machine learning. Opt. Express 2018, 26, 31767–31783. [Google Scholar] [CrossRef] [PubMed]
  15. Guo, H.Y.; Xu, Y.J.; Li, Q.; Du, S.P.; He, D.; Wang, Q.; Huang, Y.M. Improved machine learning approach for wavefront sensing. Sensors 2019, 19, 3533. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Schematic diagram of the four-quadrant discrete phase modulation (FQDPM)-based convolutional neural network adaptive optics (CNN-AO) system.
Figure 1. Schematic diagram of the four-quadrant discrete phase modulation (FQDPM)-based convolutional neural network adaptive optics (CNN-AO) system.
Sensors 20 05106 g001
Figure 2. Working principle of the FQDPM-based CNN-AO.
Figure 2. Working principle of the FQDPM-based CNN-AO.
Sensors 20 05106 g002
Figure 3. CNN architecture.
Figure 3. CNN architecture.
Sensors 20 05106 g003
Figure 4. RMS (a) and SR (b) before and after correction of 1000 samples in the test set.
Figure 4. RMS (a) and SR (b) before and after correction of 1000 samples in the test set.
Sensors 20 05106 g004
Figure 5. Correction results of 494th (a) and 932nd (b) samples in test set.
Figure 5. Correction results of 494th (a) and 932nd (b) samples in test set.
Sensors 20 05106 g005
Figure 6. Schematic diagram of the optical system structure.
Figure 6. Schematic diagram of the optical system structure.
Sensors 20 05106 g006
Figure 7. Schematic diagram of a real optical path.
Figure 7. Schematic diagram of a real optical path.
Sensors 20 05106 g007
Figure 8. Wavefront reconstruction accuracy of samples in the test set.
Figure 8. Wavefront reconstruction accuracy of samples in the test set.
Sensors 20 05106 g008
Figure 9. Reconstruction results of 54th (a) and 580th (b) samples in the test set.
Figure 9. Reconstruction results of 54th (a) and 580th (b) samples in the test set.
Sensors 20 05106 g009

Share and Cite

MDPI and ACS Style

Qiu, X.; Cheng, T.; Kong, L.; Wang, S.; Xu, B. A Single Far-Field Deep Learning Adaptive Optics System Based on Four-Quadrant Discrete Phase Modulation. Sensors 2020, 20, 5106. https://doi.org/10.3390/s20185106

AMA Style

Qiu X, Cheng T, Kong L, Wang S, Xu B. A Single Far-Field Deep Learning Adaptive Optics System Based on Four-Quadrant Discrete Phase Modulation. Sensors. 2020; 20(18):5106. https://doi.org/10.3390/s20185106

Chicago/Turabian Style

Qiu, Xuejing, Tao Cheng, Lingxi Kong, Shuai Wang, and Bing Xu. 2020. "A Single Far-Field Deep Learning Adaptive Optics System Based on Four-Quadrant Discrete Phase Modulation" Sensors 20, no. 18: 5106. https://doi.org/10.3390/s20185106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop