Abstract
Fourier Ptychographic Microscopy (FPM) is a microscopy imaging technique based on optical principles. It employs Fourier optics to separate and combine different optical information from a sample. However, noise introduced during the imaging process often results in poor resolution of the reconstructed image. This article has designed an approach based on a residual local mixture network to improve the quality of Fourier ptychographic reconstruction images. By incorporating channel attention and spatial attention into the FPM reconstruction process, the network enhances the efficiency of the network reconstruction and reduces the reconstruction time. Additionally, the introduction of the Gaussian diffusion model further reduces coherent artifacts and improves image reconstruction quality. Comparative experimental results indicate that this network achieves better reconstruction quality, and outperforming existing methods in both subjective observation and objective quantitative evaluation.
1. Introduction
Spurred by the unceasing development of scientific and technological capabilities, the process of delving into the microscopic domain has become ever more in-depth. Researchers in biology, medicine, materials science, and other fields require more precise observation and measurement of microscopic structures to better understand their characteristics, functions, and interactions. Traditional microscopy techniques are often limited by the diffraction limit of light and sample scattering, which prevents them from providing high-resolution images. In 2013, Guoan Zheng et al. proposed a new quantitative imaging technique called Fourier Ptychographic Microscopy (FPM) [,]. Unlike traditional microscopes, this technique illuminates the sample with different incident angles sequentially and captures a set of low-resolution images with corresponding spatial spectra. The low-resolution intensity images obtained from different angles are then iteratively processed in the Fourier domain to solve for the optimal solution that satisfies both the spatial amplitude constraint and the frequency domain support constraint, thereby reconstructing the high-resolution image of the sample [,].
Despite its numerous advantages, FPM faces some challenges and limitations. The imaging speed is relatively slow, necessitating improved scanning and computational methods to enhance efficiency. The reconstruction process can also be affected by noise and system errors, requiring more effective denoising and optimization algorithms to improve image quality. Traditional FPM algorithms are mainly based on frequency domain processing and back-projection principles. They process the frequency domain data of the sample and then use back-projection algorithms to reconstruct the three-dimensional structure of the sample. For instance, Zuo et al. [] and Bian et al. [,] manually adjusted parameters and performed multiple iterations to improve noise robustness, accelerate the convergence of the reconstruction process, and enhance reconstruction results. However, traditional reconstruction algorithms take a long time due to multiple iterations. The neural network-based approach employs a large-scale pre-established dataset to train an end-to-end deep convolutional neural network, which is used to reconstruct high-resolution intensity and phase images. Jiang et al. [], Sun et al. [], and Zhang et al. [] solved for high-resolution intensity and phase images using the backpropagation of neural networks. However, neural network modeling methods [,,] are essentially gradient descent-based iterative methods and do not completely overcome the drawbacks of iterative algorithms, such as slow reconstruction speed and susceptibility to noise.
Although traditional neural network algorithms can reconstruct high-resolution images, they still have some shortcomings. During the course of recent years, image processing approaches powered by deep learning have advanced rapidly, gradually extending from fields like object recognition and image classification to image super-resolution and other image reconstruction areas [,,,,]. Dong et al. [] first applied deep neural networks to image super-resolution, surpassing a series of traditional algorithms, like sparse coding [] and neighbor embedding [], in image reconstruction quality. In Kim et al. [], the concept of residual learning was first introduced [,,], reconstructing the missing high-frequency residuals in low-resolution images to improve image resolution. Sun et al. [] put forth the double-flow convolutional neural network (DFNN) approach, which supplanted traditional iterative approaches to improve the quality of single wide-field reconstructed images, but this method could not be used for large-scale image reconstruction, limiting its scope. Sun et al. [] put forth the neural network model combined with pupil recovery (FINN-P) approach, which uses a more efficient workflow and a selection of dissimilar optimizers in the imaging network, achieving better reconstruction results than the neural network-based method by Jiang []. Moreover, deploying the proposed network on neural engines or tensor processing units (TPUs) has the potential to further enhance the reconstruction speed of FINN-P. Zhang et al. [] put forth the integration of neural network and physical reconstruction model (FuNN) method, fusing FPM’s physical reconstruction model with a convolutional neural network and optimizing the weights and biases of FuNN. Compared to FINN-P, FuNN does not require alternating training processes with different network settings, thus speeding up the reconstruction process. Zhang et al. [] proposed the physics-based learning with channel attention (PbNN-CA) method, integrating a physics-based network with a channel attention module (CA). Integrating the CA module into the physics-based network renders it capable of correcting pupil aberrations and LED intensity errors at the same time. Employing the channel attention module enhances the performance and noise resilience of PbNN. Our team published a Fourier ptychographic reconstruction method based on a residual hybrid attention network in 2023 [], combined channel attention with spatial attention, apportioning channel weights and procuring high-resolution spatial features founded on residual learning to ameliorate the quality of image reconstruction.
The article describes a Fourier ptychographic reconstruction solution, which was developed based on a residual local mixture network. By introducing a mixture attention mechanism and combining it with a Gaussian diffusion model, the quality of the reconstructed images generated by this network is vastly enhanced. The mixture attention mechanism allows the model to utilize both channel and spatial information simultaneously, while the Gaussian diffusion model simulates the diffusion process of images in the Fourier space, effectively handling noise and artifacts in microscopic images. By integrating the mixture attention mechanism with the Gaussian diffusion model, we can suppress noise interference while preserving image details, resulting in clearer microscopic reconstruction images. Comparative experimental analysis verifies that our approach attains considerable advancements in the quality and the efficiency of the reconstructed images. Our foremost contributions can be set forth as follows:
- To address the issues of poor reconstruction quality and low efficiency, a mixture attention network is introduced to optimize the reconstruction efficiency, reduce computational complexity, improve quality, and minimize time costs;
- To tackle the problems of noise and artifacts during the reconstruction process, a Gaussian diffusion model is introduced to simulate data diffusion, smoothing out noise, reducing coherent artifacts, and enhancing the quality and accuracy of reconstructed images.
2. Methods
2.1. Network Architecture
The residual local mixture network (RLMN) is primarily composed of three parts, as demonstrated in Figure 1. The shallow feature extraction module employs a 3 × 3 convolution to extract shallow features through convolution operations, maintaining texture and local detail information while diminishing the likelihood of overfitting. The information obtained from the shallow feature extraction module is subsequently fed into the deep feature extraction as the input. In the deep feature extraction stage, K local residual fusion groups are embedded to effectively extract and process high-frequency information in the images. Additionally, to enrich the semantic information in the images, high-frequency information is combined with low-frequency information through residual connections. This enhances the richness of semantic information contained in the image. Concluding the process, the reconstruction module deploys upsampling module and two 3 × 3 convolutions to execute the objective of reconstructing high-frequency information. Included among these, the input data consist of a series of low-resolution images. These low-resolution data are then synthesized in the Fourier domain, and transformed into a dual-channel data format through inverse Fourier transform. This dual-channel data are then used as the input to the network.
Figure 1.
Structure of residual local mixture network.
2.2. Residual Local Mixture Group
As demonstrated in Figure 1, inside the deep feature extraction block, the input is first directed into a convolution module comprising a 3 × 3 convolution and a LeakyReLU activation function. The convolution module can enhance the model’s feature representation capabilities through the mechanisms of convolution and activation functions. Afterwards, the model inputs the information into the Gaussian diffusion model, simulating the image diffusion process in the Fourier space to handle noise and artifacts in the image. Following a 1 × 1 convolution to reduce channel dimensions, the information is then fed into the hybrid attention module to further enhance the quality of the reconstructed images.
2.3. Mixture Attention Block
The spatial attention is enhanced by introducing channel attention into the spatial attention. First, the input is passed through the channel attention module, which distinctly characterizes the dependencies among channels to generate weights for each individual channel. These extracted weights are then applied to the feature maps for further processing. Subsequently, spatial attention focuses on the dependencies between different spatial positions in the feature maps to extract fine spatial features. The structure of the mixture attention block is portrayed in Figure 2.
Figure 2.
Structure of mixture attention block.
2.4. Enhanced Spatial Attention (ESA)
As demonstrated in Figure 3, the spatial attention technique first leverages a 1 × 1 convolutional layer to lower the channel dimension. Then, to expand the receptive field, a combination of stridden convolution (stride 2) and max pooling layers are leveraged to swiftly downscale the spatial dimensions of the network. To restore the spatial size, an upsampling layer is added, and a 1 × 1 convolutional layer is leveraged to reinstate the channel count. Finally, a sigmoid layer is used to generate the attention mask. Furthermore, skip connections are introduced, enabling the direct transfer of high-resolution features preceding the spatial dimension reduction to the output of the block. This design not only optimizes the efficiency and effectiveness of the network but also achieves effective processing of information at different scales through fine-grained structural adjustments, thereby improving the overall performance.
Figure 3.
Structure of enhanced spatial attention.
2.5. SE Layer
As demonstrated in Figure 4, the structure of the SE layer evaluates the importance of the distinct channels comprising the feature map and adjusts the channel weights based on these evaluations. The SE layer consists of two key operations: squeeze and excitation. First, the input information is passed through global average pooling to transform the feature map of each channel into a single real-valued quantity, reflecting the comprehensive information of that channel. This process captures the global receptive field of each channel, enabling the learning of dependencies between channels. Subsequently, a fully connected network (that comprises two dense layers and an activation function, with the first dense layer reducing the dimensions and the second restoring the original channel count) processes the real number sequence obtained from the squeeze operation. The fully connected network learns the importance of each channel and outputs a weight vector corresponding to the number of input channels. Finally, the SE layer multiplies the weight vector obtained from the excitation operation with the original feature map on a per-channel basis, thus adjusting the weights of the distinct channel features. The SE layer helps the network focus better on important feature channels and suppress less important ones, thereby improving the network’s performance.
Figure 4.
Structure of SE layer.
2.6. Diffusion Model
The diffusion model evolves from an initial state to a target distribution through a series of iterative processes. It can be seen as a random walk in probability space to simulate the target distribution. The diffusion model is significant in image-generation tasks. Leveraging its ability to learn the probability distribution of images can produce images that are of high quality and exhibit a high degree of realism. The diffusion model comprises two processes: “noising” and “denoising.” As shown in Figure 5, during the noising process, input X0 constantly mixes with Gaussian noise. After T iterations of noising, the image XT becomes a pure noise image following a standard normal distribution. During the denoising process, the network learns T denoising steps to restore XT to X0.
Figure 5.
Diffusion model, with the noising process above and the denoising process below.
2.7. Evaluation Metrics
As a method of image reconstruction, Fourier ptychographic imaging directly reflects the effectiveness of the reconstruction algorithm as reflected in its reconstruction outcomes. The quality of image reconstruction is evaluated in this paper using the Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) as the assessment metrics.
The Peak Signal-to-Noise Ratio (PSNR) in signal processing denotes the proportion between the maximum achievable signal power and the noise power that impacts the fidelity of the signal representation. In image evaluation, it can be defined by the Mean Squared Error (MSE):
The mean square error can be expressed as
where the variable equates to the authentic image, and the variable equates to the reconstructed image; a PSNR value that is more elevated corresponds to better image quality.
The Structural Similarity Index (SSIM) is given by:
where and are the mean and standard deviation of 1 and 2, respectively; is the covariance of the two; and is a constant.
3. Experiment
3.1. Dataset
This experiment employs a simulated dataset encompassing 15,000 sets of high-resolution image data. Every set is composed of two images, one representing the intensity channel and the other the phase channel. The synthesized complex amplitude data from the two input channels are simulated using the Fourier imaging system. As part of the simulation process, Gaussian noise with a mean of 0 and a standard deviation of 3 × 10−4 is incorporated to emulate potential system error noise encountered in real-world imaging scenarios. These Gaussian noise-added data are regarded as low-resolution data pertinent to Fourier ptychographic imaging. A conventional Fourier ptychographic reconstruction algorithm, executed with a solitary iteration, is subsequently applied to combine these low-resolution data into low-resolution complex amplitudes. This produced 15,000 sets of low-resolution input data. Ultimately, the 15,000 sets of high-resolution image data were reserved to be used as the reference images for subsequent performance evaluation and comparative analysis. The architectural design of the Fourier ptychographic microscopy system is depicted in Figure 6.
Figure 6.
Fourier ptychographic microscope: (a) real FPM system; (b) FPM simulation schematic.
3.2. Optimizer Comparison Experiment
To achieve better optimization results for the experiment while keeping the loss function and number of iterations the same, Adagrad, Adamax, and AdamW optimizers were chosen for comparison, respectively. The comparison of loss curves for different optimizers is shown in Figure 7. The number of iterations is shown on the horizontal axis, and the loss value for each training session is plotted on the vertical axis of this figure. The three curves represent the three different optimizers. From the figure, it is evident that the model trained using the AdamW optimizer demonstrates quicker loss decrease and superior convergence relative to the other two optimizers. The reconstruction findings are presented in Figure 8, where HR represents the original high-resolution image, and LR corresponds to its corresponding low-resolution image. It is discernible that the intensity images reconstructed by all three optimizers achieve good results, but the network trained with the AdamW optimizer produces the highest-quality phase images, followed by Adamax, with Adagrad yielding the lowest quality. Tabulated in Table 1 are the quantitative metrics, SSIM, and PSNR, for the reconstruction images generated by the three optimizers. As can be observed from the table, the AdamW optimizer significantly outperforms the other two optimizers in image quality. For intensity images, the PSNR of the AdamW reconstruction results is higher than the other two optimizers by 0.96 and 6.91, respectively. For phase images, the SSIM of the AdamW reconstruction results is higher than the other two optimizers by 0.0676 and 0.43, respectively. This comparative experiment serves to substantiate the effectiveness of the AdamW optimizer network.
Figure 7.
Loss curve comparison for different optimizers.
Figure 8.
Reconstruction results for different optimizers.
Table 1.
Quantitative results for different optimizers.
3.3. Ablation Experiment
In this section, we conducted ablation experiments to validate the effectiveness of the Fourier ptychographic reconstruction method combining hybrid attention and Gaussian diffusion models. Among them, RLDN is the network with the standalone addition of the Gaussian diffusion model, RLMAN is the network with the standalone addition of the hybrid attention mechanism, RLMN is the residual local mixture network. Keeping the loss, learning rate, and noise parameters uniform, the network is trained using the three models. As presented in Figure 9, three model networks achieve good convergence performance, but the loss value for the mixture network remains stable throughout the descent in contrast to the other two networks. The reconstruction results generated by the three models are depicted in Figure 10. It is clear that all three models achieve good reconstruction results, but the RLMN model produces the reconstructions of the quality that are closest in resemblance to the ground truth. Table 2 presents the quantitative results of the reconstructed images’ metrics of SSIM and PSNR. The information presented in the table reveals the RLMN model’s superior performance over the other two models in both intensity and phase image metrics, demonstrating the effectiveness of the RLMN model.
Figure 9.
Loss curve comparison for different models.
Figure 10.
Reconstruction results for different models.
Table 2.
Quantitative results for different models.
3.4. Comparative Experiment under Identical Noise Conditions
In the context of FPM image acquisition, the image quality can be compromised by a range of variables, like instrumentation and illumination, resulting in noise. To simulate these real-world conditions, this section employs Gaussian noise with a mean of 0 and a standard deviation of 3 × 10−4 as the primary interference condition to additionally validate the efficacy of the presented approach. The traditional phase recovery methods A-S [] and G-S [], the neural network-based reconstruction method proposed by Jiang et al. [], the INNM method proposed by Zhang et al. [], and the SwinIR method proposed by Wang et al. [] were used as comparative algorithms against our proposed reconstruction method in this study. A trio of image sets are randomly picked from the test collection. The reconstruction outcomes of disparate approaches under identical noise conditions are illustrated in Figure 11. Table 3 presents the quantitative performance of the distinct reconstruction approaches using SSIM and PSNR evaluation. The results shown in Figure 11 and Table 3 indicate that the reconstruction approach presented in this article demonstrates a better reconstruction effect and performance metrics. The reconstructed images have higher clarity compared to the phase reconstruction results of the G-S, A-S, and Wang et al. methods. Relative to the methods of Jiang et al. and Zhang et al., the reconstruction results in this work do not exhibit significant artifacts, and they contain more textural details while reducing errors. The bolded values in Table 3 indicate the optimal results. In terms of the reconstruction metric values, the method in this chapter outperforms the other five approaches.
Figure 11.
Reconstruction results under identical noise situations.
Table 3.
Quantitative results of reconstruction under identical noise situations.
3.5. Comparative Experiment under Different Noise Conditions
In real-world reconstruction scenarios, the noise levels can vary. Therefore, this section simulates the reconstruction outcomes and performance metrics of identical images under dissimilar noise situations. The noise magnitudes employed are 1 × 10−4, 2 × 10−4, and 3 × 10−4. A random selection of images is taken from each of the three test datasets. Figure 12 showcases the reconstruction outcomes of the identical amplitude and phase images under dissimilar noise situations, and Table 4 presents the quantitative reconstruction outcomes of the identical amplitude and phase images under dissimilar noise scenarios. The results presented in Figure 12 and Table 4 indicate that the proposed reconstruction method demonstrates superior results compared to the other five reconstruction methods, proving the effectiveness of this model.
Figure 12.
Reconstruction results comparison under dissimilar noise situations.
Table 4.
Quantitative results of reconstruction under dissimilar noise situations.
3.6. Comparative Experiment on Real Images
To validate the performance of the proposed network on real data, real picture data collected using the FPM system is combined with the simulated dataset to train the network. This section uses traditional algorithms A-S [], G-S [], and the techniques of Jiang et al. [], Zhang et al. [], and Wang et al. [] as comparative experiments. The reconstruction results using actual data are illustrated in Figure 13. It is clear from the figure that the proposed method maintains better reconstruction performance and visual quality when tested on real data. Unlike the other five approaches, the approach presented in this work is able to effectively remove the coherent artifacts generated during the imaging process while also demonstrating higher clarity and more distinct textural details. Table 5 presents the time metrics for the reconstruction results of real data. As indicated by the table, the proposed approach requires the shortest reconstruction time compared to the other five approaches, with SwinIR being the second shortest. Although the A-S and G-S approaches have relatively short reconstruction times, their phase image quality is poor. In contrast, the approaches of Jiang et al. and Zhang et al. produce high-quality reconstruction images but require longer reconstruction times. These experimental results confirm the proposed approach’s superior performance.
Figure 13.
Reconstruction results of real data.
Table 5.
Time Metrics for real data reconstruction.
4. Conclusions
In this article, constructed upon a residual local mixture network architecture, a reconstruction approach for Fourier ptychographic microscopy is put forward. In order to address the issues of poor reconstruction quality and artifacts in traditional reconstruction algorithms, the network incorporates a mixture attention mechanism and the Gaussian diffusion model. Continuous training of the network aims to achieve high-resolution image reconstruction. Through the experimental investigation, the effectiveness of the proposed approach is proven using datasets under identical noise, dissimilar noise, and real data. By comparing traditional algorithms and neural network-based methods, the proposed method’s efficacy is proven. Quantitative analysis of the reconstruction results is performed using SSIM, PSNR metrics, and time from different perspectives.
Author Contributions
Conceptualization, Y.W. (Yan Wang), Y.W. (Yongshan Wang) and J.L.; methodology, Y.W. (Yan Wang) and Y.W. (Yongshan Wang); software, Y.W. (Yongshan Wang); validation, Y.W. (Yongshan Wang); investigation, Y.W. (Yongshan Wang) and X.W.; writing—original draft preparation, Y.W. (Yongshan Wang); writing—review and editing, Y.W. (Yan Wang) and Y.W. (Yongshan Wang); supervision, Y.W. (Yan Wang), X.W. and J.L.; project administration, Y.W. (Yan Wang) and Y.W. (Yongshan Wang) contributed equally to this paper. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Science and Technology Development Plan Projects of Jilin Province [NO: YDZJ202301ZYTS180].
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data sharing is not applicable to this article.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Zheng, G.; Horstmeyer, R.; Yang, C. Wide-field, high-resolution Fourier ptychographic microscopy. Nat. Photonics 2013, 7, 739–745. [Google Scholar] [CrossRef]
- Zheng, G.; Shen, C.; Jiang, S.; Song, P.; Yang, C. Concept, implementations and applications of Fourier ptychography. Nat. Rev. Phys. 2021, 3, 207–223. [Google Scholar] [CrossRef]
- Pan, A.; Zuo, C.; Yao, B. High-resolution and large field-of-view Fourier ptychographic microscopy and its applications in biomedicine. Rep. Prog. Phys. 2020, 83, 096101. [Google Scholar] [CrossRef]
- Guo, K.; Dong, S.; Zheng, G. Fourier ptychography for brightfield, phase, darkfield, reflective, multi-slice, and fluorescence imaging. IEEE J. Sel. Top. Quantum Electron. 2015, 22, 77–88. [Google Scholar] [CrossRef]
- Zuo, C.; Sun, J.; Chen, Q. Adaptive step-size strategy for noise-robust Fourier ptychographic microscopy. Opt. Express 2016, 24, 20724–20744. [Google Scholar] [CrossRef]
- Bian, L.; Suo, J.; Zheng, G.; Guo, K.; Chen, F.; Dai, Q. Fourier ptychographic reconstruction using Wirtinger flow optimization. Opt. Express 2015, 23, 4856–4866. [Google Scholar] [CrossRef]
- Bian, L.; Suo, J.; Chung, J.; Ou, X.; Yang, C.; Chen, F.; Dai, Q. Fourier ptychographic reconstruction using Poisson maximum likelihood and truncated Wirtinger gradient. Scientific reports, 2016, 6, 27384. [Google Scholar] [CrossRef]
- Jiang, S.; Guo, K.; Liao, J.; Zheng, G. Solving Fourier ptychographic imaging problems via neural network modeling and TensorFlow. Biomed. Opt. Express 2018, 9, 3306–3319. [Google Scholar] [CrossRef]
- Sun, J.; Chen, Q.; Zhang, Y.; Zuo, C. Efficient positional misalignment correction method for Fourier ptychographic microscopy. Biomed. Opt. Express 2016, 7, 1336–1350. [Google Scholar] [CrossRef]
- Zhang, J.; Xu, T.; Zhang, Y.; Chen, Y.; Wang, S.; Wang, X. Multiplex Fourier ptychographic reconstruction with model-based neural network for Internet of Things. Ad Hoc Netw. 2021, 111, 102350. [Google Scholar] [CrossRef]
- Wang, X.; Piao, Y.; Jin, Y.; Li, J.; Lin, Z.; Cui, J.; Xu, T. Fourier Ptychographic Reconstruction Method of Self-Training Physical Model. Appl. Sci. 2023, 13, 3590. [Google Scholar] [CrossRef]
- Zhang, Y.; Liu, Y.; Jiang, S.; Dixit, K.; Song, P.; Zhang, X.; Ji, X.; Li, X. Neural network model assisted Fourier ptychography with Zernike aberration recovery and total variation constraint. J. Biomed. Opt. 2021, 26, 036502. [Google Scholar] [CrossRef]
- Chen, Q.; Huang, D.; Chen, R. Fourier ptychographic microscopy with untrained deep neural network priors. Opt. Express 2022, 30, 39597–39612. [Google Scholar] [CrossRef]
- Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Tejani, A.; Totz, J.; Wang, Z.; Shi, W. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
- Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Loy, C.C.; Qiao, Y.; Tang, X. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
- Choi, H.; Lee, J.; Yang, J. N-gram in swin transformers for efficient lightweight image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 2071–2081. [Google Scholar]
- Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16. Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 191–207. [Google Scholar]
- Wang, H.; Chen, X.; Ni, B.; Liu, Y.; Liu, J. Omni aggregation networks for lightweight image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 22378–22387. [Google Scholar]
- Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef]
- Huang, S.; Sun, J.; Yang, Y.; Fang, Y.; Lin, P.; Que, Y. Robust Single-Image Super-Resolution Based on Adaptive Edge Preserving Smoothing Regularization. IEEE Trans. Image Process. 2018, 27, 2650–2663. [Google Scholar] [CrossRef]
- Gao, X.; Zhang, K.; Tao, D.; Li, X. Image Super-Resolution with Sparse Neighbor Embedding. IEEE Trans. Image Process. 2012, 21, 3194–3205. [Google Scholar]
- Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1646–1654. [Google Scholar]
- Musunuri, Y.R.; Kwon, O.S. Deep Residual Dense Network for Single Image Super-Resolution. Electronics 2021, 10, 555. [Google Scholar] [CrossRef]
- Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
- Zhang, Y.; Li, K.; Li, K.; Zhong, B.; Fu, Y. Residual non-local attention networks for image restoration. arXiv 2019, arXiv:1903.10082. [Google Scholar]
- Sun, M.; Shao, L.; Zhu, Y.; Zhang, Y.; Wang, S.; Wang, Y.; Diao, Z.; Li, D.; Mu, Q.; Xuan, L. Double-flow convolutional neural network for rapid large field of view Fourier ptychographic reconstruction. J. Biophotonics 2021, 14, e202000444. [Google Scholar] [CrossRef]
- Sun, M.; Chen, X.; Zhu, Y.; Li, D.; Mu, Q.; Xuan, L. Neural network model combined with pupil recovery for Fourier ptychographic microscopy. Opt. Express 2019, 27, 24161–24174. [Google Scholar] [CrossRef]
- Zhang, J.; Tao, X.; Yang, L.; Wang, C.; Tao, C.; Hu, J.; Wu, R.; Zheng, Z. The integration of neural network and physical reconstruction model for Fourier ptychographic microscopy. Opt. Commun. 2022, 504, 127470. [Google Scholar] [CrossRef]
- Zhang, J.; Xu, T.; Li, J.; Zhang, Y.; Jiang, S.; Chen, Y.; Zhang, J. Physics-based learning with channel attention for Fourier ptychographic microscopy. J. Biophotonics 2022, 15, e202100296. [Google Scholar] [CrossRef]
- Li, J.; Hao, J.; Wang, X.; Wang, Y.; Wang, Y.; Wang, H.; Wang, X. Fourier Ptychographic Microscopic Reconstruction Method Based on Residual Hybrid Attention Network. Sensors 2023, 23, 7301. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).