Next Article in Journal
Generation of 27 nm Spectral Bandwidth, Two-Port Output Pulses Directly from a Yb-Doped Fiber Laser
Previous Article in Journal
Photonic–Surface Plasmon Coupling Mode: Experimental Study with a Silver Thin-Film Coating on MPCC
Previous Article in Special Issue
Curved Holographic Augmented Reality Near-Eye Display System Based on Freeform Holographic Optical Element with Extended Field of View
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Application of Deep Learning in the Phase Processing of Digital Holographic Microscopy

1
School of Electrical Engineering and Electronic Information, Xihua University, Chengdu 610039, China
2
Sichuan Provincial Key Laboratory of Signal and Information Processing, Xihua University, Chengdu 610039, China
*
Author to whom correspondence should be addressed.
Photonics 2025, 12(8), 810; https://doi.org/10.3390/photonics12080810
Submission received: 12 July 2025 / Revised: 8 August 2025 / Accepted: 11 August 2025 / Published: 13 August 2025
(This article belongs to the Special Issue Holographic Information Processing)

Abstract

Digital holographic microscopy (DHM) provides numerous advantages, such as noninvasive sample analysis, real-time dynamic detection, and three-dimensional (3D) reconstruction, making it a valuable tool in fields such as biomedical research, cell mechanics, and environmental monitoring. To achieve more accurate and comprehensive imaging, it is crucial to capture detailed information on the microstructure and 3D morphology of samples. Phase processing of holograms is essential for recovering phase information, thus making it a core component of DHM. Traditional phase processing techniques often face challenges, such as low accuracy, limited robustness, and poor generalization. Recently, with the ongoing advancements in deep learning, addressing phase processing challenges in DHM has become a key research focus. This paper provides an overview of the principles behind DHM and the characteristics of each phase processing step. It offers a thorough analysis of the progress and challenges of deep learning methods in areas such as phase retrieval, filtering, phase unwrapping, and distortion compensation. The paper concludes by exploring trends, such as ultrafast 3D holographic reconstruction, high-throughput holographic data analysis, multimodal data fusion, and precise quantitative phase analysis.

1. Introduction

The concept of digital holography was first introduced by Professor Goodman in 1967, marking the beginning of holographic technology development [1]. However, in the 1970s and 1980s, digital holography faced significant challenges in realizing its full potential due to constraints in imaging hardware and computational power. In the 1990s, the development of high-resolution charge-coupled devices (CCDs) allowed CCDs to replace traditional photosensitive media, such as photographic and laser plates, for recording holograms. This shift enabled the use of digital methods to simulate optical diffraction, facilitating the full digitalization of both hologram recording and reconstruction.
Digital holographic microscopy (DHM) is a technology that combines traditional optical microscopy with holographic phase imaging techniques [2]. It provides benefits, such as noninvasive sample analysis, real-time dynamic detection, and three-dimensional (3D) reconstruction [3]. Furthermore, DHM allows for the quantitative measurement of a sample’s 3D morphological characteristics. This technology has found widespread use in fields such as biomedical research [4,5], cell mechanics studies [6], and environmental monitoring [7].
The goal of phase processing is to recover the complete phase information of an object from incomplete data, thereby obtaining its 3D shape and characteristics. However, traditional phase processing methods in DHM face several challenges, such as high computational complexity, poor resistance to noise, phase ambiguity, and limited real-time performance. Additionally, as the demand for higher resolution, increased imaging precision, and 3D imaging grows, issues such as larger data volumes, more complex scenes, and diverse equipment have become more prominent.
As noted by Rivenson et al., recent advancements in deep learning have led to significant progress in holography and coherent imaging by effectively addressing their inherent challenges while preserving their fundamental benefits [8]. Unlike traditional methods, deep learning-based phase processing techniques do not require prior knowledge and can automatically manage complex, nonlinear relationships. These methods have been progressively applied to various steps, such as phase retrieval [9], phase filtering [10], phase unwrapping [11], and phase distortion compensation [12]. Recently, the use of deep learning in solving phase processing challenges in DHM has become a major research focus in the field.
Deep learning, inspired by the structure and function of the human brain, utilizes deep neural networks composed of multiple layers of interconnected neurons through which information is propagated [13].
This paper provides a brief overview of the fundamental principles of DHM and emphasizes its potential advantages in addressing the challenges faced by traditional methods. It also examines the application of current deep learning methods in phase retrieval, phase filtering, phase unwrapping, and phase distortion compensation, while comparing the performance of various deep learning approaches. Furthermore, the paper highlights the limitations of existing methods and offers perspectives on future developments in the field, such as ultrafast 3D holographic reconstruction, high-throughput holographic data analysis, precise quantitative phase analysis, and multimodal data fusion. The aim is to offer valuable insights for advancing phase processing techniques in DHM.

2. Principles of Digital Holographic Microscopy

2.1. Theoretical Foundations of Digital Holographic Microscopy

The evolution of quantitative phase imaging (QPI) technology began with the introduction of phase contrast microscopy in 1942, followed by a major breakthrough in holography in 1962, and the further development of digital holography in 1999. Over time, QPI has become a highly accurate imaging technique that is now widely applied in various fields, such as biomedical research and materials science.
QPI addresses the challenge of phase information loss in optical imaging by quantitatively measuring the changes in diffraction patterns or interference fringes caused by variations in the refractive index, thus extracting the phase information of the object [14].
Holographic microscopy (HM), a prominent technique within the QPI framework, works by combining a reference wave with an object wave to create interference, converting phase information into intensity, which is then captured in a hologram. When a reconstruction wave is used to illuminate the hologram, the phase and amplitude distribution of the object can be restored. With the development of CCDs and complementary metal-oxide-semiconductor (CMOS) sensors, hologram recording shifted from photographic plates to digital methods. The interference pattern is projected onto a high-resolution CCD or CMOS sensor for capture, and digital signal processing techniques are applied for decoding and reconstruction, allowing for the recording and reproduction of 3D object images [15].

2.1.1. Digital Holographic Microscopy Recording

The measurement principle of digital holography follows the same basic concept as traditional optical holography, with the key difference being that digital holography utilizes an electronic imaging sensor instead of a traditional holographic plate to record a digital hologram. Figure 1 illustrates the optical path for recording a digital hologram. In this setup, X’0’Y’ represents the object plane, X0Y signifies the recording plane where the CCD photosensitive surface is located, and d indicates the recording distance of the digital holography system. When the object wave and reference wave are illuminated simultaneously, the hologram is captured by the CCD based on the principle of light interference, thereby achieving the digitalization of the hologram.
Let the object wave  O x , y , generated on the object plane by the illumination of the observed object, and the reference wave  R x , y , be represented by the following mathematical expressions:
O x , y = A 0 x , y exp i φ 0 x , y
R x , y = A R x , y exp i k   sin   θ y
where  A 0  and  A R  denote the amplitudes of the object wave and reference wave, respectively;  k = 2 π / λ , where  λ  is the light’s wavelength; and θ is the angle between the reference wave and the z-axis.
The intensity distribution of the interference pattern created by the two waves at the recording plane is expressed as
I x , y = O x , y 2 + R x , y 2 + O x , y R * x , y + O * x , y R x , y
where  O * x , y  and  R * x , y  are the complex conjugates of the object wave and reference wave, respectively. The CCD captures the interference light intensity and converts it into an electrical signal. The structural schematic of the CCD is presented in Figure 2. Let the dimensions of the CCD’s photosensitive surface be  L x  and  L y , the pixel size be  x × y , the number of pixels be  M x × M y , and the pixel interval be  d x × d y , where  M x = L x / x  and  M y = L y / y . The discretized interference intensity recorded by the CCD can then be expressed as
I p , q = I H x , y rect x L x , y L y p = M x 2 M x 2 1 q = M y 2 M y 2 1 δ x p Δ x , y q Δ y
r e c t ( t ) = { 1 ,   if t 0.5 0 ,   if t > 0.5
where p and q refer to integers within specified ranges  M x / 2 p M x / 2 1  and  M y / 2 q M y / 2 1 , and  δ ( x , y )  signifies a two-dimensional Dirac delta function, which is an extension of the one-dimensional Dirac delta function,  δ ( x ) . The one-dimensional Dirac delta function,  δ ( x ) , is defined as  δ ( x ) = { ,   if   x = 0 0 ,   if   x 0 . The two-dimensional Dirac delta function,  δ ( x , y ) , can be formulated as the product of two one-dimensional delta functions:  δ ( x , y ) = δ ( x ) δ ( y ) . Finally, the hologram recorded by the CCD is stored as a numerical matrix in the computer.

2.1.2. Digital Holographic Microscopy Reconstruction

The digital holography reconstruction process involves digitizing the recorded interference image and inputting it into a computer system. The computer then simulates the illumination of the hologram by light waves and the diffraction of emitted light reaching the observation plane. This allows for the calculation of both the intensity and phase information of the object being measured. Figure 3 illustrates the hologram reconstruction process, where the hologram is positioned in the XOY plane, and the reconstructed image is in the X″O″Y″ plane. The distance d1 represents the reconstruction distance.
Let the complex amplitude of the light wave after illuminating the hologram be denoted as  U ( x , y ) . During the reconstruction process, a numerical reconstruction algorithm is employed to obtain the complex amplitude of the object wave at the image plane, denoted as  U 1 ( x 1 , y 1 ) . From the complex amplitude distribution, the intensity  I 1 x 1 , y 1  and phase  Φ 1 x 1 , y 1  of the reconstructed image can then be calculated as follows:
I 1 x 1 , y 1 = U 1 x 1 , y 1 U 1 * x 1 , y 1
Φ 1 x 1 , y 1 = arctan Im U 1 x 1 , y 1 Re U 1 x 1 , y 1
where  Im U 1 x 1 , y 1  and  Re U 1 x 1 , y 1  denote the imaginary and real parts of  U 1 x 1 , y 1 , respectively. Since the phase calculated using the arctangent function is wrapped within the range [ π , π ], phase unwrapping is necessary to recover the true phase of the object.

2.2. Signal Changes in Phase Processing

Phase processing is a critical aspect of DHM, encompassing several key steps, such as phase recovery, filtering, phase unwrapping, and distortion compensation. These processes together guarantee the quality and precision of the imaging results.
As illustrated in Figure 4, which presents a flowchart for phase processing steps in DHM, the red dashed section is designed as phase extraction. This phase extraction process involves three main steps: phase filtering, phase unwrapping, and phase distortion compensation.
In the phase recovery step, the intensity  I H ( x , y )  of the recorded interference light field is extracted using Equation (3). DHM algorithms, Fourier transform, or deep learning techniques are then applied to recover the phase information, allowing the complex amplitude  U ( x , y )  to be computed, which encompasses both intensity and phase. The phase that needs to be recovered is denoted as  φ ( x , y ) .
U x , y = O x , y e i φ x , y
In the phase filtering step, the process involves applying a Fourier transform to convert the complex amplitude  U ( x , y )  into the frequency domain, performing the filtering operation, and subsequently transforming it back to the spatial domain. Suppose the complex amplitude  U ( x , y )  is processed with the frequency domain filter  H ( f x , f y ) , resulting in  U 2 x , y .
U 2 x , y = F 1 F U x , y H f x , f y
where  F U x , y  implies the frequency domain representation of  U x , y  and  H f x , f y  implies the frequency domain filter.  F 1  denotes the inverse Fourier transform.
After phase filtering, the updated phase information  φ ( x , y )  can be extracted from  U 2 x , y , and its expression is denoted as
U 2 x , y = O x , y e i φ x , y
where  O x , y  represents the intensity information of the object wave after filtering, and  φ ( x , y )  denotes the phase information after filtering.
During phase unwrapping, the unwrapped phase information is given by
Φ x , y = φ x , y + 2 π N x , y
where  N x , y  signifies an integer function, indicating the integer multiple of 2π that should be added during the phase unwrapping process. The complex amplitude following phase unwrapping is expressed as
    U 3 x , y = O x , y e i Φ x , y
In the phase distortion compensation step, assuming that the unwrapped phase  Φ x , y  contains a certain known phase distortion  Δ φ ( x , y ) , the phase information after distortion compensation,  Φ x , y , can be expressed as
Φ x , y = Φ x , y Δ φ x , y
The complex amplitude after distortion compensation can be expressed as
U 4 x , y = O x , y e i ( Φ x , y Δ φ x , y + Δ ω x , y )
where  Δ ω x , y  denotes the estimated aberration.
Following the above steps, the phase complex amplitude can be represented as
U x , y = O x , y e i Φ x , y Δ φ x , y + Δ ω x , y + Δ Φ x , y
where  Δ Φ ( x , y )  indicates the phase information derived from the difference between the phase after distortion compensation and the original signal during the phase processing. Finally, the real phase  Φ x , y  obtained through phase processing is
Φ x , y = Φ x , y Δ φ x , y + Δ ω x , y + Δ Φ x , y = Φ 1 x 1 , y 1

2.3. Theoretical Framework for Phase Processing in Deep Learning

Figure 5 introduces a structured, three-layer classification framework designed to clarify how deep learning techniques are applied to DHM phase processing across three dimensions: the application scenario layer (e.g., biomedical, materials science), the technical principle layer (data-driven, model-driven, hybrid), and the learning paradigm layer (supervised, unsupervised, self-supervised learning).
Additionally, the theoretical framework for deep learning-based phase processing is structured into four hierarchical layers: the application scenario, technical principle, learning paradigm, and algorithm implementation.

2.3.1. Application Scenario Dimension

From the perspective of application scenarios, deep learning-based phase processing in DHM serves distinct roles across fields. In biomedical research, it enables label-free, dynamic analysis of biological samples by suppressing phase noise and reconstructing 3D structures for high-resolution, long-term live-cell tracking.
In materials science, deep learning-based phase processing facilitates the extraction of microstructural features, such as crystal orientations and interfacial defects, supporting process optimization and performance prediction for advanced materials.
While both fields rely on precise phase analysis at the microscale, biomedical applications emphasize non-destructive monitoring of dynamic processes, whereas materials science focuses on high-throughput analysis of static structure–property relationships.

2.3.2. Technical Principle Dimension

From the perspective of technical principles, three paradigms have been adopted in DHM phase processing using deep learning. In the data-driven paradigm, models are trained on large annotated datasets to automatically learn phase reconstruction mapping relationships from raw holograms. This approach, reliant on data volume and model capacity, is suited for data-rich scenarios involving complex, non-explicit physical mechanisms.
In the physics model-driven paradigm, optical priors, such as light wave propagation and diffraction theory, are embedded into model structures or loss functions. By incorporating physical constraints or differentiable simulation modules, model outputs are aligned with underlying physical laws. This paradigm offers greater interpretability and generalization, particularly in data-limited cases with well-defined physics.
The hybrid paradigm combines both strategies. Deep learning is used to model nonlinearities and noise, while physical constraints are introduced—e.g., via wave equation regularization terms based on wave equations to neural networks. This method enhances phase reconstruction accuracy while preserving physical consistency, making it well-suited for biomedical and materials science applications demanding both high precision and interpretability.

2.3.3. Learning Paradigm Dimension

The characteristics of the three learning paradigms are discussed in what follows.
Supervised learning involves training models on large volumes of annotated data to directly learn the mapping between inputs and corresponding outputs. The accuracy of such models is highly dependent on both the quality and quantity of the annotations provided. In applications requiring phase continuity constraints, loss functions are employed to enforce consistency between the model’s output and the continuous phase of the target labels.
Unsupervised learning performs phase unwrapping by harnessing the intrinsic structure of the data, thereby eliminating the need for annotated datasets. This approach is particularly suitable for scenarios where annotation is costly or prior domain knowledge is scarce. Phase continuity constraints are implemented by designing prior constraints or leveraging the intrinsic structure of the data.
Self-supervised learning automatically circumvents manual annotation by constructing supervisory signals through proxy tasks designed directly from unlabeled data. Incorporating physical priors improves the model’s generalization to unseen samples. Phase continuity constraints are enforced either through transformations—such as rotation or occlusion—that require consistent phase predictions, or by leveraging phase correlation across multi-frame images or data from different sensors.

3. Application of Deep Learning in Phase Recovery of Digital Holographic Microscopy

Phase information is essential for reconstructing the 3D structure and complex details of an object. However, direct measurement of phase information from a hologram is not possible. Phase retrieval techniques reconstruct the object’s phase from the recorded holographic intensity, forming the foundation for subsequent phase reconstruction and 3D imaging from indirect or incomplete data. This process enhances image quality and contrast while enabling the analysis of internal structures and dynamic changes in the object. Traditional phase retrieval methods, such as the Gerchberg–Saxton algorithm [16] and the Fienup algorithm [17], have robust anti-interference capabilities. However, they face the following limitations:
(1)
They rely on iterative calculations, which are time-consuming and inefficient.
(2)
The complex operational processes lead to lower-quality reconstructed images.
(3)
They require that the intensity and phase be processed separately in sequential steps, as they lack an integrated framework for efficient joint reconstruction.
To address the issue (1), Zhang et al. from Tsinghua University proposed a model in 2018 based on a U-shaped (U-Net) convolutional neural network (CNN) [18]. This model incorporates the U-Net structure, enabling faster recovery of the original phase of the sample and replacing traditional iterative methods, thus reducing computation time. However, this study did not provide specific evaluation metrics.
To address the issue (2), Yuki Nagahama from the Tokyo University of Agriculture and Technology introduced a phase retrieval method using a U-Net network model in 2022 [19]. Figure 6 depicts the U-Net CNN model. This approach can process a single hologram, significantly simplifying the process, effectively removing overlapping conjugate images, and enhancing the quality of reconstructed images. In terms of the evaluation metric mean absolute error (MAE), this method reduces the error by 0.058–0.069 π compared to traditional methods. However, comparisons of this method to other deep learning models were not provided.
To address the issue (3), Wang et al. from Northwestern Polytechnic University proposed the one-to-two Y-Net network in 2019 [20]. This network was developed to simultaneously reconstruct both intensity (light intensity distribution) and phase information from a single digital hologram. However, the Y-Net network requires three rounds of training, which affects its training efficiency and necessitates further optimization.
In 2023, Chen et al. from the North China University of Science and Technology proposed the Mimo-Net network [21], which requires only a single training session to simultaneously reconstruct intensity and phase information at three different scales. Compared to Y-Net [20], Mimo-Net offers superior reconstruction performance in terms of image quality. Figure 7 compares the reconstruction results of Mimo-Net and Y-Net at the same scale of 256 × 256.
To provide a structured overview of phase processing research, the subsequent analysis is organized across multiple dimensions. Table 1 classifies existing phase recovery networks, highlighting structural design principles that impact phase recovery performance. Table 2 compares the core mechanisms, advantages, and disadvantages of phase recovery algorithms, distilling key insights to validate algorithmic applicability across diverse scenarios.

4. Application of Deep Learning in Phase Extraction for Digital Holographic Microscopy

Figure 8 illustrates the process of phase extraction in DHM, which involves three key steps: phase filtering, phase unwrapping, and phase distortion compensation. The phase extraction process begins with phase retrieval to acquire the complex amplitude image. This image is then subjected to phase filtering to produce the wrapped phase map. Subsequently, phase unwrapping is performed on the wrapped phase map to generate the phase unwrapping map. Finally, phase distortion compensation is applied to the unwrapped phase map to obtain the final real phase information.

4.1. Application of Deep Learning in Phase Filtering

Phase information obtained through phase retrieval can be affected by high-frequency noise or artifacts due to environmental noise, optical system limitations, or computational errors. These noises can cause the phase images to become unstable or distorted. Phase filtering helps remove such unwanted noise, smooth the phase map, and optimize its spatial distribution, effectively reducing artifacts and minimizing imaging errors caused by noise.
Traditional speckle noise filtering methods are divided into spatial domain, transform domain, and deep learning-based techniques, based on their theoretical foundations and processing stages. Spatial domain filtering methods include median filtering [22], mean filtering [23], blind convolution filtering [24], block matching [25], and 3D filtering [26], while transform domain methods include Wiener filtering [27], windowed Fourier transform [28], and wavelet filtering [29]. Although these methods classify and process objects, several challenges remain in practice:
(1)
Difficulty in obtaining noise-free phase maps for labeling.
(2)
Unclear feature representation.
(3)
Inefficient in denoising.
Recent studies have shown that deep learning approaches enable robust single-image super-resolution without iterative processing, greatly enhancing the speed of high-resolution hologram acquisition and reconstruction.
To address the issue (1), Wu et al. from Guangdong University of Technology proposed a network comprising a deep bilateral network (DBSN) and a non-local feature network (NLFN) in 2022 [30]. The framework of the DBSN and NLFN network is shown in Figure 9. The network is trained using a self-supervised loss function and outperforms traditional median filtering, block-matching, and 3D filtering in noise removal, providing superior noise suppression and generating noise-free phase maps.
To address the issue (2), Tang et al. from Zhejiang University of Science and Technology (2024) identified the high similarity between digital holographic mixed phase noise (DHHPN) and Perlin noise and proposed a deep learning-based continuous phase denoising method utilizing this similarity [31]. This approach not only resolved challenges associated with experimental data collection and labeling but also improved the generalizability of the trained CNN model across different DHM systems, effectively addressing unclear feature representation.
In 2025, Awais et al. proposed a lightweight wrapped-phase denoising network, WPD-Net [32], that enhances feature selection capability in noisy regions through residual dense attention blocks (RDABs), thereby providing an efficient solution for real-time biomedical imaging.
The Table 3 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase filtering, focusing on the challenges in acquiring noise-free phase labels and unclear feature representations.
To tackle the issue (3), Fang et al. from the Kunming University of Science and Technology (2022) introduced a deep learning-based speckle denoising algorithm using conditional generative adversarial networks (cGANs) [33], which enhanced feature extraction and denoising performance. In 2023, they integrated the Fourier neural operator into the CNN model [34], proposing the Fourier-UNet network. This architecture boosted denoising accuracy while reducing the number of training parameters. In 2024, they further enhanced their model by incorporating a 4-f optical speckle simulation module into the CycleGAN network [35]. This approach eliminated the need for pre-labeled paired data, improving both training accuracy and speed, and resulting in a 6.9% performance improvement over traditional methods when applied to simulated data. Figure 10 compares various methods with those reported in [24] for speckle noise denoising performance.
The Table 4 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase filtering, focusing on denoising inefficiencies.
The Table 5 below presents a classification and comparative summary of deep learning and traditional methods for phase filtering.
The Table 6 below presents a comparative analysis of deep learning methods for improving phase filtering denoising effects.
M A E = 1 m i = 1 m y i y i
M S E = 1 m i = 1 m y i y i 2
P S N R = 10   log 10 M A X I 2 M S E = 20   log 10 M A X I M S E
S S I M = 2 μ x μ y + c 1 σ x y + c 2 μ x 2 + μ y 2 + c 1 σ x 2 + σ y 2 + c 2
R F = t = 1 M j = 1 N d i , j d i , j 1 M N
C F = i = 2 M j = 1 N d i , j d i 1 , j M N
S F = R F 2 + C F 2
where  y i   a n d   y i  denote the true and predicted values of the  i  sample, respectively.  M A X I  represents the maximum value of the image point color, which is 255 when each sample point is represented by 8 bits.  u x  and  u y  represent the mean values of images X and Y, respectively.  σ x  and  σ y  are the variances of images X and Y, respectively.  σ x y  indicates the covariance between images X and Y, and  c 1  and  c 2  are constants.
In phase filtering, spatial frequency serves as a key metric for assessing the effectiveness of denoising. Equations (21) and (22) define the row frequency and column frequency, which quantify the frequency variations along the row and column directions of an image, respectively. Equation (23) combines row frequency and column frequency to compute the overall spatial frequencies, providing a comprehensive measure of the image’s frequency distribution.
Currently, deep learning-based denoising techniques are confronted with challenges, such as limited accuracy, high reliance on data, poor generalization, substantial computational demands, and model complexity. Future research is expected to focus on enhancing adaptability, improving accuracy, and simplifying model architectures to address these limitations.

4.2. Application of Deep Learning in Phase Unwrapping

After phase filtering, phase values are often confined within the 2π periodic range, resulting in jumps or discontinuities in the phase signal  [ π π ] . The goal of phase unwrapping is to remove these periodic jumps, enabling the phase variation to represent a continuous and physically plausible process. This process involves sequentially correcting the phase data to eliminate errors introduced by periodic constraints, thereby ensuring an accurate representation of the object’s surface morphology and subtle phase variations.
Traditional phase unwrapping techniques are generally categorized into time phase unwrapping (TPU) and spatial phase unwrapping (SPU) methods. TPU methods address phase unwrapping by acquiring multiple phase maps at different frequencies over time, with multi-frequency approaches [36] being commonly used to resolve phase surface discontinuities. SPU methods, which focus on phase unwrapping the phase in a single two-dimensional wrapped phase image, include path-following and minimum norm methods. The path-following approach recovers the continuous phase by defining integration paths, with common techniques, such as branch cut [37], quality-guided [38], and region growth [39]. The minimum norm method transforms phase unwrapping into a global optimization problem, aiming to minimize the difference between the phase gradient and the true phase gradient, with least squares being a typical representative technique.
Traditional phase unwrapping methods take into account the image features of the phase map and the algorithm’s stability, but they still encounter several challenges, including
(1)
Prolonged phase unwrapping time;
(2)
Limited robustness;
(3)
Poor generalization;
(4)
Low accuracy.
Deep learning approaches to phase unwrapping are typically categorized into two main types: regression-based methods and segmentation-based methods. Regression-based methods approach phase unwrapping as an image recovery problem, whereas segmentation-based methods treat it as an image segmentation issue.
To address the issue (1), Seonghwan Park et al. from South Korea introduced a deep learning model called “UnwrapGAN” in 2021 [40]. This model uses phase unwrapping as an image recovery task, effectively overcoming the challenges posed by phase discontinuities. Compared to traditional phase unwrapping methods, UnwrapGAN is able to perform phase unwrapping at twice the speed.
In 2024, Wang et al. from Hefei University of Technology introduced the Res-UNet network [41], as depicted in the detailed schematic in Figure 11. This network leverages the correlation between residual networks and U-Net networks. Compared to the “UnwrapGAN” model [40], the Res-UNet network achieves phase unwrapping in approximately 1 s, even for larger images. Additionally, in comparison to path-based algorithms, this network not only demonstrates superior noise suppression capabilities but also delivers better performance across varying image sizes.
The Table 7 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase unwrapping, focusing on mitigating the phase unwrapping time problem.
The Table 8 below presents a comparative analysis of the impact of deep learning methods on reducing phase unwrapping time.
To address the issue (2), the following methods frame the problem as an image segmentation task. In 2019, Zhang et al. from Hangzhou Dianzi University introduced the DeepLabV3+ network [42], based on deep CNNs. Compared to conventional phase unwrapping techniques, this network delivers satisfactory phase unwrapping results, even in highly noisy conditions. Figure 12 in their study illustrates the phase unwrapping of real circular data of the methods in reference [42] and the traditional methods sorting by reliability following a noncontinuous path (SRFNP), transport of intensity equation (TIE), iterative transport of intensity equation (ITIE), and robust transport of intensity equation (RTIE) under significant noise.
In 2020, Spoorthi et al. from the Indian Institute of Technology introduced a new phase unwrapping learning framework called PhaseNet2.0 [43]. Unlike the previously proposed PhaseNet method [44], this framework incorporates a novel loss function (such as L1 and residual loss) and uses L1 loss to address class imbalance. Moreover, it eliminates the need for post-processing.
In 2024, Zhang et al. from Northwestern Polytechnical University, drawing inspiration from SegFormer [45], introduced a new phase unwrapping approach based on SFNet. This method outperforms traditional phase unwrapping approaches and the DeepLabV3+ network [42] in terms of noise resistance, accuracy, and generalization, especially in cases involving mixed noise and discontinuities.
In 2025, Awais et al. proposed the DenSFA-PU [46] network to address issues of poor robustness, error accumulation, and low computational efficiency in phase unwrapping under strong noise conditions.
The Table 9 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase unwrapping, focusing on addressing limited robustness.
The Table 10 below presents a comparative analysis of the impact of deep learning methods on improving robustness.
To address the problem (3), the following networks treat the phase unwrapping algorithm as an image restoration task. In 2022, Zhao et al. from the University of Electronic Science and Technology of China introduced the VDE-Net network [47], building upon the VUR-Net network [48]. This network incorporated a weighted skip-edge attention mechanism for the first time, and, compared to VUR-Net, the VDE-Net demonstrated improved performance in undersampled wrapped phase maps.
In 2023, Chen et al. from the University of Shanghai for Science and Technology developed the U2-Net network [49]. This model enhances generalization by incorporating deeper nested U-structures and residual modules, allowing it to maintain strong performance across various noise conditions and diverse datasets.
In 2024, Li et al. from Harbin Engineering University proposed the phase unwrapping deep convolutional network (PUDCN) for phase unwrapping in the presence of mixed noise and discontinuous regions [50]. The PUDCN was applied to phase unwrapping of optical fibers in interferometric measurements, demonstrating fewer fluctuations, outliers, and distortions, which indicates its robust generalization ability. Figure 13 presents a 3D visualization of the phase height of coaxial dual-waveguide fiber side projections, calculated using different methods.
The Table 11 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase unwrapping, focusing on addressing poor generalization.
The Table 12 below presents a comparative analysis of the impact of deep learning methods in addressing poor generalization.
To address the problem (4), Li et al. from Guangxi University of Science and Technology proposed a deep learning-based center difference information filtering PU (DLCDIFPU) algorithm in 2023 [51]. This algorithm approaches the phase unwrapping problem as an image restoration task and introduces a novel, efficient, and robust filter (CDIF) combined with efficient local phase gradient estimation and a path-tracking strategy based on heap sorting. This approach improved the accuracy of phase unwrapping by approximately 15%.
In 2024, Zhao et al. from Guilin University of Electronic Technology introduced the C-HRNet network [52], which framed the phase unwrapping problem as an image segmentation challenge. By incorporating a high-resolution network and an object context representation module, the C-HRNet network significantly enhanced phase unwrapping accuracy. Notably, when compared to the DeeplabV3+ network, it exhibited lower root mean square error (RMSE) and higher PSNR.
The Table 13 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase unwrapping, focusing on addressing low accuracy.
The following Table 14 presents a comparative analysis of deep learning-based methods for enhancing accuracy.
R M S E = 1 N i = 1 N y i y i ^ 2
A U = 1 X Y x = 1 X y = 1 Y BEM x , y × 1
where N represents the number of observations,  y i  denotes the actual value of the i-th observation, and  y ^ i  indicates the predicted value of the i-th observation. X and Y signify the number of rows and columns of the unwrapped phase image, respectively. BEM(x, y) is the pixel value on the binary error map, where the pixel is 1 if the phase is correctly unwrapped, and 0 otherwise.
In phase unpacking, the absolute error graph map serves as a critical metric for evaluating unwrapping accuracy. Equation (25) quantifies the error of understanding by computing the difference between the unwrapped and the ground truth phase images.
The Table 15 below presents a classification and comparative summary of deep learning and traditional methods for phase unwrapping.

4.3. Application of Deep Learning in Phase Distortion Compensation

Phase distortion can arise due to factors such as aberrations in the optical system, uneven illumination, tilt between the recording surface and the object, and the surface morphology of the object. These distortions can introduce unwanted low-frequency backgrounds, nonlinear errors, or high-frequency noise during the holographic reconstruction process, ultimately affecting the accuracy of the reconstructed image and the true optical characteristics of the sample.
Phase distortion compensation is a crucial process for eliminating these artifacts caused by imperfect systems or environments, correcting non-physical backgrounds, and restoring the true phase distribution of the target. This process improves the quality of holographic image reconstruction and enhances the accuracy of quantitative analysis. Traditional phase distortion compensation methods can generally be categorized into physical and numerical approaches. Physical methods typically involve electronically adjustable lenses, co-located DHM configurations, telecentric setups, and the use of identical objective lenses. Numerical methods rely on post-processing techniques and do not require additional holograms or optical components. These include fitting calculations, principal component analysis, spectral analysis, geometric transformations, and nonlinear optimization, typically implemented using traditional software.
While these traditional methods are grounded in mature physical and mathematical theories, they still face several limitations:
(1)
They require extensive preprocessing.
(2)
They are constrained by perturbation assumptions.
(3)
They often require manual intervention.
To address the issue (1), Xiao et al. from Beihang University proposed a CNN-based multi-variable regression distortion compensation method in 2021 [53]. This method predicts the optimal compensation coefficients and automatically compensates for phase distortions without needing to detect background areas or any physical parameters. Figure 14 compares the compensation results of this method with those of manual nonlinear programming-based phase compensation (NPL) methods across multiple time periods.
In 2023, Li et al. from Xi’an University of Technology introduced a digitally-driven phase distortion compensation network (PACUnet3+) [54]. This network directly generates a reference hologram from sample holograms, eliminating the need for any preprocessing techniques. PACUnet3+ outperforms the original Unet3+ network [55] in eliminating phase distortion and also surpasses traditional methods, such as phase aberration compensation (PAC) and fitting.
To address the issue (2), Tang et al. from Northwestern Polytechnical University introduced a self-supervised sparse constraint network (SSCNet) in 2023 [56]. The SSCNet combines sparse constraints with Zernike model enhancement and requires only a single measured phase for self-supervision. This network achieves high accuracy and stability in PAC, without the need for target masks or perturbation assumptions. When it comes to adaptive compensation for dynamic distortions, SSCNet outperforms the double exposure technique. Figure 15 compares the compensation results of SSCNet with those of various algorithms.
The Table 16 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase distortion compensation, focusing on extensive preprocessing and perturbation assumption constraints.
To address the issue (3), Thanh Nguyen et al. from the Catholic University of America proposed a fully automated technique in 2017 that combined deep learning-based CNNs and Zernike polynomial fitting (ZPF) [57]. This method’s advantage lies in its ability to automatically monitor the background region and compensate for higher-order aberrations, enhancing the accuracy of real-time measurements and enabling dynamic process monitoring.
In 2021, MA et al. from Northeastern University [58] introduced a phase distortion compensation method for DHM based on a two-stage generative adversarial network (GAN). This method eliminates the need for a complex spectral centering process or prior knowledge, significantly simplifying the operational workflow while improving both precision and efficiency. It successfully automates the compensation process without requiring manual intervention. Figure 16 compares the distortion compensation results of this method with those of traditional approaches.
The Table 17 below presents an analysis and comparative summary of the mechanisms of deep learning and traditional methods for phase distortion compensation, focusing on addressing the manual intervention issue.
In summary, traditional phase distortion compensation methods offer advantages in terms of theoretical framework and algorithmic simplicity, while deep learning methods stand out for their adaptability and ability to manage complex relationships. However, there is still considerable potential for improving the effectiveness of phase distortion compensation using deep learning techniques in the future.
The Table 18 below presents a classification and comparative summary of deep learning and traditional methods for phase distortion compensation.
The Table 19 below presents a comparative analysis of deep learning-based phase distortion compensation methods.
P V = ϕ m a x ϕ m i n
where  ϕ m a x  represents the maximum phase value in the phase distribution, and  ϕ m i n  denotes the minimum phase value in the phase distribution.

4.4. Evaluation Metrics for Phase Processing Results

In DHM, the quality of phase processing is essential, as it directly impacts the accuracy, fidelity, and applicability of the results. These results influence subsequent analyses and applications. Both qualitative and quantitative evaluations are commonly used to assess phase processing.
Qualitative evaluation generally focuses on three key factors: visual quality, detail preservation, and background smoothness. Quantitative evaluation involves numerical metrics to objectively assess the reconstruction from several perspectives, including phase accuracy, image quality, and noise suppression. Accuracy is often evaluated using metrics, such as the mean squared error (MSE) and MAE. Image quality is assessed with the PSNR and SSIM. Noise suppression is evaluated using the RMSE.

5. Summary and Outlook

Deep learning approaches in DHM have shown improvements in accuracy, processing speed, and generalization compared to traditional phase processing methods. However, research in this field is still in its early stages. Key areas for further exploration include the following:
(1)
Ultrafast 3D Holographic Reconstruction: Current 3D reconstruction in DHM relies on time-consuming iterative algorithms. Future work could focus on deep learning-based fast inversion algorithms, such as Bayesian–physical joint modeling [59] or a 3D reconstruction method based on unpaired data learning [60]. By training on a large dataset of paired holographic interference patterns and their 3D reconstructions, these models can be trained to learn the direct mapping. This pattern enables the rapid conversion of interference patterns into 3D reconstructions, significantly improving real-time imaging—particularly for capturing dynamic processes, such as cell movement.
(2)
High-Throughput Holographic Data Analysis: DHM generates vast amounts of data in high-throughput experiments (such as drug screening). Automating data analysis using deep learning pipelines, such as few-shot learning and transfer learning, could improve classification, segmentation, and feature extraction. In practice, the model can first be pre-trained on a large-scale general holographic dataset to learn a broad feature representation of the data. Subsequently, few-shot learning can be applied to quickly adapt the model to specific high-throughput experimental data, enabling accurate analysis and enhancing the efficiency and automation of these experiments.
(3)
Precise Quantification of Phase Analysis: In biomedical applications, accurate phase analysis is crucial for the non-destructive measurement of optical thickness and refractive index in cells and tissues. Deep learning could help develop more precise phase recovery algorithms. A CNN-based architecture can be tailored to effectively extract phase information from holograms. By introducing an attention mechanism, the network can prioritize key regions with significant phase features, enabling automatic extraction of quantitative data for complex biological analysis.
(4)
Multimodal Data Fusion: Combining DHM with other imaging techniques (e.g., electron or fluorescence microscopy) could significantly enhance the information gathered from images. Deep learning could integrate these modalities. A multimodal fusion network is specifically designed to extract features from each modality separately. These features are then combined through the fusion layer to reveal deeper structural and biological information, enhancing resolution, contrasts, and the diversity of biological markers in biomedical research.

Author Contributions

W.J.: Conceptualization, Validation, Data curation, Writing-review and editing, Supervision, Funding acquisition. L.L.: Conceptualization, Methodology, Validation, Data curation, Investigation, Writing-original draft preparation. Y.B.: Methodology, Investigation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Sichuan Science and Technology Program (grant no. 2021JDJQ0027) and the Natural Science Foundation of China (grant no. 61875166).

Data Availability Statement

The data underlying the results presented in this manuscript are not publicly available at this time but may be obtained from the authors upon reasonable request.

Acknowledgments

Wenbo Jiang gratefully acknowledges the support of the Sichuan Provincial Academic and Technical Leader Training Plan and the Overseas Training Plan of Xihua University (09/2014-09/2015, University of Michigan, Ann Arbor, USA).

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Goodman, J.W. Digital image formation from electronically detected holograms. Proc. SPIE-Int. Soc. Opt. Eng. 1967, 10, 176–181. [Google Scholar] [CrossRef]
  2. Zuo, C.; Qian, J.; Feng, S.; Yin, W.; Li, Y.; Fan, P.; Han, J.; Qian, K.; Chen, Q. Deep learning in optical metrology: A review. Light Sci. Appl. 2022, 11, 39. [Google Scholar] [CrossRef]
  3. Shen, X.; Wang, L.; Li, W.; Wang, H.; Zhou, H.; Zhu, Y.; Yan, W.; Qu, J. Ultralow laser power three-dimensional superresolution microscopy based on digitally enhanced STED. Biosensors 2022, 12, 539. [Google Scholar] [CrossRef]
  4. Kim, M.K. Applications of digital holography in biomedical microscopy. J. Opt. Soc. Korea 2010, 14, 77–89. [Google Scholar] [CrossRef]
  5. Potter, J.C.; Xiong, Z.; McLeod, E. Clinical and biomedical applications of lensless holographic microscopy. Laser Photonics Rev. 2024, 18, 2400197. [Google Scholar] [CrossRef]
  6. Mann, C.J.; Yu, L.; Kim, M.K. Movies of cellular and sub-cellular motion by digital holographic microscopy. Biomed. Eng. OnLine 2006, 5, 21. [Google Scholar] [CrossRef]
  7. Wu, Y.; Ozcan, A. Lensless digital holographic microscopy and its applications in biomedicine and environmental monitoring. Methods 2017, 136, 4–16. [Google Scholar] [CrossRef]
  8. Rivenson, Y.; Wu, Y.; Ozcan, A. Deep learning in holography and coherent imaging. Light Sci. Appl. 2019, 8, 85. [Google Scholar] [CrossRef]
  9. Wang, K.; Song, L.; Wang, C.; Ren, Z.; Zhao, G.; Dou, J.; Di, J.; Barbastathis, G.; Zhou, R.; Zhao, J.; et al. On the use of deep learning for phase recovery. Light Sci. Appl. 2024, 13, 4. [Google Scholar] [CrossRef]
  10. Bianco, V.; Memmolo, P.; Leo, M.; Montresor, S.; Distante, C.; Paturzo, M.; Picart, P.; Javidi, B.; Ferraro, P. Strategies for reducing speckle noise in digital holography. Light Sci. Appl. 2018, 7, 48. [Google Scholar] [CrossRef]
  11. Zuo, C.; Huang, L.; Zhang, M.; Chen, Q.; Asundi, A. Temporal phase unwrapping algorithms for fringe projection profilometry: A comparative review. Opt. Lasers Eng. 2016, 85, 84–103. [Google Scholar] [CrossRef]
  12. Wang, L.; Yan, W.; Li, R.; Weng, X.; Zhang, J.; Yang, Z.; Liu, L.; Ye, T.; Qu, J. Aberration correction for improving the image quality in STED microscopy using the genetic algorithm. Nanophotonics 2018, 7, 1971–1980. [Google Scholar] [CrossRef]
  13. Zeng, T.; Zhu, Y.; Lam, E.Y. Deep learning for digital holography: A review. Opt. Express 2021, 29, 40572–40593. [Google Scholar] [CrossRef]
  14. Park, Y.K.; Depeursinge, C.; Popescu, G. Quantitative phase imaging in biomedicine. Nat. Photonics 2018, 12, 578–589. [Google Scholar] [CrossRef]
  15. Kim, M.K. Principles and techniques of digital holographic microscopy. J. Photonics Energy 2009, 1, 8005. [Google Scholar] [CrossRef]
  16. Liu, K.; He, Z.; Cao, L. Double amplitude freedom Gerchberg-Saxton algorithm for generation of phase-only hologram with speckle suppression. Appl. Phys. Lett. 2022, 120, 061103. [Google Scholar] [CrossRef]
  17. Pauwels, E.J.R.; Beck, A.; Eldar, Y.C.; Sabach, S. On Fienup methods for sparse phase retrieval. IEEE Trans. Signal Process. 2018, 66, 982–991. [Google Scholar] [CrossRef]
  18. Zhang, G.; Guan, T.; Shen, Z.; Wang, X.; Hu, T.; Wang, D.; He, Y.; Xie, N. Fast phase retrieval in off-axis digital holographic microscopy through deep learning. Opt. Express 2018, 26, 19388–19405. [Google Scholar] [CrossRef]
  19. Nagahama, Y. Phase retrieval using hologram transformation with U-Net in digital holography. Opt. Contin. 2022, 1, 1506–1515. [Google Scholar] [CrossRef]
  20. Wang, K.; Dou, J.; Kemao, Q.; Di, J.; Zhao, J. Y-Net: A one-to-two deep learning framework for digital holographic reconstruction. Opt. Lett. 2019, 44, 4765–4768. [Google Scholar] [CrossRef]
  21. Chen, B.; Li, Z.; Zhou, Y.; Zhang, Y.; Jia, J.; Wang, Y. Deep-learning multiscale digital holographic intensity and phase reconstruction. Appl. Sci. 2023, 13, 9806. [Google Scholar] [CrossRef]
  22. Zhu, Y.; Huang, C. An improved median filtering algorithm for image noise reduction. Phys. Procedia 2012, 25, 609–616. [Google Scholar] [CrossRef]
  23. Rakshit, S.; Ghosh, A.; Shankar, B.U. Fast mean filtering technique (FMFT). Pattern Recognit. 2007, 40, 890–897. [Google Scholar] [CrossRef]
  24. Kundur, D.; Hatzinakos, D. Blind image deconvolution. IEEE Signal Process. Mag. 1996, 13, 43–64. [Google Scholar] [CrossRef]
  25. Barjatya, A. Block matching algorithms for motion estimation. IEEE Trans. Evol. Comput. 2004, 8, 225–239. [Google Scholar]
  26. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising with Block-Matching and 3D Filtering. In Image Processing: Algorithms and Systems, Neural Networks, and Machine Learning, Proceedings of the Electronic Imaging 2006, 15–19 January 2006, San Jose, CA, USA; SPIE: Bellingham, WA, USA, 2006; Volume 6064, pp. 354–365. [Google Scholar]
  27. Chen, J.; Benesty, J.; Huang, Y.; Doclo, S. New insights into the noise reduction Wiener filter. IEEE Trans. Audio Speech Lang. Process. 2006, 14, 1218–1234. [Google Scholar] [CrossRef]
  28. Huang, L.; Kemao, Q.; Pan, B.; Asundi, A.K. Comparison of Fourier transform, windowed Fourier transform, and wavelet transform methods for phase extraction from a single fringe pattern in fringe projection profilometry. Opt. Lasers Eng. 2010, 48, 141–148. [Google Scholar] [CrossRef]
  29. Villasenor, J.D.; Belzer, B.; Liao, J. Wavelet filter evaluation for image compression. IEEE Trans. Image Process. 1995, 4, 1053–1060. [Google Scholar] [CrossRef]
  30. Wu, J.; Tang, J.; Zhang, J.; Di, J. Coherent noise suppression in digital holographic microscopy based on label-free deep learning. Front. Phys. 2022, 10, 880403. [Google Scholar] [CrossRef]
  31. Tang, J.; Chen, B.; Yan, L.; Huang, L. Continuous phase denoising via deep learning based on Perlin noise similarity in digital holographic microscopy. IEEE Trans. Ind. Inform. 2024, 20, 8707–8716. [Google Scholar] [CrossRef]
  32. Awais, M.; Kim, Y.; Yoon, T.; Choi, W.; Lee, B. A Lightweight Neural Network for Denoising Wrapped-Phase Images Generated with Full-Field Optical Interferometry. Appl. Sci. 2025, 15, 5514. [Google Scholar] [CrossRef]
  33. Fang, Q.; Xia, H.-T.; Song, Q.; Zhang, M.; Guo, R.; Montresor, S.; Picart, P. Speckle denoising based on deep learning via a conditional generative adversarial network in digital holographic interferometry. Opt. Express 2022, 30, 20666–20683. [Google Scholar] [CrossRef] [PubMed]
  34. Fang, Q.; Li, Q.; Song, Q.; Montresor, S.; Picart, P.; Xia, H. Convolutional and Fourier neural networks for speckle denoising of wrapped phase in digital holographic Interferometry. Opt. Commun. 2024, 550, 129955. [Google Scholar] [CrossRef]
  35. Yu, H.B.; Fang, Q.; Song, Q.H.; Montresor, S.; Picart, P.; Xia, H. Unsupervised speckle denoising in digital holographic interferometry based on 4-f optical simulation integrated cycle-consistent generative adversarial network. Appl. Opt. 2024, 63, 3557–3569. [Google Scholar] [CrossRef] [PubMed]
  36. Demer, D.A.; Soule, M.A.; Hewitt, R.P. A multiple-frequency method for potentially improving the accuracy and precision of in situ target strength measurements. J. Acoust. Soc. Am. 1999, 105, 2359–2376. [Google Scholar] [CrossRef]
  37. Belenguer, J.-M.; Benavent, E.; Prins, C.; Prodhon, C.; Calvo, R.W. A branch-and-cut method for the capacitated location-routing problem. Comput. Oper. Res. 2011, 38, 931–941. [Google Scholar] [CrossRef]
  38. Zhao, M.; Huang, L.; Zhang, Q.; Su, X.; Asundi, A.; Kemao, Q. Quality-guided phase unwrapping technique: Comparison of quality maps and guiding strategies. Appl. Opt. 2011, 50, 6214–6224. [Google Scholar] [CrossRef]
  39. Cheng, Z.; Wang, J. Improved region growing method for image segmentation of three-phase materials. Powder Technol. 2020, 368, 80–89. [Google Scholar] [CrossRef]
  40. Park, S.; Kim, Y.; Moon, I. Automated phase unwrapping in digital holography with deep learning. Biomed. Opt. Express 2021, 12, 7064–7081. [Google Scholar] [CrossRef]
  41. Wang, B.; Cao, X.; Lan, M.; Wu, C.; Wang, Y. An Anti-noise-designed residual phase unwrapping neural network for digital speckle pattern interferometry. Optics 2024, 5, 44–55. [Google Scholar] [CrossRef]
  42. Zhang, T.; Jiang, S.; Zhao, Z.; Dixit, K.; Zhou, X.; Hou, J.; Zhang, Y.; Yan, C. Rapid and robust two-dimensional phase unwrapping via deep learning. Opt. Express 2019, 27, 23173–23185. [Google Scholar] [CrossRef]
  43. Spoorthi, G.E.; Gorthi, R.K.S.S.; Gorthi, S. PhaseNet 2.0: Phase unwrapping of noisy data based on deep learning approach. IEEE Trans. Image Process. 2020, 29, 4862–4872. [Google Scholar] [CrossRef]
  44. Spoorthi, G.E.; Gorthi, S.; Gorthi, R.K.S.S. PhaseNet: A deep convolutional neural network for two-dimensional phase unwrapping. IEEE Signal Process. Lett. 2018, 26, 54–58. [Google Scholar] [CrossRef]
  45. Zhang, Z.; Wang, X.; Liu, C.; Han, Z.; Xiao, Q.; Zhang, Z.; Feng, W.; Liu, M.; Lu, Q. Efficient and robust phase unwrapping method based on SFNet. Opt. Express 2024, 32, 15410–15432. [Google Scholar] [CrossRef]
  46. Awais, M.; Yoon, T.; Hwang, C.O.; Lee, B. DenSFA-PU: Learning to unwrap phase in severe noisy conditions. Opt. Laser Technol. 2025, 187, 112757. [Google Scholar] [CrossRef]
  47. Zhao, J.; Liu, L.; Wang, T.; Wang, X.; Du, X.; Hao, R.; Liu, J.; Liu, Y.; Zhang, J. VDE-Net: A two-stage deep learning method for phase unwrapping. Opt. Express 2022, 30, 39794–39815. [Google Scholar] [CrossRef] [PubMed]
  48. Qin, Y.; Wan, S.; Wan, Y.; Weng, J.; Liu, W.; Gong, Q. Direct and accurate phase unwrapping with deep neural network. Appl. Opt. 2020, 59, 7258–7267. [Google Scholar] [CrossRef] [PubMed]
  49. Chen, J.; Kong, Y.; Zhang, D.; Fu, Y.; Zhuang, S. Two-dimensional phase unwrapping based on U2-Net in complex noise environment. Opt. Express 2023, 31, 29792–29812. [Google Scholar] [CrossRef]
  50. Li, Y.; Meng, L.; Zhang, K.; Zhang, Y.; Xie, Y.; Yuan, L. PUDCN: Two-dimensional phase unwrapping with a deformable convolutional network. Opt. Express 2024, 32, 27206–27220. [Google Scholar] [CrossRef]
  51. Jiaying, L. Central difference information filtering phase unwrapping algorithm based on deep learning. Opt. Lasers Eng. 2023, 163, 107484. [Google Scholar] [CrossRef]
  52. Zhao, W.; Yan, J.; Jin, D.; Ling, J. C-HRNet: High resolution network based on contexts for single-frame phase unwrapping. IEEE Photonics J. 2024, 16, 5100210. [Google Scholar] [CrossRef]
  53. Xiao, W.; Xin, L.; Cao, R.; Wu, X.; Tian, R.; Che, L.; Sun, L.; Ferraro, P.; Pan, F. Sensing morphogenesis of bone cells under microfluidic shear stress by holographic microscopy and automatic aberration compensation with deep learning. Lab A Chip 2021, 21, 1385–1394. [Google Scholar] [CrossRef]
  54. Li, Z.; Wang, F.; Jin, P.; Zhang, H.; Feng, B.; Guo, R. Accurate phase aberration compensation with convolutional neural network PACUnet3+ in digital holographic microscopy. Opt. Lasers Eng. 2023, 171, 107829. [Google Scholar] [CrossRef]
  55. Yin, M.; Wang, P.; Ni, C.; Hao, W. Cloud and snow detection of remote sensing images based on improved Unet3+. Sci. Rep. 2022, 12, 14415. [Google Scholar] [CrossRef] [PubMed]
  56. Tang, J.; Zhang, J.; Zhang, S.; Mao, S.; Ren, Z.; Di, J.; Zhao, J. Phase aberration compensation via a self-supervised sparse constraint network in digital holographic microscopy. Opt. Lasers Eng. 2023, 168, 107671. [Google Scholar] [CrossRef]
  57. Nguyen, T.; Bui, V.; Lam, V.; Raub, C.B.; Chang, L.-C.; Nehmetallah, G. AutAomatic phase aberration compensation for digital holographic microscopy based on deep learning background detection. Opt. Express 2017, 25, 15043–15057. [Google Scholar] [CrossRef] [PubMed]
  58. Ma, S.; Liu, Q.; Yu, Y.; Luo, Y.; Wang, S. Quantitative phase imaging in digital holographic microscopy based on image inpainting using a two-stage generative adversarial network. Opt. Express 2021, 29, 24928–24946. [Google Scholar] [CrossRef]
  59. Tan, J.; Niu, H.; Su, W.; He, Z. Structured light 3D shape measurement for translucent media base on deep Bayesian inference. Opt. Laser Technol. 2025, 181, 111758. [Google Scholar] [CrossRef]
  60. Tan, J.; Liu, J.; Wang, X.; He, Z.; Su, W.; Huang, T.; Xie, S. Large depth range binary-focusing projection 3D shape reconstruction via unpaired data learning. Opt. Lasers Eng. 2024, 181, 108442. [Google Scholar] [CrossRef]
Figure 1. Schematic of the digital holography recording process.
Figure 1. Schematic of the digital holography recording process.
Photonics 12 00810 g001
Figure 2. Schematic of the CCD image sensor structure.
Figure 2. Schematic of the CCD image sensor structure.
Photonics 12 00810 g002
Figure 3. Schematic of the digital holographic reconstruction process.
Figure 3. Schematic of the digital holographic reconstruction process.
Photonics 12 00810 g003
Figure 4. Flowchart of phase processing.
Figure 4. Flowchart of phase processing.
Photonics 12 00810 g004
Figure 5. Theoretical framework for phase processing in digital holographic microscopy.
Figure 5. Theoretical framework for phase processing in digital holographic microscopy.
Photonics 12 00810 g005
Figure 6. U-Net CNN model [19], where green line represents the combination of LeakyRelu + Conv2D + BatchNorm, pink line represents the combination of Relu + DeConv2D + BatchNorm, red line represents the combination of Relu + DeConv2D + BatchNorm + Dropout, and orange line represents the combination of Relu + DeConv2D + Tanh.
Figure 6. U-Net CNN model [19], where green line represents the combination of LeakyRelu + Conv2D + BatchNorm, pink line represents the combination of Relu + DeConv2D + BatchNorm, red line represents the combination of Relu + DeConv2D + BatchNorm + Dropout, and orange line represents the combination of Relu + DeConv2D + Tanh.
Photonics 12 00810 g006
Figure 7. Comparison of network reconstruction results between Mimo-Net and Y-Net at a 256 × 256 scale [21].
Figure 7. Comparison of network reconstruction results between Mimo-Net and Y-Net at a 256 × 256 scale [21].
Photonics 12 00810 g007
Figure 8. Flowchart of phase extraction, where the red dashed box represents local magnification.
Figure 8. Flowchart of phase extraction, where the red dashed box represents local magnification.
Photonics 12 00810 g008
Figure 9. Network framework diagrams of DBSN and NLFN [30], where (A) represent the self-supervised learning scheme, (B) represent the DBSN network model, and (C) represent the NLFN network model.
Figure 9. Network framework diagrams of DBSN and NLFN [30], where (A) represent the self-supervised learning scheme, (B) represent the DBSN network model, and (C) represent the NLFN network model.
Photonics 12 00810 g009
Figure 10. Comparison of speckle noise image denoising using different methods [35].
Figure 10. Comparison of speckle noise image denoising using different methods [35].
Photonics 12 00810 g010
Figure 11. Detailed schematic of the neural network structure [41].
Figure 11. Detailed schematic of the neural network structure [41].
Photonics 12 00810 g011
Figure 12. Phase unwrapping of real circular data under strong noise ((af) represent the wrapped phase map, SRFNP, TIE, ITIE, RTIE, and the unwrapped phase map generated by the method in reference [42], respectively).
Figure 12. Phase unwrapping of real circular data under strong noise ((af) represent the wrapped phase map, SRFNP, TIE, ITIE, RTIE, and the unwrapped phase map generated by the method in reference [42], respectively).
Photonics 12 00810 g012
Figure 13. Three-dimensional visualization of the side projection phase height of coaxial dual-waveguide fibers calculated using different methods [50].
Figure 13. Three-dimensional visualization of the side projection phase height of coaxial dual-waveguide fibers calculated using different methods [50].
Photonics 12 00810 g013
Figure 14. Comparison of compensation results between the method in reference [53] and manual NPL over multiple time periods, where (ad,f) represent distorted images at different time periods, (gk) represent distorted images after distortion compensation by CNN at different time periods, and (lp) represent distorted images after distortion compensation by Manual NPL at different time periods.
Figure 14. Comparison of compensation results between the method in reference [53] and manual NPL over multiple time periods, where (ad,f) represent distorted images at different time periods, (gk) represent distorted images after distortion compensation by CNN at different time periods, and (lp) represent distorted images after distortion compensation by Manual NPL at different time periods.
Photonics 12 00810 g014
Figure 15. Compensation effects of SSCNet and various algorithms [56].
Figure 15. Compensation effects of SSCNet and various algorithms [56].
Photonics 12 00810 g015
Figure 16. (a) Result obtained by the proposed method; (b) 3D rendering of (a). (c) Result obtained by BS+ZPF; (d) 3D rendering of (c). (e) Result obtained by ZPF; (f) 3D rendering of (e) [58].
Figure 16. (a) Result obtained by the proposed method; (b) 3D rendering of (a). (c) Result obtained by BS+ZPF; (d) 3D rendering of (c). (e) Result obtained by ZPF; (f) 3D rendering of (e) [58].
Photonics 12 00810 g016
Table 1. Classification of phase recovery networks across key dimensions.
Table 1. Classification of phase recovery networks across key dimensions.
MethodApplication Scenario DimensionTechnical Principle DimensionLearning Paradigm Dimension
U-Net [18]Biomedicine (off-axis DHM).Data-driven (end-to-end mapping from defocused holograms to focused phases using U-Net).Supervised learning (L2 Loss + high-frequency weighting for phase reconstruction accuracy).
U-Net [19]Biomedicine/materials science (in-line DHM).Data-driven (multi-scale feature fusion via U-Net for separating object waves and conjugate images in single holograms).Supervised learning (using MAE loss function for optimizing the mapping relationship between complex amplitude and light intensity).
Y-Net [20]Real-time biomedical imaging; simultaneous intensity and phase reconstruction (off-axis DHM).Dual-branch U-Net with single/double/triple convolutional kernels for multi-scale feature extraction.Supervised learning.
Mimo-Net [21]Multi-scale holographic reconstruction (off-axis DHM).Multi-branch input with multi-scale feature extraction modules and cross-layer fusion.Supervised learning.
Table 2. Comparison of advantages, disadvantages, and mechanisms analysis of phase recovery algorithms.
Table 2. Comparison of advantages, disadvantages, and mechanisms analysis of phase recovery algorithms.
MethodCore MechanismAdvantagesLimitationsConclusion
Gerchberg–Saxton iterative algorithm [16], Fienup iterative algorithm [17]Alternating projections between spatial and frequency domains to iteratively approximate the true phase using known amplitude.High interpretability and low data dependency.Low computational efficiency, sub-optimal noise resistance, and limited adaptability to complex scenes.Traditional algorithms offer strong physical interpretability in simple and low-noise settings but are constrained by limited real-time performance and scalability in complex scenarios.
U-Net [18]Extracts multi-level features from defocused holograms via multi-layer convolutions, retains structural details through skip connections, and directly outputs the focused phase.High real-time performance, strong noise resistance, detail preservation, and good adaptability to complex scenarios.Dependent on simulated data and sensitive to parameter tuning.Well-suited for real-time dynamic applications, such as rapid phase monitoring of living cells in biomedical imaging.
U-Net [19]The U-Net encoder captures noise patterns from holograms, while the decoder separates object waves and conjugates images via a nonlinear transformation to output pure phases.High efficiency in single-image processing, targeted noise reduction, and a simplified workflow.Limited generalization to diverse noise types and high computational resource demands.Optimized for single-image noise reduction, making it suitable for high-throughput hologram analysis in materials science.
Y-Net [20]Utilizes a dual-branch U-Net to integrate multi-scale convolutional features for concurrent intensity and phase reconstruction from digital holograms.Symmetrical structure with 30% fewer parameters than U-Net; achieves SSIM > 0.96 for simultaneous intensity and phase reconstruction.Limited generalization to complex noise environments and multi-scale scenarios.Well-suited for real-time biomedical imaging with a reduced parameter count; generalization to complex and multi-scale conditions requires enhancement.
Mimo-Net [21]Applies a multi-branch network to extract global and local features at multiple scales, enabling phase reconstruction through cross-layer fusion.Supports three scales within a single model; delivers 62% faster reconstruction with superior PSNR and SSIM compared to Y-Net.Higher computational overhead for multi-scale training; slightly reduced phase reconstruction accuracy.Effective for high-throughput tasks via multi-branch, multi-scale feature extraction; however, phase accuracy lags behind intensity reconstruction.
Table 3. Mechanism analysis and comparison of traditional methods and deep learning algorithms for phase filtering, focusing on the challenges in acquiring noise-free phase labels and unclear feature representations.
Table 3. Mechanism analysis and comparison of traditional methods and deep learning algorithms for phase filtering, focusing on the challenges in acquiring noise-free phase labels and unclear feature representations.
MethodCore MechanismAdvantagesLimitationsConclusion
References [22,23,24,25,26,27,28,29]Signal processing based on prior models, including spatial filtering, frequency-domain filtering, and statistical models.Strong physical interpretability and low dependency on training data.Poor adaptability to complex or structured noise, sensitivity to parameters, and potential for structural distortion.Traditional algorithms retain advantages in physical interpretability for simple and low-noise scenarios, but they are limited in real-time performance and effectiveness under complex scenarios.
DBSN and NLFN [30]Joint training of a denoising network (DBSN) and a noise level estimation network (NLFN) using Bayesian rules and negative log-likelihood loss, eliminating the need for clean labels.No requirement for labels and dynamic noise estimation capability of noise levels.Model assumption limitations, high computational complexity.Designed for biomedical scenarios where label acquisition is difficult; however, generalizability is limited when noise deviates from Gaussian distributions.
DHHPN [31]Observes that holographic noise resembles Perlin noise, which is then used to simulate and generate training data.Accurate noise simulation and low labeling cost.Simulation bias and absence of frequency-domain feature representation.Enhances denoising accuracy through physically inspired noise modeling, but its effectiveness depends on how closely simulated noise matches real noise and the absence of frequency-domain feature representation.
WPD-Net [32]Residual dense attention blocks (RDABs) + dense feature fusion (DFF) with a dynamic hybrid loss functionLightweight model with strong mixed noise suppression and excellent real-time performancePerformance degrades under extreme noise conditions (SNR < 0 dB)Suitable for high-noise scenarios in biomedicine and materials science imaging
Table 4. Mechanism-level analysis and comparison of deep learning algorithms for phase filtering, focusing on denoising inefficiencies.
Table 4. Mechanism-level analysis and comparison of deep learning algorithms for phase filtering, focusing on denoising inefficiencies.
MethodCore MechanismAdvantagesLimitationsConclusion
cGAN [33]The generator (U-Net+DenseNet) maps holograms to clean phases, and the discriminator enforces the generated images to match optical physical characteristics by combining PSNR loss.High-fidelity reconstruction and strong generalizability.High data dependency and unstable training.Enables high-fidelity reconstruction of phase images under strong speckle noise, but requires high-quality labels and is susceptible to generative adversarial network (GAN) mode collapse.
Fourier-Unet [34]Combines CNN for local spatial feature extraction with a Fourier neural operator to capture global frequency information.Multi-domain complementarity and physically informed design.High computational cost and complex architecture.Enhances denoising accuracy and subsequent phase unwrapping effects through joint spatial-frequency processing, but demands significant computational complexity and careful architectural parameters.
4-fCycleGAN [35]Integrates a 4-f optical system to simulate a speckle noise generator (F) and employs CycleGAN with cyclic consistency loss (Lcyc) to ensure phase reversibility before and after denoising.Fully unsupervised; accommodates diverse noise patterns.Limited by fixed simulation parameters; potential performance bottlenecks.Designed for label-free industrial optical inspection with fixed simulation parameters and tasks, but its generalizability to complex real-world noise is lower than that of supervised methods.
Table 5. Classification of phase filtering network dimensions.
Table 5. Classification of phase filtering network dimensions.
MethodApplication Scenario DimensionTechnical Principle DimensionLearning Paradigm Dimension
DBSN and NLFN [30]Biomedicine, materials science (Off-axis DHM).Physical noise modeling with self-supervised learning structure.Unsupervised learning (trained solely on noisy data).
DHHPN [31]Precision measurement, biomedicine (in-line DHM).Supervised learning with physical noise simulation based on the Perlin noise model.Supervised learning (paired training with simulated datasets).
WPD-Net [32]Biomedicine, materials science (in-line DHM).Data-driven (RDAB + DFF) + attention mechanismsSupervised learning (dynamic hybrid loss)
cGAN [33]Industrial inspection, optical metrology (in-line DHM).Generative adversarial networks (GANs) with optical simulation datasets.Supervised learning with noise-clean image pairing.
Fourier-Unet [34]Phase denoising in holographic interferometry, 3D topography measurement (in-line DHM).Hybrid approach: CNN for spatial features combined with Fourier frequency-domain features.Supervised learning (using both simulated and experimental data).
4-fCycleGAN [35]Industrial optical inspection, holographic metrology (in-line DHM).Unsupervised learning combined with speckle noise simulation via 4-f optical system.Unsupervised learning (trained on unpaired data).
Table 6. Comparative analysis of deep learning methods for enhancing denoising performance in phase filtering.
Table 6. Comparative analysis of deep learning methods for enhancing denoising performance in phase filtering.
YearMethodMAEPSNRSSIMSFSuggestions for Improvement
2025WPD-Net [32]0.006826.5860.987/Enhance WPD-Net’s performance under extreme noise conditions (SNR < 0 dB) by integrating temporal dynamics into the training process
2022cGAN [33]0.005525.250.99741.3567Improve adaptability to dynamic scenes and enhance overall denoising performance.
2023Fourier-Unet [34]0.003943.280.9999/Strengthen denoising capabilities under complex backgrounds and improve phase unwrapping accuracy.
20244-fCycleGAN [35]0.0954916.1647/0.03614Improve generalization to non-Gaussian noise and enhance real-time processing efficiency.
Note: The data presented above are the experimental results obtained from their respective datasets. Evaluation metrics: The calculation formulas for MAE, peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and signal fidelity (SF) are provided as follows.
Table 7. Comparative analysis of traditional and deep learning methods in mitigating the phase unwrapping time problem.
Table 7. Comparative analysis of traditional and deep learning methods in mitigating the phase unwrapping time problem.
MethodCore MechanismAdvantagesLimitationsConclusion
References [36,37,38,39]The path tracking method utilizes a phase quality map to guide integration paths. The minimum norm method performs global optimization to minimize phase gradient errors.High physical interpretability; low dependence on training data.Sensitive to noise; low computational efficiency; fails in complex phase scenarios.These traditional methods offer clarity in physical principles and require no training data; however, they are sensitive to noise and phase mutations and are best suited for simple scenarios with low noise and continuous-phase settings.
UnwrapGAN [40]Generative adversarial learning (GAN) improves the authenticity of phase structures through discriminators.Good real-time performance and strong generalization ability.Requires high-quality paired data; unstable training.Enables automatic focusing and phase unwrapping, at over twice the speed of traditional algorithms, but depends heavily on high-quality paired data and is vulnerable to mode collapse during training.
Res-Unet [41]Incorporates residual blocks and attention mechanisms to enhance edge feature extraction and structural detail.Strong noise resistance; robust in the presence of phase discontinuities.High model complexity; large parameter count; significant computational demands.Demonstrates strong robustness to complex noise and undersampling conditions. However, its practical deployment is constrained by high computing resources.
Table 8. Comparative analysis of deep learning methods for reducing phase unwrapping time.
Table 8. Comparative analysis of deep learning methods for reducing phase unwrapping time.
YearMethodSSIMTime (Seconds)Suggestions for Improvement
2021UnwrapGAN [40]0.90000.5021Enhance generalization by expanding the training dataset to cover more diverse cell types and complex phase structures; explore lightweight network designs to reduce computational latency in real-time applications.
2024Res-Unet [41]0.99361.087Incorporate physics-based constraints into the loss function to improve handling of phase discontinuities under extreme noise scenarios; implement a multi-scale network architecture to enhance performance on high-resolution images without image stitching.
Table 9. Comparative mechanism analysis of deep learning methods addressing limited robustness in phase unwrapping.
Table 9. Comparative mechanism analysis of deep learning methods addressing limited robustness in phase unwrapping.
MethodCore MechanismAdvantagesLimitationsConclusion
DeepLabV3+ [42]Utilizes the encoder-decoder architecture for semantic segmentation. Multi-scale features are extracted through atrous spatial pyramid pooling (ASPP), and high- and low-level features are fused through skip connections to predict wrap counts in phase unwrapping.Enables multi-scale feature fusion, suitable for unwrapping complex phase structures.Limited performance on small-target details; requires additional post-processing optimization.Effectively addresses accuracy and efficiency issues in low-noise, regular-structure scenarios. However, performance on large-sized images requires downsampling, and training depends on high-quality paired datasets.
PhaseNet2.0 [43]Utilizes a fully convolutional DenseNet architecture, framing phase unwrapping as a dense classification problem. A novel loss function combining gradient difference minimization and L1 loss is employed to mitigate class imbalance, directly and predict wrap counts.Strong resistance to noise; supports high-dynamic range phase recovery; end-to-end training reduces manual intervention.High reliance on simulated data; limited generalization to unseen noise types.Effectively addresses the insufficient robustness against complex noise and high-dynamic-range phase conditions. However, it exhibits a large parameter volume and limited generalization to highly complex or geometric structures.
SFNet [45]Builds on a lightweight variant of SegFormer using a hierarchical transformer encoder without positional encoding to capture global phase dependencies, combines an MLP decoder to fuse multi-scale features, and enhances phase unwrapping accuracy in complex noise and discontinuous scenarios via self-attention mechanisms.Demonstrates strong global feature modeling capability with low parameter count; facilitates real-time inference.Limited handling of fine local details; relying on the Transformer structure results in slightly higher computational complexity.This approach addresses the limitations of traditional methods in global feature modeling and real-time performance. However, tile-based processing is needed for extremely high-resolution images, with edge detail accuracy left to be improved.
DenSFA-PU [46]Densely connected encoder + SFA module (Bi-LSTM + BAM)High robustness to strong noise; fast unwrapping speedRequires optimization of initial conditions under extreme noiseSuitable for real-time holographic imaging and unwrapping
Table 10. Comparative analysis of the effects of deep learning methods in addressing limited robustness in phase unwrapping.
Table 10. Comparative analysis of the effects of deep learning methods in addressing limited robustness in phase unwrapping.
YearMethodSSIMRMSEPSNRSuggestions for Improvement
2019DeepLabV3+ [42]/1.2325/Explore multi-scale training strategies to improve performance on high-resolution images, and integrate physics-informed constraints into the network to enhance noise suppression in extreme scenarios.
2020PhaseNet2.0 [43]0.92000.03540.5Develop an adaptive loss function capable of handling diverse noise types, and explore lightweight network architectures to improve real-time processing efficiency.
2024SFNet [45]0.99630.5268101.22Include dynamic attention mechanisms to better capture local phase discontinuities, and expand the training dataset to include more complex real-world noise patterns.
2025DenSFA-PU [46]0.999//Improve stability under extreme noise conditions (e.g., SNR < 0 dB) and enhance the network’s ability to capture long-range dependencies.
Table 11. Mechanism comparison of deep learning methods addressing poor generalization in phase unwrapping.
Table 11. Mechanism comparison of deep learning methods addressing poor generalization in phase unwrapping.
MethodCore MechanismAdvantagesLimitationsConclusion
VDE-Net [47]Utilizes a two-stage architecture that emphasizes phase jump regions through a weighted jump-edge attention mechanism; applies dilated convolutions to expand the receptive field and achieves robust phase unwrapping to noise and undersampling.Enhances edge feature enhancement; robust to noise and undersampling; incorporates physical priors in the preprocessing stage to improve interpretability.Increased computational complexity due to attention modules; reduced stability under extreme noise environments.Effectively addresses the shortcomings of traditional methods in edge detail processing and noise robustness.
U2-Net [49]Constructs an encoder–decoder architecture using nested U-shaped residual blocks, enabling multi-scale feature fusion and deep supervision to improve phase unwrapping accuracy and structural stability in complex noise environments.Provides strong multi-resolution feature extraction capability, superior noise robustness over traditional U-Net, and supports wide dynamic range.High parameter count and extended training time; performance degrades in extremely low signal-to-noise ratio.Strikes a balance between parameter scale and noise resilience than traditional deep learning models.
PUDCN [50]Incorporates deformable convolutions to dynamically capture irregular phase edges, combining a pyramid feature dynamic selection module and fusion feature dynamic selection module for multi-scale feature extraction and detail-optimized phase unwrapping via a coarse-to-fine strategy.Deformable convolutions adapt to phase jumps of arbitrary shapes; lightweight design with only 1.16 M parameters, suitable for deployment on edge devices.The two-stage training process is complex; inference speed decreases significantly for extremely high-resolution images.Effectively addresses the limitations of traditional CNNs in handling irregular edges.
Table 12. Comparative analysis of the effects of deep learning methods in addressing poor generalization.
Table 12. Comparative analysis of the effects of deep learning methods in addressing poor generalization.
YearMethodSSIMRMSEPSNRSuggestions for Improvement
2022VDE-Net [47]0.9984//Dynamically adjust the weighted factors in the jump-edge attention mechanism to improve the adaptability to various noise distributions and phase discontinuities.
2023U2-Net [49]0.9989/63.6289Incorporate lightweight attention modules into U2-Net to reduce computational overhead while preserving its multi-scale feature extraction capability for real-time applications.
2024PUDCN [50]0.99860.2625/Employ self-adaptive deformable convolution kernels to effectively extract irregular phase edges and enhance generalization under extreme noise conditions.
Table 13. Mechanism analysis of deep learning mechanisms addressing low accuracy in phase unwrapping.
Table 13. Mechanism analysis of deep learning mechanisms addressing low accuracy in phase unwrapping.
MethodCore MechanismAdvantagesLimitationsConclusion
DLCDIFPU [51]Utilizes a deep learning-based region segmentation model to divide wrapped phase maps into irregular regions along fringe edges; integrates CDIF with a heap-sort path-tracking strategy to unwrap each region independently, merging the results via phase consistency and achieving noise-resistant, and localized phase unwrapping.Integrates data-driven learning with physical models; achieves an RMSE of only 0.24π under strong noise. Regional segmentation effectively suppresses cross-edge error propagation.Depends on a dual-network architecture and necessitates high-precision annotated data, increasing implementation complexity.Enhances phase unwrapping accuracy and efficiency by leveraging regional processing and physical prior, such as fringe edge detection, to guide localized unwrapping.
C-HRNet [52]Extends HRNet by adding a fifth stage for enhanced multi-resolution feature fusion and incorporates the OCR module to aggregate pixel-object relationships. Treats phase unwrapping as a semantic segmentation task, predicting fringe order maps directly from single-frame wrapped phases, thereby improving unwrapping accuracy in complex noise and isolated regions.Supports multi-scale feature fusion and contextual mechanisms; enables attention, and end-to-end learning for improved phase unwrapping accuracy.Relies heavily on large-scale, data; lacks clear physical interpretability; incurs high computational cost.Enhances the robustness of phase unwrapping in complex scenarios through multi-scale feature extraction and contextual semantic modeling.
Table 14. Comparative analysis of deep learning methods for addressing low accuracy in phase unwrapping.
Table 14. Comparative analysis of deep learning methods for addressing low accuracy in phase unwrapping.
YearMethodRMSESuggestions for Improvement
2023DLCDIFPU [51]/Incorporate adaptive noise suppression mechanisms into the central difference information filter to enhance the robustness of complex fringe structures and under severe noise variations.
2024C-HRNet [52]1.76Establish a dynamic context aggregation mechanism in the object contextual representation module to effectively extract diverse phase discontinuities in isolated regions.
Note: The data presented above are the experimental results obtained from their respective datasets.
Table 15. Classification of phase unwrapping network dimensions.
Table 15. Classification of phase unwrapping network dimensions.
MethodApplication DomainTechnical Principle DimensionLearning Paradigm
UnwrapGAN [40]Biomedicine (off-axis DHM).Pix2Pix-based GAN with an encoder-decoder structure; incorporates adversarial and L1 loss functions.Supervised learning (paired data).
Res-Unet [41]Digital holographic interferometry; industrial inspection (off-axis DHM).Residual U-Net architecture enhanced with weighted skip-edge attention mechanisms.Supervised learning (training on simulated data).
DeepLabV3+ [42]Optical metrology under complex noise conditions (off-axis DHM).Semantic segmentation framework using dilated convolutions and spatial pyramid pooling.Supervised learning (augmented noisy data).
PhaseNet2.0 [43]Optical metrology; SAR interferometry (off-axis DHM).Fully convolutional DenseNet architecture; utilizes composite loss functions (cross-entropy, L1, and residual loss).Supervised learning (multimodal noisy data).
SFNet [45]Industrial optical inspection; 3D shape reconstruction (off-axis DHM).Transformer-based encoder with MLP decoder and self-attention mechanism.Supervised learning (combined simulated and real data).
DenSFA-PU [46]Digital holographic interferometry (off-axis DHM).Densely connected network + spatial feature aggregator (SFA)Supervised learning (composite loss function)
VDE-Net [47]Biomedical imaging (off-axis DHM).Two-stage deep architecture incorporating dilated convolutions and weighted edge attention mechanism.Supervised learning (attention-guided learning strategy).
U2-Net [49]Dynamic flame imaging; complex surface structure measurement (in-line DHM).Nested U-shaped residual blocks with deep supervision loss for multi-scale feature extraction.Supervised learning (multi-scale fusion).
PUDCN [50]Fiber optic interferometry, industrial defect detection (off-axis DHM).Deformable convolution layers with a pyramid feature selection module for dynamic representation.Supervised learning (coarse-to-fine two-stage training).
DLCDIFPU [51]Industrial inspection (in-line DHM).Deep region segmentation combined with central difference information filtering (CDIF) and path tracking.Supervised learning (simulated phase maps with real phase labels).
C-HRNet [52]Precision metrology, biomedical imaging, industrial inspection (off-axis DHM).High-resolution network (HRNet) with object-contextual representation (OCR) module for semantic segmentation.Supervised learning (mixed simulated dataset and real data using cross-entropy loss).
Table 16. Comparative analysis of deep learning mechanisms in extensive preprocessing and perturbation assumption constraints.
Table 16. Comparative analysis of deep learning mechanisms in extensive preprocessing and perturbation assumption constraints.
MethodCore MechanismAdvantagesLimitationsConclusion
CNN+ZPF [53]Employs a CNN to automatically identify background regions in DHM, fitting phase aberrations via Zernike polynomial expression.High-accuracy background detection with automated compensation.Limited generalizability and dependency on annotated datasets.Enables long-term phase distortion correction without manual intervention by integrating automatic segmentation with model-based phase fitting.
PACUnet3+ [54]Constructs an encoder-decoder architecture based on Unet3+ to learn the mapping between object holograms and sample-free references, enabling data-driven background subtraction for phase correction.Eliminates reliance on physical modeling while maintaining high accuracy.Computationally intensive and requires large-scale training data.Achieves accurate PAC using data-driven hologram reconstruction, bypassing explicit physical model assumptions.
SSCNet [56]Combines a Zernike polynomial-constrained phase model with self-supervised learning and sparse optimization to estimate phase aberration coefficients using a single hologram input.Capable of compensating for dynamic phase distortions without reference data.Sensitive to initial conditions and exhibits slow convergence.Enables adaptive correction of phase errors in dynamically varying systems through self-supervised optimization aligned with physical constraints.
Table 17. Comparative analysis of deep learning mechanisms addressing the manual intervention issue.
Table 17. Comparative analysis of deep learning mechanisms addressing the manual intervention issue.
MethodCore MechanismAdvantagesLimitationsConclusion
CNN+ZPF [57]Utilizes a CNN to segment background regions in the phase image, followed by Zernike polynomial fitting to decouple sample-induced phase from system aberrations, enabling fully automated compensation.Enables high-throughput processing with simple operation.Limited robustness under complex background textures or high noise.Facilitates rapid and automated background identification, making it particularly effective for high-throughput applications, such as cell morphology analysis.
Two-stage GAN [58]Implements a two-step generative adversarial network: the first network reconstructs structural edges of occluded regions, while the second completes the missing content, thereby generating sample-independent reference holograms for phase correction.Strong capability for large-area restoration with realistic details.Training is unstable; performance degrades when occlusion exceeds 40%.Applies unsupervised adversarial learning to restore phase data in highly occluded conditions, outperforming traditional methods in challenging scenarios.
Table 18. Classification table of phase distortion compensation network dimensions.
Table 18. Classification table of phase distortion compensation network dimensions.
MethodApplication Scenario DimensionTechnical Principle DimensionLearning Paradigm Dimension
CNN+ZPF [53]Biomedical applications and real-time phase analysis (off-axis DHM).Data-driven + physical model fitting.Supervised learning.
PACUnet3+ [54]High-precision reconstruction of microstructures (off-axis DHM).Data-driven (end-to-end hologram generation).Supervised learning.
SSCNet [56]Adaptive phase correction in dynamic environments (off-axis DHM).Physical model + self-supervised sparse constraints.Self-supervised learning.
CNN+ZPF [57]Restoration in complex and large-field-of-view samples. (off-axis DHM).Data-driven (GAN-generated restoration) + edge priors.Unsupervised learning.
Two-stage GAN [58]Single-frame unwrapping in noise-intensive conditions (off-axis DHM).Multi-scale feature fusion + contextual semantic enhancement.Supervised learning.
Table 19. Summary of deep learning-based methods for phase distortion compensation.
Table 19. Summary of deep learning-based methods for phase distortion compensation.
YearMethodSSIMMAERMSEPVSuggestions for Improvement
2021CNN+ZPF [53]0.9430.021//Enhance generalization capabilities to accommodate complex cell morphologies or dynamic noise environments.
2023PACUnet3+ [54]0.93/0.09630.6308Optimize the network architectures to minimize computational complexity and enhance real-time performance.
2023SSCNet [56]0.99240.0120.4415.490Investigate more robust sparsity-constrained optimization strategies to accelerate convergence and improve stability.
2017CNN+ZPF [57]0.9270.028//Enhance the precision of background segmentation in scenarios for complex backgrounds or low-contrast sample features.
2021Two-stage GAN [58]0.98850.0029//Optimize GAN training strategies to ensure repair consistency and detail fidelity for high-occlusion-rate samples.
Note: The data presented above are the experimental results obtained from their respective datasets. Evaluation Metric: The formula for calculating PV is as follows.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, W.; Liu, L.; Bu, Y. Application of Deep Learning in the Phase Processing of Digital Holographic Microscopy. Photonics 2025, 12, 810. https://doi.org/10.3390/photonics12080810

AMA Style

Jiang W, Liu L, Bu Y. Application of Deep Learning in the Phase Processing of Digital Holographic Microscopy. Photonics. 2025; 12(8):810. https://doi.org/10.3390/photonics12080810

Chicago/Turabian Style

Jiang, Wenbo, Lirui Liu, and Yun Bu. 2025. "Application of Deep Learning in the Phase Processing of Digital Holographic Microscopy" Photonics 12, no. 8: 810. https://doi.org/10.3390/photonics12080810

APA Style

Jiang, W., Liu, L., & Bu, Y. (2025). Application of Deep Learning in the Phase Processing of Digital Holographic Microscopy. Photonics, 12(8), 810. https://doi.org/10.3390/photonics12080810

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop