1. Introduction
Correlation imaging is a technique based on second-order intensity correlation [
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12]. In traditional correlation imaging, a beam emitted by a light source is split into two beams, which then travel along different optical paths, usually referred to as the signal and the reference beam. The signal beam passes through an object and the intensity of transmitted light is measured by a detector, whereas the reference beam does not interact with the object. The image of the object is retrieved by measuring the correlations between the intensities at the two detectors [
1,
2,
3]. In 2008, a computational correlating imaging system was proposed by Shapiro [
13]. In this system, the source is a programmable light source [
8,
13], which is realized using a programmable spatial light modulator (SLM) and a laser [
13]. Later on, it was found that a programmable light source can also be achieved using a digital light projector (DLP) [
14,
15], which itself contains a digital micro-mirror device (DMD) and a light emitting diode, to produce structured digital light fields. Compared to SLM, DMD is characterized by a higher modulation speed [
16]. The programmable light source creates deterministic illumination patterns at the object’s position, such that the reference beam is no longer needed, and only a detector is required for imaging [
8,
13]. This greatly simplifies the overall imaging system and makes its application much easier. Correlation imaging techniques have several practical advantages compared to traditional imaging, including higher sensitivity and robustness [
4], which have been exploited in different fields, such as optical lithography [
17], remote imaging [
18], microscopy imaging [
19], X-ray imaging [
20] and imaging for occluded objects [
7]. In all the above applications, the spatial resolution is a crucial factor [
21] and methods to improve the resolution have received widespread attention in recent years [
21,
22,
23,
24,
25].
In traditional imaging systems, the resolution is limited by the Rayleigh criterion [
5], i.e., by the properties of the point spread function (PSF) [
26,
27]. The same happens for correlation imaging systems when the pixel size of the detector is much smaller than the average size of the speckles [
21,
22]. In general, the resolution is quantified by the full width at half-maximum (FWHM) of the PSF, which is approximately equal to the average size of the speckles [
5,
21,
28]. Several schemes to improve the spatial resolution of correlation imaging have been proposed. They may be divided into two main classes according to whether they exploit physical principles or are based on image reconstruction. Among physical methods, Han et al. reported a two-arm microscope scheme that employs second-order intensity correlation imaging to narrow the PSF [
29]. Shih et al. reported a super-resolution method that uses the spatial frequency filtered intensity fluctuation correlation to reduce the FWHM of the PSF [
30]. The narrowing of the PSF by the higher-order correlation of non-Rayleigh speckle fields has been also reported [
31]. A sub-Rayleigh resolution experiment has been performed via the spatial low-pass filtering of the instantaneous intensity to narrow the PSF [
21]. Li et al. used localization and thresholding to reduce the effect of the PSF in a correlation imaging system [
32]. Preconditioned deconvolution methods [
33] and the use of the coefficient of skewness [
34] have been also suggested. Among the schemes based on image reconstruction, compressed sensing correlation imaging reduced the effect of the PSF on the imaging quality, using sparsity constraints to improve the spatial resolution of correlation imaging [
24,
35,
36,
37]. The use of deep neural network constraints [
38] and sped up robust features for a new sum of modified Laplacian (SURF-NSML) [
39] have been also suggested, and high-resolution correlation imaging through complex scattering media has been obtained via temporal correction [
40].
In this paper, we focus on how the fluctuation characteristics of the second-order correlation may be exploited to improve the resolution of correlation imaging. The paper is organized as follows. In
Section 2, we review the concept of resolution for traditional correlation imaging techniques, whereas, in
Section 3, we analyze the fluctuation characteristics of the second-order correlation function. It is shown that the PSF of correlation imaging can be narrowed via the fluctuation information. It is also shown that the grayscale information of a grayscale object can be extracted through this fluctuation information. Based on these characteristics, the high-resolution image of the object can be obtained by using the fluctuation information. In
Section 4, we describe our experimental setup and results, which confirm the theoretical predictions.
Section 5 closes the paper with some concluding remarks.
3. Enhance the Resolution via the Fluctuation Information
If we consider that the power of a light source cannot be kept stable,
in Equation (
5) and
in Equation (
6) should be substituted with
and
, respectively. Fluctuations of
lead to fluctuations of
and
[
41]. In other words, the two quantities
and
are statistically fluctuating functions. In order to account for this effect, we employ the cumulant-generating function
, defined as
where
is the nth-order cumulant of
,
, and
. The nth-order cumulant is given by
From Equations (
7) and (
8), we see that the second-order cumulant of
is given by
According to Equation (
9),
contains information about the fluctuations of
, due to the fluctuations of the light source. Upon the comparison of Equation (
5) with Equation (
9), we also see that this quantity has a narrower PSF than that of
. Then, we integrate over
and obtain the new imaging information as
In other words, using
instead of
to image the object, one has a narrower PSF. From Equation (
9), we also see that the fluctuations of
lead to fluctuations of
and, in turn, after integrating over
, we obtain an overall narrower PSF of the imaging system
If we use
,
and
to denote
,
and
, respectively, we may write, according to Equation (
6) and Equations (
10) and (
11),
where
n is the number of iterations. The meaning of each iteration is that we go one step further in considering cumulants to extract the fluctuation information. From Equation (
12), we find that
is the information consisting of all
.
is the fluctuation information of
when
n ⩾ 1, which is equivalent to
for
n = 0.
In Ref. [
41], the authors have presented a scheme to enhance the resolution of correlation imaging, referred to as second-order cumulant ghost imaging (SCGI). The imaging information of the SCGI can be written as [
41]
where
is the fluctuation information of
.
is cross-information, which reduces the resolution of
[
41]. From Equation (
10) and Equation (
13), we see that
includes
and cross-information
, and that
greatly enhances the resolution but
decreases the resolution. In the framework of SCGI,
is the cumulant-generating function of
, and
is the second-order cumulant obtained from
. Thus,
includes not only the fluctuation information of
for different
points, but also the cross-information generated correlating
and
for different
and
(
) [
41]. This is the reason for
appearing. In our case, we have a spatially resolving charge-coupled detector
instead of the bucket detector of [
41]. If the distance between the object and the
is smaller than their longitudinal coherence length, we consider
, i.e., the cumulant-generating function of
. The second-order cumulant obtained by
is the fluctuation information of
. In other words, the term
disappears in our scheme, and this leads to a better resolution limit than that of SCGI.
By explicitly evaluating Equation (
12), we see that the PSF at the n-th iteration is the 2
n-th power function of sinc
2. Since the PSF of
is the 2-th power function of sinc (see Equation (
6)), the FWHM of the PSF for
is narrower by a factor of
than that of
. For a binary object, the resolution of the correlation imaging can be thus enhanced by a factor
. In order to address a concrete example, a schematic diagram of our numerical simulations is shown in
Figure 1a. We take 20,000 measurements for every simulation. The wavelength of the source is
= 500 nm, and the radius is
R = 1.28 ×
m, whereas
=
= 0.022 m. The object consists of two squares, as shown in
Figure 2a. The side length of both squares is
r = 1.6 ×
m. The distance between the centers of the two squares is
d = 4 ×
m. The distance between the object and the
is
= 0.002 m. According to Equations (
6)–(
8) in Ref. [
42], we have that
is less than the spatial longitudinal coherence length between the object plane and the
.
At first, we reconstruct the image by the
and
, respectively. Results are shown in
Figure 2b and
Figure 2c, respectively. From
Figure 2b, we see that imaging of the object is not possible using the
due to the interference effect of light. From
Figure 2c, we find that also
fails to provide acceptable imaging. This is because the distance between the square centers is smaller than allowed by the Rayleigh criterion, i.e,
. Using SCGI (i.e.,
), we obtain
Figure 2d, which shows a better resolution, since the PSF of
is narrower than that of
. Finally, we have reconstructed the image using
and
, respectively. The results are shown in
Figure 2e,f. We see that the resolution provided by
is better than
. This is due to the disappearance of the
term. Finally, if we then compare
Figure 2e to
Figure 2f, we see that the resolution provided by
is better than that of
, since the FWHM of the PSF of
is smaller than that in
. For a binary object, the resolution becomes better and better as
n grows.
In order to further compare the quality of the reconstructed image, we evaluate the Peak Signal-to-Noise Ratio (PSNR), which is defined as [
43]
where
,
and
denote the reconstructed image and the original object, respectively;
N is the total number of pixel points for each of
; and
is the maximum gray value of
. By using Equation (
14), we calculate the PSNRs of the normalized images in
Figure 2b–f, obtaining PSNR = 15.86, 17.19, 20.65, 23.41 and 24.17, as shown in
Table 1. This is consistent with our previous analysis and further confirms the reliability and accuracy of the new protocol.
Moreover, for a grayscale object, the resolution may be enhanced using
. In particular, according to Equation (
12), we see that the information about the higher grayscale regions of the object can be gradually extracted from
with increasing
n. As a consequence, we can extract high-resolution images of the different grayscale regions by combining
with some centroid localization algorithms. Then, the full high-resolution image of the grayscale object can be constructed by combining the images of the different regions, as in single-molecule localization microscopy (SMLM) [
44]. Concerning the centroid localization, several algorithms may be used, such as thresholding [
45] and DAOSTORM [
46]. For the sake of simplicity, we employ a simple thresholding method and obtain the maximum value of
by considering
and setting the threshold
where 0 <
P < 1. The smaller is
P, the higher is the accuracy and the slower is the overall speed of the imaging technique, i.e., a trade-off level has to be set according to the specific application. Here, we set
P = 0.5 and obtain the threshold by substituting
P into Equation (
16). Only the values of
larger than threshold are accepted to set the centroid of the image. The overall imaging process thus proceeds as follows. At first, we use
to reconstruct the high grayscale regions of the object. Then, we obtain the centroids of these regions by thresholding and eliminate the information about the centroid by blocking its position or using some specific algorithms. Then, we again use
to image the low grayscale regions. As an example, we consider an object composed of three squares with sides equal to 2.56 ×
m and a center distance 4 ×
m, as shown in
Figure 3a. The transmittances of the squares are set to 1, 0.75 and 0.5, respectively, from left to right. The other parameters are set to the same values as in the previous simulations. The reconstructions obtained by
,
and
are shown in
Figure 3b–d. As is apparent from the plot, the information in the high grayscale region can be extracted and improves for increasing
n.
Since there is a grayscale difference between
and
, the peak of
Figure 3d is reduced by a factor of
where
is the peak coordinates of
Figure 3d. To better recover the grayscale information, we multiply
Figure 3d by this compensation factor
, and, by this procedure, the image of the left square can be retrieved. We use thresholding to obtain the centroid of
Figure 3d and eliminate the information about the centroid from the imaging information. Then, we use the steps described above to obtain the image of the middle square, as shown in
Figure 3e. Similarly, we obtain the image of the right square, as shown in
Figure 3f. After eliminating the information of the three centroids, we obtain the image in
Figure 3g. Upon the comparison of
Figure 3d with
Figure 3g, we see that
Figure 3g contains only a fraction of the information of
Figure 3d. This is because
Figure 3g is already included in
Figure 3d, but its intensity is negligible compared to
Figure 3d. At this point, we assume that the imaging of all the regions has been completed and the full image may be obtained by combining the images of the different grayscale regions, i.e., combining the images of the left square from
Figure 3d multiplied by the factor
, the middle square of
Figure 3e and the right square of
Figure 3f. We then normalize the image, as shown in
Figure 3h. We also compare the quality of the reconstructed images quantitatively by using Equation (
14). The PSNRs of the normalized images of
Figure 3b,
Figure 3c and
Figure 3h are 16.43, 22.52 and 25.18, respectively, as shown in
Table 2. The results show that our protocol can be used for grayscale object imaging and has a better resolution than traditional correlation imaging.
4. Experimental Results and Discussion
To verify our theoretical results, experiments are carried out. We use the simple single-arm setup shown in
Figure 4a (which is equivalent to
Figure 1a [
14,
21], because the intensity distribution on the
(
) can be obtained by calculations). The light source is a projector (XE11F), and there is a digital mirror device (DMD) in the source. The object is a double-slit object with width
a = 0.223 mm, a distance between the centers of the two slits
b = 0.445 mm and a slit height equal to
g = 1.114 mm.
The object is part of the resolution target (USAF1-228 LP/MM), as shown in
Figure 4b. The distance between the source plane and the object is 25 cm. A camera (MV-VEM033SM) is taken, used as
. The distance between the object and the
is 2 cm. In this experiment, we employ both the traditional correlation imaging and our method based on
. All the results are obtained by averaging over 20,000 exposure frames. For
, we extract
every 2000 steps. The FWHM of the different methods can be obtained by imaging a single pixel [
31], and for
,
and
, we obtain 69 pixels, 49 pixels and 35 pixels, respectively, as summarized in
Table 3. Then, we obtain the image of the object by substituting
into Equation (
12). The results are shown in
Figure 5.
The image of the object, as obtained by traditional correlation imaging, is reported in
Figure 5a, which shows that the two slits cannot be separated.
Figure 5b shows the image of the object obtained using
, where the two slits are now visible, i.e.,
provides a better resolution than the traditional correlation imaging. This is because the resolution of the system is limited by the PSF, and
has a narrower PSF than
.
Figure 5c shows the image of the object as obtained by
. The image is apparently clearer than those in
Figure 5a,b. A quantitative comparison can be obtained from the normalized horizontal marginals, shown in
Figure 5d–f. The PSF of
is indeed the narrowest among the three models. The above experimental results show that the resolution obtained from
is the best, and the resolution of
is better than that of
(
). This means that the resolution provided by
improves for increasing
n, which is consistent with our theoretical analysis.
Since the double-slit object is a real object, we cannot obtain perfect data for
, and therefore we are unable to calculate the PSNR for the images in
Figure 5a–c by Equation (
14). In order to compare quantitatively the quality of the reconstructed images, we evaluate the resolution using the intensity at saddle point position
for the images in
Figure 5d–f, respectively. The value of
can be extracted from
, where
is the center of the left slit in the object, and
is the center of the right slit. According to Ref. [
47], for a binary object, the double slit cannot be separated when the normalized intensity at
is greater than 0.81. The normalized intensity at
for the different methods is summarized in
Table 3.
Table 3 shows that the double slit cannot be separated by traditional correlation imaging. However, it can be separated by
and
. From the FWHM of the PSF in
Table 3, we see that the resolution of
is improved by a factor of
compared to that of
. This is consistent with our theoretical analysis.
Finally, in order to show that the proposed method can be applied also to the imaging of grayscale objects, we take the double-slit object considered above and cover the left slit with a film. Results are shown in
Figure 6. As is apparent from the plots, the image cannot be obtained by traditional correlation imaging (see
Figure 6a), but it is well retrieved by
(see
Figure 6b). This is because the PSF of
is narrower than that of
. As a matter of fact, low grayscale regions are discarded if
n in
is set too high. However, even without using centroid localization algorithms, the resolution of the imaging may be improved by setting the appropriate
n according to the grayscale distribution of the object. We then obtain the image by combining
with a thresholding method. The detailed imaging process is the same as that used for the simulations, and the result is shown in
Figure 6c. The improvement in the resolution is apparent and is confirmed by the normalized horizontal sections shown in
Figure 6d–f. Since the object is a grayscale one, the intensity at
is not equal to the intensity at
. Compared to the image of the right slit (higher grayscale area), the image of the left slit (lower grayscale area) is more difficult to observe, and thus we further assess the resolution of the images in
Figure 6d–f by calculating the ratio between the intensity at
and at
. Results are shown in
Table 4.
Table 4 shows that the double slit cannot be separated by the traditional correlation imaging, whereas it can be separated by
and
with the thresholding method. From
Table 4, we also see that the resolution of
with the thresholding method is the best one. This is because the PSF of
narrows as
n grows, and the high-resolution image of the different grayscale regions can be obtained by thresholding. The image of the grayscale object can be thus obtained by combining the images of the different grayscale regions. The experimental results are in excellent agreement with our theoretical analysis.
Since contains information about the fluctuations of and those of when n ≥ 1, the number of measurements required for imaging using increases as n grows. We thus conclude that traditional correlation imaging is, in general, faster than our scheme, whereas our scheme shows a better resolution than traditional correlation imaging.