1. Introduction
The first experiments on imaging techniques based upon the correlation of intensity fluctuations measured at two or more spatial–temporal points for thermal sources of light were carried out by R. Hanbury Brown and R. Q. Twiss (HBT) [
1,
2,
3,
4]. From the outset, it was clear that this first exploration in quantum optics had many advantages over conventional telescopes and amplitude interferometers. Among these advantages are the followsing:
- ▪
HBT does not require the propagation of collected light beams to combiner units as in existing amplitude interferometers. Routing optics and beam combiners are eliminated.
- ▪
HBT does not require nanometer-level control of nor nanometer knowledge of the relative positions of the optical components.
- ▪
The signal-to-noise ratio (SNR) does not suffer from beam splitting and the throughput losses that plague amplitude interferometry.
- ▪
HBT does not require expensive diffraction-limited optics. Photo detective hardware consists of simple, flux collecting apertures.
- ▪
HBT is insensitive to atmospheric turbulence and therefore can be used by ground-based as well as space-based facilities.
- ▪
Above all, each aperture transmits its own intensity fluctuation signal. All pairs of these signals are correlated via a separate computational facility. Thus each pair of apertures are electronically coupled so that the relative distance vectors between apertures (the baseline vectors) can be arbitrarily large, allowing for the possibility of indefinitely fine image resolution.
Unfortunately, for HBT correlators, the measurement signal is proportional to the square of the average light intensity. Most astronomical light sources are not bright enough to determine stellar diameters with reasonable integration times (hours or minutes rather than millions of hours). Until very recently, integration time has been the major obstacle for measurements of stellar diameters. A second difficulty is that HBT has been concerned with the calculation of stellar diameters modeled as radially symmetric intensity distributions that require only the measurement of coherence magnitudes. For much broader astronomical investigations, an algorithm capable of two-dimensional imaging is needed. This requires both coherence magnitude and coherence phase retrieval. Finally, in the interests of astronomy, we concentrate on imaging an illuminated body against a black sky where the positions of the black pixels are unknown but are determined by ICI.
The present discussion is a survey of a sequence of four papers [
5,
6,
7,
8] that developed an algorithm for two-dimensional imaging that enormously reduces the integration times for the HBT effect while retaining the advantages listed above. First of all, Reference [
5] developed the noise reducing phase retrieval (NRPR) algorithm. This led to three papers [
6,
7,
8] produced by the author as an independent researcher, and these three most recent papers were the most instrumental in solving the curse of the weak signal. The four articles were arranged in both topical and chronological order with rigorous technical detail. While the salient features of the mathematical development are herein presented, this paper encourages its users to appreciate the rather simple process for the determination of high-resolution images from the extremely noisy database. We emphasize that the HBT technique has been called “Intensity Correlation Interferometry”. However, this significant step forward merits the name “Intensity Correlation Imaging” (which we designate as ICI hereafter).
The plan of the present development is as follows. In the next section, we explain the coherence magnitude measurement process and a reformulation of the HBT SNR (signal-to-noise ratio) that reflects the fact that the conditional probabilities arising from the image domain constraints (needed in the Cramér–Rao bound) must be accounted for in any estimate of measurement accuracy as a function of integration time. In other words, the conventional SNR treatments [
1,
2,
3,
4] are inappropriate in the context of two-dimensional image reconstruction. Evidence of this appeared in earlier studies, which showed that conventional estimates of integration time for many imaging examples were grossly conservative. This section explains why the ICI data collection process must utilize coherence magnitudes that are strictly nonnegative in order to correct the HBT SNR.
Next, in
Section 3, we offer an overview of the noise reducing phase retrieval (NRPR) algorithm [
5]. This section postulates a complete knowledge of the positions of the black pixels, wherein an illuminated, square foreground of a specified size is centered amid the constrained black pixels of the background. This is termed the “Box”. Using NRPR to compute a multitude of “Box” sizes, we present computational studies that examine the correct integration time under the assumption that integration times are the same for all aperture pairs (i.e., all baselines in the u-v plane). The results show an enormously more precise estimate for integration time as a function of image size and accuracy.
Section 4 describes the technique wherein the black pixels in the background having specified the “Box” size are able to determine the zero-noise image. The process embeds the NRPR algorithm (equipped with the correct integration times) within a stochastic search algorithm [
8]. By varying the box sizes on each NRPR run, the occurrence of two completely correlated images (aside from 180-degree rotations and translations) indicates the correct zero-noise image. The correct identification is attained via very few iterations.
Section 5 discusses theorems that lead to an even faster identification of the correct, zero-noise image. Then, in the
Section 7, we discuss further possibilities for both the advancement of the ICI algorithm within the photonic community and its application to advanced image resolution.
2. ICI Data Collection Process
For simplicity, we postulate a set of flux collecting apertures forming a square grid on the u-v plane. Furthermore, the parameters of all apertures, including the integration times (denoted by
), are identical. For a given pair of apertures, the fundamental data collected by Brown and Twiss to obtain an estimate of the coherence magnitude are a running time average of the intensity fluctuation cross-correlation as measured by each pair of intensity sensors. We designate
as the baseline vector in the u-v plane associated with the two apertures. Here, we reconsider the statistics of intensity fluctuation cross-correlation that prevail when the object of study has very low spectral radiance. For the present discussion, we assume that (1) the averaging time,
, is much larger than the detector response time,
; (2) the field is quasi-monochromatic, i.e., that it is confined to a narrow spectral band,
, centered at frequency
such that
; (3) the light is approximately cross-spectrally pure, i.e., the normalized mutual coherence takes the form
, which implies that the coherence time of the light satisfies
(
denoting the optical bandwidth); (4) the slow detector assumes that
; and (5) the thermal light radiated by the targets to be imaged has very small total photodetection events per second per Hertz (a restriction almost universally valid). For simplicity of exposition, we also assume that the intensity and mutual coherence magnitudes are approximately constant over the spatial extent of each aperture pair. In other words, the partial coherence factor calculated in Reference [
2] is approximately unity.
Let us introduce some further notation. The symbol
denotes the ensemble average, and
denotes the time average, defined as follows:
The prefixes “” and “” indicate fluctuations about the ensemble average and about the time average, respectively. Finally, let the electronic outputs of the two detectors be denoted by . Then, the time-averaged intensity fluctuation cross-correlation (the principal datum of the measurement) is .
The statistics of the time-averaged intensity fluctuation cross-correlation (the quantity
) has been treated in great detail in [
1,
2,
3,
4]. Accordingly, assumption (5), which holds for all but the brightest objects, leads to the conclusion that in the presence of conditions (1)–(4) and at any particular integration time where
is a Gaussian random variable, the mean and standard deviation are as follows:
is a zero mean, unit variance Gaussian random variable where and are statistically independent if . Note that as the integration time approaches infinity, converges to a constant multiplier times the magnitude squared of the normalized degree of optical coherence, . This is the primary discovery of Brown and Twiss for the case of thermal light.
At this point, we observe that the coefficient of the noise term can be determined at the outset by calibrating the detectors, using the aperture parameters, and measuring the intensity observed by the individual aperture sensors. In other words, the quantity
can be obtained before any intensity fluctuation cross-correlation measurements are attempted. Hence, it is convenient to divide
by this factor, thereby normalizing the noise component to unity standard deviation. Therefore, we can write the equation as follows:
may be recognized as the zero baseline SNR for a unit integration time, which can be determined at the outset via measurements from the individual apertures and equipment calibrations. It is important to note that the statistical properties of and are completely determined by the parameters aside from and . Any statistical data regarding supposedly specific radio astronomy, adaptive optics, or other noise distributions are carried within these parameters.
Clearly, the meaning of is the same as the time-averaged cross-correlation, and they are interchangeable. However, is more easily interpreted, as it is the “empirical SNR” that one can use to compare the signal component of the measurement to the noise component, whose standard deviation is always one. Hence, as the integration time increases, one can compare to unity to ascertain if the desired SNR is reached with a 50% probability. can be used to illustrate the process of experiment design and data gathering and demonstrates the essential differences between the measurement of stellar diameters versus the capture of two-dimensional images.
Let us discuss the initial experiment design as practiced by Brown and Twiss. Obviously, one cannot prolong the running time average to an impractical degree. At the outset, one must estimate the length of the final time average, denoted here by
so that the final empirical SNR is roughly of the same magnitude as the desired SNR,
SNRd, allowing for an approximately unit standard deviation fluctuation. Based upon the above description of the estimation process, Equation (3) leads to the following specification of the limiting integration time, denoted by
:
where
has been replaced by
in (3) to arrive at this expression and
is some estimate of the coherence magnitude. As expected, any reduction of the factor
(traditionally characterized by the inefficiency factor) lengthens the necessary integration time so that accurate equipment calibration is essential. We also see that design of in-the-field measurements needs an a priori estimate of
.
The process adopted by Brown and Twiss to measure stellar diameters involves measuring only the coherence magnitudes and curve fitting the results to a simple, radially symmetric model of the stellar intensity pattern. Because of the simplicity of the technique, one finds that reasonable accuracy for the Brown–Twiss calculation may be obtained using a
close to unity (note that the normalized coherence has the properties
). In fact, some equalities in Refs. [
1,
2,
3,
4] actually assign the factor
to unity, although a more precise estimate is likely to be
. This estimate of integration time is roughly correct for stellar interferometry and provides a good starting value to the needed integration time. Also, the quantity
may be considered an efficient estimator of the coherence magnitude squared in that it is inherently positive and does not depend on the value of the quantity being estimated (namely the normalized magnitude of the optical coherence). In contrast, a fully two-dimensional image of a complex object, based on interferometric experience, requires at least
. Thus, in this case, the method used by Brown and Twiss to determine a two-dimensional image is clearly inappropriate.
The inconvenient aspect of the raw, noisy data delivered by the empirical SNR is that it cannot be guaranteed to be nonnegative, and therefore, the noisy data cannot be considered an efficient estimator for the purpose of two-dimensional imaging. Further, the Cramér–Rao bound entails that the Fisher information include conditional probabilities that are necessary for phase retrieval algorithms that demand positivity. Indeed,
can assume negative values with high probability when
is sufficiently small. Before proceeding further, we remedy this by formulating a nonnegative function of the measurements, without somehow ignoring negative-valued data. We take some pairs of independent measurements, say, the following:
then multiply one of them by the unit imaginary, add the results, evaluate the square root of the magnitude, and divide by
, which is known in advance, to obtain the following:
where the nonnegative quantity,
, now takes the place of the coherence magnitude measurement. Substituting the expressions in (5) for
, we have the following:
However,
is equivalent to
, where
and
are independent Gaussian random variables with zero mean and unit variance. Therefore, the following nonnegative quantity may be taken as a single measurement:
The quantity
is the SNR of the combined measurement at position
and observation time
. It is possible to control the relative SNR values of the various baseline measurements used as image reconstruction data. This can be calculated by taking an a priori estimate of
, calling it
, and choosing to stop observation at duration
such that the signal-to-noise ratio is
:
where
is the same positive constant for all
. Consequently,
becomes
We remark that the formulation of permits the introduction of whatever a priori information on the image that might be available (overall dimensions, foreground configuration, etc.).
Before leaving this subject and for later use, we establish the formulation, based upon (10), that is closest to the traditional practice, namely the situation wherein there is no a priori information on
for any baseline. Under this circumstance, it is obvious that the desired SNR for all baselines cannot be obtained unless
in which case the noise model becomes
To close this section, we recognize the most important difference between stellar diameter determination via ICI and two-dimensional imaging by ICI. In the former, only measurements of coherence magnitude are made, while by virtue of the Van Cittert–Zernike theorem, the latter requires both magnitude and phase. Hence, phase retrieval algorithms are needed that use additional image domain information to supplement the lack of available phase information. This process must operate in the presence of very significant noise. The next section reviews the noise reducing phase retrieval (NRPR) algorithm, which has been shown [
5] to be effective in handling very large amounts of noise.
3. Noise Reducing Phase Retrieval Algorithm and Reduction of Integration Times
In References [
5,
6], we have described a new phase retrieval algorithm and its application to imaging in the context of ICI. The algorithm recognizes that much of the noise in the averaging data is inconsistent with the image domain constraints and can be rendered harmless if both the Fourier domain and Image domain constraints can be made to intersect. The algorithm accepts the noisy coherence magnitude data and uses a relaxation technique to project these data onto a subspace wherein the image domain constraints can be satisfied. When it is run to completion for a single set of coherence magnitude data, we have shown by computational examples that the impact of much of the noise is eliminated, even for extremely large amounts of noise. For this section, it is assumed that the positions of the black pixels are known a priori.
As mentioned previously, we assume that measurements of the coherence magnitude are accumulated at points on a square grid, N pixels on the side. The image is likewise pixelated into a square grid. Thus, both the image and the coherence magnitudes compose nonnegative matrices of the same dimensions. For this discussion, it is supposed that the image consists of a bright object against a black (zero intensity) sky and that we correctly identify a boundary enclosing the bright object, such that the object cannot be translated or rotated by 180° without being clipped. The foreground and background pixels are defined as the pixel sets that are contained within and outside of, respectively, the boundary. It is assumed that the size of the background is roughly three times the size of the foreground.
It is convenient to consider both image and coherence as
-dimensional vectors. To be specific, if
denotes the two-dimensional matrix representing the image, then
, where
is the Kronecker algebra operator that stacks the columns of the two-dimensional matrix into a single-column vector. We further define the following notation:
In this discussion, the measured coherence magnitudes are assumed to have the form (12). Thus, the integration time is presumed to be the same for all baselines. Occasionally, we may wish to work with the two-dimensional matrix forms of , etc. These are distinguished by an under-bar, as in , etc., and the Fourier transform of is denoted by , where indicates the Hadamard product.
The following phase retrieval algorithm [
5] is considered:
where
and
denotes the Hadamard product.
denotes the Frobenius norm. If
A is a complex-valued vector or matrix,
denotes the matrix wherein each element of
A is replaced by its absolute magnitude. Likewise,
is the result of replacing each element of
A by its real part. If
A and
B are real matrices of equal dimension, then each element of
is the larger of the corresponding elements of
A and
B. Element-by-element inequalities are denoted by double inequality symbols, e.g.,
A >> B,
A << B, etc. Finally,
is a vector of statistically independent, uniform random variables in
.
A brief explanation of the algorithm is in order. Steps B, C, D, and F are taken from the Hybrid Input/Output (HIO) algorithm of Fienup [
9]. For small or zero-noise data, HIO converges rapidly, whereas other methods are prone to very slow convergence. On the other hand, if noise is more than unity, HIO fails to converge and in fact will sustain rapid oscillatory motions. Thus, steps B and E are necessary.
The fundamental input to the algorithm is a noisy batch of coherence magnitude measurements as formulated in (12). The algorithm begins with a random guess at the image such that each pixel intensity is uniformly distributed within a small interval,
. There follows the core of NRPR consisting of a recursive algorithm of six steps. Step A merely subjects the previous value of the image to Fourier transformation to produce the corresponding coherence (both amplitude and phase). Step B is crucial to noise reduction. It modifies the coherence magnitude data by discounting slightly (through the relaxation parameter
) the existing data values while adding
times the magnitude of the latest coherence estimate. Useful values of
are typically
. This step permits the joint convergence of both the image domain constraints and the coherence magnitude modified data that would otherwise be impossible in the presence of significant noise. An analysis of the algorithm [
5] shows that from iteration to iteration, the noise in the image is inexorably reduced relative to the signal component. The remaining steps, except for E, are very similar to the HIO algorithm of Fienup [
9]. Step C takes
and preserves its phase but replaces its magnitude with the modified coherence magnitude data to obtain the preliminary coherence estimate
. This is essential for the gradual reduction of the constrained pixel intensities. Step D calculates the corresponding preliminary image,
. Step E speeds convergence by ensuring that the constrained pixel intensities are real-valued, and the unconstrained intensities are real and nonnegative. Finally, step F reduces the intensities of the constrained pixels in accordance with the HIO algorithm. The parameter
equals 0.7. After each iterate of the recursive algorithm, the 1-norm of the constrained pixel intensities,
, is computed. In this work, convergence to an image having errors indetectable by visual inspection are obtained with
.
In Reference [
5], various properties of the algorithm were explored. In particular, when run to convergence, the algorithm is equivalent to an operator that projects the coherence magnitude measurement data into the subspace that is consistent with the image domain constraints, thereby eliminating a large portion of the coherence magnitude measurement noise. These conclusions do not depend upon the noise model.
Before leaving this section, we determine the value of the reciprocal SNR,
. Ref. [
8] gives an overview of the imaging examples using the integration times as calculated by Brown and Twiss. These results reveal that the small features of the image converge first. Using the Hirschman entropic uncertainty analysis [
10,
11,
12] shows that the Brown–Twiss integration time is much larger than needed. To correct this,
must include terms involving the constrained pixels. Also, to treat all possible factors, the sum of the number of constrained pixels must appear in all of the terms
in
. Finally, since
is a constant for all coherence magnitude measurements, we conclude the following:
Now, we proceed to determine the value of
. The validity of (17) is explored using randomly generated image examples capable of encompassing a multitude of image sizes. As mentioned previously, we assume that there is an illuminated foreground that contains the object to be imaged and that it is centered within the field of view, while the background contains the constrained black pixels (
). The boundary of the foreground/background can be almost any convex contour, but as it turns out, the most convenient shape is a square with
pixels on the side. This is the contour that separates the unconstrained pixels in the foreground from the black pixels in the background. Reference [
7] designates this contour as the
box.
Now, we consider the constant C. With the above foreground/background prescription, we run the NRPR algorithm (15–16) for many box sizes, some of which are depicted in
Figure 1. To prevent 180-degree rotation of the image during NRPR convergence, a dark square “tab” is placed in the upper left corner of the foreground, within the box.
The intensity of each pixel of the foreground is uniformly distributed in , with all pixels statistically independent. Model (12) is implemented through the precise calculation of in every case. For each , the algorithm is run 20 times for a selected value, and the convergence is checked to see if the correct convergence is achieved in at least 18 out of 20 cases. We note that NRPR always converges so that the background constraints are satisfied, while “correct” convergence is deemed to occur if the foreground image is in agreement with the actual image to the extent that the 1-norm of the error is less than 0.001. Failure to correctly converge is accompanied not only by significant foreground error but also a very noticeable decline in the average intensity of the foreground.
Using this definition of convergence failure, we start with
, then increase
until correct convergence in more than 10% of cases fails (in all cases when
, there was 100% convergence). Then, the next lowest value of
is recorded. This process is undertaken for each value of
. We then summarized the results in
Figure 2, which shows the largest values of
for which convergence succeeds at least 90% of the time as a function of
. The resulting variation of
with
is approximately as follows:
Of course, by the reasoning of (17), the increase in sigma is due to the number of image domain constraints (the number of constrained pixels), denoted here by
. By virtue of the foreground/background proportions,
. Consequently, we have the following:
where
denotes the minimum SNR required for NRPR convergence. This result substantially validates [
6]. Note that in comparison with the conventional value of
corresponding to a desired SNR of unity, the integration time is diminished by the factor
. For a relatively small image that is only 100 pixels on a side, this factor amounts to slightly more than two hundred thousand. Thus the
value is an asymptotic result.
It is also evident that the above relations lead to a simple criterion for NRPR convergence, namely that the product of the
and the number of constraints has a nonzero, positive lower bound:
To test whether or not the above integration time results are sensitive to image content, we examined several non-random images using the same foreground/background arrangement and computational procedure as is described following Equation (19).
Figure 3 depicts the noise-free images being tested, where only the foreground in each case is shown. Corresponding to each recovered image, the stars indicate the largest values of
for which convergence succeeds at least 90% of the time as a function of
. The straight line shows the previously derived estimate for
. At least for the cases tested, there appears to be little sensitivity to image content.
The conclusion to this section is that the formula for the computation of the coherence magnitude values is as follows:
Note that the ratio of the above integration time to the Brown–Twiss calculation is as follows:
where
is typically 10 to 100. Hence, this ratio of the two is many orders of magnitude smaller.
A limitation of the present section is that the set of constrained pixels is determined a priori. To apply ICI to image space objects of unknown configuration, we must be able to discover the constrained set without prior knowledge (except for the fact that we are dealing with a bright object against a black sky). This is treated in the following section.
4. NRPR Characteristics and Stochastic Search
Having obtained the NRPR algorithm with a box size along with its appropriate integration time, we can run NRPR to completion for multiple box sizes. The outcome of each iterate is a hypothesis concerning the correct, zero-noise image. In this section, we show how to embed NRPR within a stochastic search mechanism that quickly discovers the true image. We begin with the stochastic search algorithm approach introduced in [
7] and shown in
Figure 4.
The first step is to collect the intensity fluctuation data and compute all the cross-correlations, using the correct integration time. The stochastic search algorithm uses only these data. The first stage of the search algorithm (shown in green) sets up a “Box” of size . The second loop (shown in red) runs NRPR up to n > 1 times. Each run uses an initial guess of the image such that the intensities of all pixels are statistically independent and uniformly distributed in , where is a small real and nonnegative number. The second part of the loop ends by updating the image correlation matrix. Then (shown in black), we test the correlation matrix to see if there is a pair of fully correlated images among the n NRPR trials (allowing for 180-degree rotations and translations in the field of view). If this is not the case, then we set up a new box size and try again. If there are two correlated images, then we stop the stochastic search because as we will show, both correlated images are the zero-noise image. Generally, excellent image quality is attained when the image domain constraint error, , is less than .
Both the analysis [
6,
7] and thousands of computational runs show that the NRPR algorithm always converges to some
. Thus, we may consider NRPR as a mapping from the intial values
to the converged result, denoted by
:
as well as a mapping from
to
:
One expects that the randomness of
will result in the randomness of
so that
may not be the true, noise-free image
. However, there is one instance where
independently of
, and that is when
equals the true
; we call it
. This is illustrated in
Figure 5. For all cases the coherence magnitude data are fixed, and the results are all obtained after 500 iterates. The upper left shows
equal to the true constraint matrix. The top sequence of results shows a sample of independent trials of
(taken from 10,000 samples), obviously duplicating the true image repeatedly. In contrast, the lower left shows an erroneous specification of
. Repeated trials of
, using different random samples of
each time result in randomly varying images. This basic characteristic of NRPR suggests that at the very least, the correct background region can be discovered by trial and error using NRPR as a test. Of course, a purely trial-and-error approach would entail the evaluation of an enormous number of possible trials (perhaps as much as
), especially for any reasonably large image. A quicker technique rests upon another characteristic feature of NRPR convergence, namely the following: Given random trial values of
(that could include
), there is a nonzero probability that the repeated trials will produce two or more identical repetitions of the correct image. Once such a repetition occurs the identical images produced can be identified as the correct image. The probability of such image repetition grows with the number of correctly assigned elements of
.
Therefore, there is the possibility of identifying the correct image domain constraints with relatively few random trials.
We illustrate the algorithm’s characteristics using the simple spacecraft image first introduced in [
5] and introduced here as
Figure 6. This image is set in a total field of view of 90 by 90 pixels. The NRPR parameters are specified to be as follows:
The choice of these algorithm parameters warrants some explanation. First, consider epsilon. If the epsilon is less than or equal to zero, NRPR devolves into HIO and generally does not converge for large values of noise. Thus, epsilon must be positive. An enormous number of trials of NRPR on a great many different images has shown excellent performance for epsilon in the range epsilon = to .
The first step in developing this example is to calculate the raw coherence magnitude data batch that will be used repeatedly in every sample run. This is given by . The value for is the initial value , which is subsequently adjusted via Step B of NRPR. For this example, we estimate that (it turns out that the minimum value is ). The number of constraints, , is given as a function of the box size, denoted by and by .
With the raw coherence magnitude data in hand and the box size specified, the procedure is to compute a sequence of NRPR cases (each run for 1000 iterations). Each member of this sequence is an instance of . This sequence terminates when there are two images that are identical (aside from the trivial ambiguities and within 0.001 relative error), whereupon these are deemed the true, noise-free image.
We begin by showing a typical example of the image sequence for
, which is a reasonable starting value. The results shown in
Figure 7 are quite surprising. There are two distinct categories of image: a set of random and non-repeatable images (images 1, 2, 4, and 5) and two identical images (numbers 3 and 6) that perfectly match (within the above mentioned error criteria) the correct noise-free image. In the latter, not only are the constrained pixels outside the box correctly prescribed, but all the truly black pixels within the box are also rendered correctly, as are the unconstrained pixels of the noise-free image.
The above phenomena can be explained by the structure of NRPR considered as a discrete time dynamical system with state
, where
k is the iteration number. Here, we distinguish between the assumed value of
and the value of
that includes all of the black pixels belonging to the true, noise-free image, denoted by
. There are two primary noise reduction processes operating in NRPR. The first is step F, which reduces the noise in the constrained region of the state space. Assuming that the specified box is not smaller than the bright object, then
. The second mechanism for noise reduction is Step B. The application of Step B distinguishes two subsets of the initial values of the pixels denoted by
. One subset evolves via the action of Step B into the noise reduction of pixels
that do not contain
. This set converges to an image that retains noise in the foreground of the image, as illustrated by the random configurations in images 1, 2, 4, and 5 of
Figure 7. The second set, however, is the one under which Step B reduces the noise for the entirety of the unconstrained set
. In this case, the whole image converges to the noise-free image (as in images 3 and 6), and the final result is independent of the initial noise in the coherence magnitude measurements. A few more deductions can be made on the basis of this discussion. For example, the probability of obtaining two identical images from a sequence of runs should increase as the box size decreases as long as the bright object can be contained in the box. On the other hand, if the box is too small to contain the bright object, we have an irreconcilable conflict between the image-domain and Fourier-domain constraints so that there can only be random images (as we illustrate below).
Clearly, a trained human operator can use the above stochastic search method to rapidly determine the correct constraints and the image. For example, the results for already indicate the approximate size of the object, so a skilled technician would go immediately to without passing through the intermediate steps and then secure the correct image.
To illustrate this example further, we proceed to examine the statistical characterisitics of the number of trials (for a given box size) needed to secure the noise-free image. The main element determining the computational labor is the number of NRPR trials needed to identify the noise-free image for a given box size. Here, we show typical statistical estimates using histograms calculated in [
7]. For each box size, we obtain 500 statistical samples. Each sample consisting of repeated NRPR runs, with statistically independent initial conditions, provides the number of independent NRPR trials,
, needed to obtain two images that have a normalized correlation coefficient within
of unity.
Figure 8 shows the resulting histograms of the computed number of NRPR trials,
, for box sizes of 70, 60, 50, 40, and 30. These histograms, denoted by
, display for each
R the probability of obtaining the noise-free image for each repetition,
R, of the stochastic search, noting that each repetition ends with the second noise-free image. In practice, a reasonable starting value of box size would be roughly half the size of the field-of-view, corresponding in this case to
. As we reduce the box size down to the autocorrelation of the true foreground, the lowest histogram in the figure shows that all samples arrive at the noise-free image in just two iterations of NRPR. When we consider that a purely stochastic search would take on the order of
NRPR runs, it is clear that the stochastic search algorithm is highly efficient and that the data processing time for the combined stochastic search/NRPR algorithm is quite modest.
Another observation is that for
below 30 and greater than 16, the correct convergence to the noise-free image again requires only two iterations of NRPR. It is to be noted that as long as the box is large enough to contain the foreground, the imaged object will be located within the box. However, for 16 and below, the box is too small to contain the bright object, and therefore, NRPR cannot correctly converge to a pattern that is entirely distributed within the box. In fact, repeated trials produce completely different image patterns that remind one of a disordered fragmentation that is widely distributed beyond the confines of the box.
Figure 9 illustrates this behavior. The main conclusion is that the algorithm informs the user when the specified box size is too small.
5. Analysis and Refinement
Having explored the features of the stochastic search method via computational experimentation in
Section 4, we review the supporting theorems discussed in Reference [
8] to validate the previous results. This will also result in an even faster method of image identification. To illustrate and validate the analytical derivations presented in [
8], we adopt a more complex test image, shown in
Figure 10. This is a rendering of a GOES 16-type geophysical satellite at geostationary orbit with 20 cm resolution. For the present investigation, the values of the parameters in the above are fixed at
and it is customary practice to choose the field of view size,
, so that it is more than three times the size of the illuminated object.
The first step is to recast the Equations (15) and (16) into a discrete-time dynamic system such that all the principal quantities are indicated by the sequence of integers
. Consider the equations pertaining to
:
where the function
is a Kronecker vector having statistically independent elements that are uniformly distributed in
, as defined previously. Because
and
are very small, the residual terms are discarded for all
k. This leads to the discrete dynamic system:
Now, retaining the Kronecker vector notation,
is represented by the column vector:
In other words,
is a vector composed of the matrix elements of the background pixels, and
is the vector composed of the matrix elements of the foreground pixels. Corresponding to this is the following:
is the unitary matrix corresponding to the Fourier transform operator in the Kronecker notation. Finally,
is represented by
Reference [
8] commenced the computation of an asymptotic approximation of the background, introducing the averages of
as defined below:
To the first order, the maximum estimate for the limit on
can be found to be
The resulting asymptotic approximation is
The computations for the first 40 iterations are shown in
Figure 11 and
Figure 12, which give the complete 7000 iterations.
A comparison of the predicted and computed results in
Figure 11 and
Figure 12 establishes that the averaged background intensity is reduced by an order of magnitude at the outset and then levels off beyond iterate
to form a floor that does not vary appreciably beyond
and within several hundred iterations thereafter. These features match the computed results rather well, especially for the immediate drop-off for
, where it is established that
is very weakly dependent on
.
Figure 12 shows the complete averaged background from k = 1 to k = 7000. The conclusion of this derivation is that after
,
is so suppressed relative to
that its influence on the foreground may be neglected. Thus, we set
. In addition, for the first few iterates, the initial value of
is hardly affected so that we are left with
At this point, References [
8,
13,
14] executed a complex derivation to transform the above system into two quantities.
defined on the finite support set:
. The transformed equations are
where
is the correlation operator. Next, we have several proofs of the nonnegativity of
:
Theorem 1. is the real, nonnegative autocorrelation of .
Proof. Since is nonnegative, we have ; thus, the result follows Parseval’s theorem. □
Corollary 1. Since both are real and nonnegative, so is .
Proof. . □
Thus, , , and g are nonnegative and real. The following shows a necessary condition for convergence to the noise-free image.
Theorem 2. If the box contains the entirety of the region of support of , then (a) the correlation, and is maximum; (b) this is a necessary condition for g to converge to the noise-free image.
Proof. Earlier in the analysis (see Equation (33)), it follows that
converges to a maximum value. Given this,
is bijective to
and
. Then, the main conclusion follows from the work of Klibanov [
15,
16]. □
Theorem 3. If the box contains the entirety of the region of support of the noise-free image, , and no less, then this is a necessary and sufficient condition for the convergence to the noise-free image.
Proof. The region of support of
encompasses that of
, implying the loss of constraints on the elements of
. Since a maximum cannot be reduced via the elimination of constraints, the stated condition remains necessary. Then, in view of the analysis in [
8], the stated condition is necessary and sufficient. □
Comment. A multitude of computations indicate that if the regions of support of overlap to the extent of 50% of area, this is almost invariable that g converges to the noise-free image. Thus, with the probability of nearly one, we may relax Theorems 2 and 3.
The following will be of use:
Theorem 4. If the box is too small to completely contain the noise-free image, then NRPR cannot converge to the noise-free image. Furthermore, the converged NRPR run exhibits illuminated pixels outside of the box.
Proof. The fundamental operation of NRPR is that it simultaneously attempts to satisfy the constraints imposed by the coherence magnitude data and the image domain background constraints. These are in direct conflict so that correct convergence is impossible. □
Now, in
Figure 13, we illustrate the fact that
and
occupy statistically independent positions within the foreground box, and the brighter portions of
and
encompass the region of support of the true image (or the region of support for the possible convolutions of the true image). One notices from
Figure 13 that there is almost no overlap between
and
. In conformity with Theorems 2 and 3 and the comment above,
Figure 13 shows the convergence to a poor image, where it is apparent that the magnetometer boom is on the wrong side of the satellite. These results show that with high correlation and good image quality, there is significant overlap between
and
.
The foregoing observations suggest a rather simple proof of the stochastic algorithm for high correlation and convergence to the noise-free image. This is summarized in
Figure 14 as follows:
We can now construct a complete model for the probability of the zero-noise image convergence as a function of the repetitions,
R, of the stochastic search algorithm for a given box size. The first step is a model for the probability for a correct convergence on any single trial as a function of the box size. This function, denoted here by
, involves both
and
. For the situation on the right side of
Figure 14, since the relative positions within the box are completely random, the probability of an overlap between
and
can be approximated simply as the ratio of the areas
and
, i.e.,
. Note that this satisfies Theorem 3 when the situation on the right side of
Figure 14 holds. Finally, when
, Theorem 4 is satisfied. In summary, we derive the following:
Figure 15 shows the above formula rendered as the blue line. The red stars represent the percentage of correct images out of one hundred trials as calculated by NRPR. The agreement is close. Also, there are a few correct images even for the case wherein the box size coincides with the whole field of view (in other words, the number of constraints are equal to zero).
It must be emphasized that this expression is the probability of NRPR attaining the correct image in a single trial, but a successful determination of the true image via the stochastic search method requires the occurrence of two identical and correct images within
R independent trials, where the last trial produces the second correct image in the sequence of
R trials. Let us replace
in
Section 4 with
, the probability that the correct image is discovered in
n trials given that
is the probability of obtaining the correct image in one trial. The following theorem displays the required result:
Theorem 5. The probability of obtaining the correct image via the occurrence of two correlated images in n independent trials is Proof. In the stochastic search algorithm, the
trial is defined as the second successful correct image, thus
The above probability has the form of a Bernoulli distribution since each trial is independent. Therefore, we derive the following:
□
Figure 16 shows the probability distributions pertaining to the test image for a range of box sizes. These results verify that the probabilities of identifying the correct image via the stochastic search process are quite high.
Figure 16 is also in line with the computational results shown in [
8]. Furthermore, the progression of distributions from large to small box sizes suggests the following process. Having made a rough estimate of the object size, one starts with a box size that is twice the estimate. As seen in the third from the top distribution in
Figure 16, 90% of the probability mass is contained in the first 20 trials, and the maximum probability is only 5 trials. Thus, one would set the maximum set of trials at approximately 10. In the event of no success, one should reduce the box size by approximately a quarter until two identical images occur. If one reduces the box size to less than the object size, then by Theorem 4, NRPR will not converge properly. According to [
8], such an event can be recognized by the fact that the distribution of intensity in the image plane is several times the box size and the object size, in which case one must enlarge the box size slightly and continue to completion. In the next section, a refinement to the above process is introduced that exploits the reverse procedure of starting with a too small box size and proceeds to enlarge the box size.