1. Introduction
The Coronavirus Disease 2019 (COVID-19) pandemic forced lockdowns all over the globe, and claimed more than seven million lives worldwide [
1]. In order to handle and contain pandemics, and particularly COVID-19, large portions of the population should be frequently tested [
2]. One of the main difficulties in doing so stems from the limited testing resources and the lengthy duration required to identify the presence of an infection [
3,
4]. In particular, the main diagnosis tool for COVID-19 tests is based on Ribonucleic acid (RNA) extraction via qualitative reverse transcription polymerase chain reaction (RT-qPCR). Although various technological alternatives have been proposed [
5,
6], such as the rapid lateral-flow antigen tests, and rapid lateral-flow tests based on Clustered Regularly Interspaced Short Palindromic Repeats (CRISPR) gene-editing technology [
7] and the Reverse Transcription Loop-Mediated Isothermal Amplification (RT-LAMP) [
8], the qualitative reverse transcription polymerase chain reaction (RT-qPCR) process remains the most accurate method for COVID-19 testing. The output of this procedure represents an estimate of the viral load in the tested sample [
9]. This estimation indicates the infection process and progress (when it exists and starts), e.g., for Syndrome Coronavirus 2 (SARS-CoV-2) in the cell and between cells [
10]. The more efficient the virus is at entering and replicating in cells, the faster the viral load builds up. In practice, viral load analysis (when it is detected) is a complicated phenomenon that is widely considered in the medical and biological community, and there are certain factors impacting it, such as the presence/absence of specific Angiotensin-Converting Enzyme 2 (ACE2) mutations in an individual or the specific mutations accumulated in the S-protein of SARS-CoV-2. These factors are crucial to learn and conclude on the viral infection. In this work, we focus on efficient viral load detection from tested samples. We refer to the literature for more information on the viral load analysis [
11,
12]. Now, the main bottleneck associated with this form of COVID-19 testing follows from the fact that each qualitative reverse transcription polymerase chain reaction (RT-qPCR) machine can simultaneously process a fixed number of samples, and its procedure tends to be quite lengthy, on the order of a few hours for each test [
13,
14]. Computational and algorithmic methods have shown great promise in alleviating these challenges, especially during the time of limited testing resources [
15].
A promising method to tackle this lengthy measurement procedure and thus to increase the efficiency of COVID-19 tests is based on pooling [
16,
17]. Here, each sample processed, i.e., each input to the qualitative reverse transcription polymerase chain reaction (RT-qPCR) machine, is comprised of a mixture of several samples taken from different patients. When the infected patients constitute a small subset of the overall tested individuals, pooling-based schemes allow accurate recovery using a number of tests which is notably smaller than the number of patients [
18]. Several recovery schemes for pooled COVID-19 tests have been recently proposed [
17,
19,
20,
21,
22,
23,
24,
25], which can be divided according to the two main mathematical frameworks for such recovery procedures: The first is group-testing theory, originally derived for detecting infections in large populations [
26], used in [
16,
17,
19]. Group testing (GT) was first suggested to identify syphilis-infected draftees during World War II [
26], and has been long studied and utilized in many fields, such as computational biology [
27], biology and chemistry [
28,
29], communications [
30,
31,
32], sensor networks [
33], pattern matching [
34], web services [
35], and cyber security [
36,
37,
38]. The second framework is compressed sensing, which deals with the recovery of sparse signals [
39], and was utilized for pooled COVID-19 tests in [
20,
21,
22,
23,
24,
25]. Another interesting recovery method, which can be seen as an extension of classical group testing (GT), is the Semi-Quantitative Group Testing (SQGT) scheme [
40]. This method assumes that varying material can be extracted from different items, resulting in discrete non-binary pooling matrices. This scheme can be useful for conflict resolution in multiple access channels and genotyping [
40].
One of the main differences between classical group testing (GT) and compressed sensing (CS) is that GT deals with group detection problems, which results in binary variables. Specifically, in GT, each subject can either be infected or not infected [
18], while compressed sensing (CS)-based methods result in real-valued variables. GT is traditionally adaptive, requiring multiple sequential tests [
26] in order to achieve a minimal number of overall tests from which the presence of infection can be inferred. Nonetheless, GT can also be applied in a one-shot (non-adaptive) manner [
41], avoiding the need to mix new samples during the testing procedure. CS focuses on the one-shot recovery of sparse real-valued vectors from lower-dimensional linear projections, and thus each subject can take any real value number [
39]. The additional domain knowledge of GT, namely, the more restricted binary domain over which it operates compared to CS, allows it in some applications to operate using fewer measurements compared to CS, as shown in [
42,
43] in the context of quantization of sparse signals.
When testing for contagious diseases, and particularly for COVID-19, one is often interested in obtaining some score on the level of the viral load of the patients due to its epidemiological value [
21,
44,
45]. This score is called the infection level. The infection level can correspond to an estimate of the viral load, which in turn can map to the threshold cycle value of RT-qPCR testing. For example, [
46] uses viral loads to determine the effectiveness of different quarantine methods. This can be achieved using CS tools. The fact that GT and CS have their own pros and cons for pooled testing motivates the design of a recovery method which combines GT with one-shot operation and multi-level detection, as in CS.
Neither GT nor CS fully address the challenges posed by the need to identify infection levels during an ongoing pandemic of a large scale. In this work, we propose a multi-level GT recovery scheme for pooled testing. Our proposed GT-based method is designed to account for the unique characteristics of pooled tests for contagious diseases, and particularly those arising in COVID-19 testing. The proposed technique extends GT schemes to detect multiple levels of viral load, building upon our previous results on combining GT with CS concepts and multi-level discretization in [
42]. The resulting multi-level GT scheme operates in a one-shot manner, and is designed to avoid dilution due to mixing too many samples in each pool [
4]. This is a novel approach that identifies the non-binary infection levels, as opposed to classical GT, while requiring significantly less pooled tests than CS-based approaches. In fact, we demonstrate that our method achieves the upper bound on the number of tests of a maximum likelihood (ML) decoder, while having significantly lower complexity than a maximum likelihood (ML) decoder. All of these characteristics, alongside the proven low complexity regime in which our method works in the context of COVID-19, makes it particularly attractive for large-scale infection level detection among large populations. A comparison between the methods is available in
Table 1. Note that our proposed multi-level GT method can be readily altered to return viral loads instead of infection levels, by removing the quantization at the end.
We begin by identifying the specific requirements which arise from the setup of pooled COVID-19 testing. From these requirements, we derive the multi-level GT method. Our scheme is comprised of a dedicated testing matrix, determining which patients are pooled together into each test; and a GT-based recovery method operating in a two-stage manner, by first identifying the defective patients building upon classic GT tools, followed by a dedicated mechanism for characterizing the level of infection for the identified patients.
We theoretically analyze the proposed multi-level GT scheme. We first characterize its associated computational complexity, which is shown to be dominated by the identification of the defective subjects in the first stage of the proposed algorithm. As the complexity formulation results in a random quantity which depends on the statistical modeling of the measurement procedure, we derive the expected computational burden, and show that it results in a number of computations which is of the same order as that of low-complexity CS-based methods. Next, we derive sufficient conditions for the algorithm to yield a unique solution. While similar guarantees are also available for CS-based methods, we numerically demonstrate that our proposed scheme achieves improved accuracy over CS-based pooled recovery. While GT has been shown to be effective in real settings, and for COVID-19 in particular [
47,
48,
49], obtaining real-world data can be challenging [
50]. We therefore conduct our experimental results using the model proposed in [
21] for pooled RT-qPCR testing. For these setups, we demonstrate that our multi-level GT scheme reliably recovers the infection levels, while operating at a limited computational burden, and achieves improved accuracy over existing CS-based approaches.
The rest of this paper is organized as follows:
Section 2 reviews the system model, focusing on pooled COVID-19 testing and identifies the unique requirements of this procedure.
Section 3 presents the proposed multi-level GT scheme. In
Section 4, we analyze our approach, identifying sufficient conditions for it to reliably detect the level of infection of each patient, and characterize its complexity.
Section 5 details the simulation study, and
Section 6 provides concluding remarks.
Throughout the paper, we use boldface lower-case letters for vectors, e.g., . Matrices are denoted with boldface upper-case letters, e.g., . Let denote the element at the i-th row and j-th column of . Sets are expressed with calligraphic letters, e.g., , and is the n-th order Cartesian power of . The stochastic expectation is denoted by , ⋁ and ⊕ are the Boolean OR and XOR operations, respectively, and is the set of non-negative real numbers.
3. Multi-Level Group Testing
In this section, we provide an efficient scheme which implements GT with multiple decisions. Our design builds upon the fact that the sparsity Assumption 1 implies that the recovery of pooled RT-qPCR tests can be treated as a sparse recovery problem, which is typically studied under either the framework of GT [
18], or that of CS [
39]. Broadly speaking, GT deals with sparse recovery of binary variables, i.e., it can recover whether a subject is infected or not. In order to evaluate the actual levels of each tested subject, as requested in Assumption 2, one would have to re-run the RT-qPCR test, violating requirement Assumption 4. The alternative approach of CS operates over the domain of real numbers, namely, it attempts to identify the exact cycle threshold or viral load for each subject, and thus tends to be less accurate compared to GT, as it does not exploit the fact that one is only interested in a discrete grade value by Assumption 2. This motivates the derivation of a dedicated algorithm for pooled COVID-19 recovery, which harnesses the strength of GT theory while extending it to the multi-level domain. The proposed multi-level GT method is presented in
Section 3.1, followed by a discussion in
Section 3.2.
3.1. Pooled-Testing Algorithm
Multi-level GT is comprised of two components: the design of the testing matrix , which determines the pooling operation; and the recovery algorithm which determines the discrete level associated with each subject based on the results of the pooled tests. We next elaborate on each of these components.
3.1.1. Pooling Procedure
To determine
, we first fix the number of the pool-tests
m. In principle,
m should satisfy
(where the
is often omitted for ease of read), for some
, as this is the sufficient number of pool-test for reliable recovery in GT using the optimal maximum likelihood (ML) decoder [
55,
56]. The parameter
controls the probability of error of the procedure [
55], as we illustrate in
Section 5. In practice, the number of pools-tests is often dictated by the measurement setup, e.g., it may be constrained to be an integer multiple of the number of inputs accepted by an RT-qPCR machine. Unless stated otherwise, we assume that
, and in
Section 4, we prove that this number of samples is sufficient to detect the infection levels using the proposed algorithm.
Once
m is fixed, we proceed to setting
, whose binary entries dictate which patient is mixed into which pool. The traditional GT method of generating
draws its elements in an i.i.d. fashion according to a Bernoulli distribution with parameter
p. A common choice for
p is
, for which the probability of each element in
to be zero is
. We note that an approximation of
is often used [
57]. This approximation yields an error term of the order of
; hence, it is negligible for sufficiently large values of
k. When
k is unknown,
p is chosen using a rough approximation of
k. A major drawback of this approach is that Assumption 5 is not necessarily satisfied, and there is some chance that too many patients will be mixed into the same pool, causing dilution. We therefore consider an alternative method, which forces the columns of
, as well as the rows of
, to be “typical”. That is, we want every column/row to have exactly
and
ones, respectively. This ensures that with high probability, half of the elements of
are zero, which, in turn, reduces the required pooled tests, as demonstrated in
Section 5. This requirement cannot be guaranteed in the non-asymptotic regime by generating the elements of
in an i.i.d. fashion; hence, we force
to be typical.
Since in practical testing setups, one is interested in using a fixed deterministic matrix rather than having to work with random matrices, we generate once before the pooling starts. A typical matrix is not unique, and in fact, there are many typical matrices. Generating a typical matrix that satisfies Assumption 5 and the number of required pool-tests for recovery can be performed readily. The same testing matrix can be used for multiple pooling experiments.
3.1.2. Recovery Algorithm
The proposed recovery algorithm is given in Algorithm 1. It operates in two main steps: The first step treats the recovery as a GT setup, and thus aims to divide the patients into infected and non-infected. The second step accounts for the multi-level requirement Assumption 2, and uses decisions made in the first step to determine the infection levels.
Step 1: Detection of Defective Subjects. The first part of the algorithm identifies efficiently all of the definitely defective items in two stages, without determining the infection levels. It does so by treating the problem, in which the observations and the viral loads take continuous non-negative values in general, as a GT setup which operates over binary variables. Hence, in this step, we use a binary representation of the real-valued
, denoted
, where
divides each measured pool into infected and non-infected. Recalling that the viral level decisions are defined by the thresholds
, the
i-th element of
, denoted
, is given by
Algorithm 1 Multi-level GT recovery. |
Input: |
Output: | ▹ contains index and infection level tuples |
1: |
Step 1: Detection of defective subjects |
2: | ▹ contains PD items |
3: | ▹ contains DD items |
4: |
Step 2: Infection level recovery |
Option 1: Least squares method |
5: Solve least squares (LS) problem (3) |
6: for all i s.t. do |
7: |
8: end for |
Option 2: Iterative method |
9: |
10: while do |
11: for all i s.t. do |
12: index s.t. |
13: |
14: |
15: | ▹ From line 12 |
16: |
17: end for |
18: end while |
Definitely Not Defective (DND) [41] |
19: procedure DND() |
20: |
21: for all i s.t. do |
22: for all j s.t. do |
23: |
24: end for |
25: end for |
26: return |
27: end procedure |
Maximum Likelihood (ML) [55] over Possibly Defective (PD) subjects |
28: procedure ML() |
29: return |
30: end procedure |
In the first stage of the first step, the definitely not defective (DND) algorithm [
58] is used (lines 2 and 19–27). This stage removes all items which are definitely not defective, resulting only in a smaller possibly defective set
. Recall that the number of pool-tests
m, dictated by the testing matrix
, is fixed to allow maximum likelihood (ML) detection. The definitely not defective (DND) algorithm attempts to match the columns of
with the vector
. In particular, if column
j of
has a non-zero entry while the corresponding element in
is zero, the column is assumed not to correspond to a defective subject. This algorithm finds most of the subjects that are definitely not defective (DND) and drastically reduces the number of possible defective (PD) items. The set of subjects declared as possible defective (PD) after this stage, denoted by
, is shown to be indeed much smaller than the number of patients
n; this is proved rigorously in
Section 4, and demonstrated empirically in our numerical study in
Section 5. In particular, in
Section 4, we give a precise expression for the expected size of
. We show that
in the non-asymptotic regime that is of interest for COVID-19 pooled testing, and numerically assert these claims, showing that, typically,
. The remaining
subjects are declared not defective.
The fact that the number of possible defective (PD) subjects is notably smaller than the overall number of patients is exploited in the second stage of the first step, which determines the set of definitely defective (DD) patients. Here, the Boolean ML algorithm [
55] is used only over the smaller set of PD subjects
, to identify exactly the set
of
k definitely defective (DD) subjects (lines 3 and 28–30). We use ML over the subset
rather than over all the patients, allowing us to carry out this typically computationally prohibitive method at affordable complexity, as we show in
Section 4.5. The ML algorithm looks for a collection of
k columns in
, for which
is most likely. The ML decision rule is given by
In the ML rule, we denote by
the set of defective subjects, and by
the set of
combinations of
k defective subjects in
. While the formulation here is given with the true number of infected patients
k, in practice, it is carried out with
k (which is unknown) replaced by an approximation or an available upper bound on it.
Step 2: Infection Level Recovery. The output of the first step is a division of the patients into infected and non-infected, encapsulated in the set of identified defective subjects . In the second step, the algorithm estimates the infection level of each subject in . We provide two methods to do so:
Option 1—least squares: The first option uses least squares (LS) estimation [
59] to identify the viral loads of all definitely defective (DD) items (line 5):
where
denotes the matrix created by taking the columns of
from
. The output of the algorithm is the infected items, and the quantized infection levels
(line 10) using the threshold-based quantization mapping
defined in
Section 2.2, where a value of “not infected” is given to all items not in
. For the LS solution in (
3) to be unique, we should have
. In this case, (
3) is given by
When the measurement process induces some known non-linearity, it can be incorporated into (
3), resulting in a non-linear LS formulation. The LS operation outputs a real-valued vector, while the discrete levels are obtained by quantizing the entries of (
4). Note that
has to be non-negative; thus, one can compute the LS in (
3) while constraining the entries of
to be non-negative. Here, we omit this constraint; the simulations conducted in
Section 5 show that omitting this constraint does not affect the performance of the algorithm in the case studied in this paper. However, in the general setting, this constraint can be easily included if needed [
60]. We choose this method in our analyses below, namely, the complexity analysis and the theoretical performance guarantees characterized in the sequel consider Algorithm 1 implemented with the LS recovery option. Nonetheless, we also provide an additional method to recover the infection levels based on iterative detection.
Option 2—iterative detection: An alternative approach is to iteratively estimate the viral loads using pools containing a single patient whose viral load was not yet recovered, and to update the measurements accordingly. In contrast to option 1, which jointly maps the entries of into an estimate of the viral loads, here, we iteratively search for the pools which represent a single unknown value, identifying their corresponding infection levels separately, and then canceling their contribution on the remaining pools. This approach facilitates recovery in setups where not only the infection levels, but also the viral loads themselves are discrete.
The iterative method for recovering the infection levels from is summarized in lines 9–18 of Algorithm 1. Here, we let be the i-th element of . For a test in which only one infected subject participates according to the testing matrix (lines 11–12) (note that ∑ here represents addition over integers and not Boolean OR), the algorithm can recover the viral load directly from the measurement (lines 13–14). To obtain a discrete score, the measured value is quantized using a threshold-based mapping . Then the algorithm subtracts the viral load of that subject from all the tests in which it participates (lines 15–16), and repeats until it recovers the infection levels of all declared infected subjects (line 10).
In
Section 5, we compare the performance of the algorithm with the LS method vis-à-vis the iterative method. In our numerical results, which focus on a scenario with continuous-valued viral loads, the LS method is shown to achieve more accurate recovery. This follows from its ability to mitigate the effects of the noise induced in the measurement process, allowing it to retrieve viral loads that are close enough to the true viral loads. Option 2, in which detected pools are iteratively subtracted from non-detected ones, is more sensitive to noise and inaccurate identifications made in previous iterations, whose error may propagate to subsequent decisions. Note that increasing the granularity of the quantization can move us from discrete scores to continuous scores. We also would like to note that in the discussed regime of COVID-19, where there are strict limitations on how many tests can be combined (see
Section 2), both options are feasible in terms of computational complexity and running time. Finally, viral load analysis is a complicated phenomenon whose study falls beyond the scope of this paper, as discussed in
Section 1 and
Section 6.
3.2. Discussion
The novelty of the multi-level GT algorithm stems from its two-step procedure; the first step utilizes efficient GT methods to notably reduce the number of examined patients, while the second step inspects the remaining patients and applies dedicated mechanisms to recover their level of infection. Here, we note a few remarks arising from the proposed scheme.
In
Section 3.1, we describe how the matrix
is generated. The description involves random generation, for which the resulting matrix is not guaranteed to satisfy Assumption 5. The motivation for using such randomization stems from its provable performance guarantees in GT theory [
18]. In practice, once a typical testing matrix satisfying Assumption 5 is selected, one does not have to generate a new matrix for each group of
n patients.
For
n i.i.d. tested individuals, the probability of finding the infected items (though not necessarily their levels) is maximized by the ML rule. However, its complexity is burdensome, as it has to consider
options [
61,
62]. An efficient alternative is the DND algorithm, also known as column matching [
41,
58,
63], whose time complexity is
[
56]. However, it requires a greater amount of pooled measurements compared to ML in order to reliably identify the detective items. The first step of our proposed multi-level GT method combines the concepts of DND with the ML algorithm, while the second step extends them to operate over non-binary fields, i.e., recover multiple levels rather than just identifying which subject is defective. Performing DND on all
n items using the number of tests set to allow ML detection, i.e.,
as opposed to the number of tests required for DND, which is
[
41], results in a smaller set of PD subjects
. This, in turn, notably reduces the complexity of the recovery method. In
Section 4, we give exact expressions for
after performing DND. Given
, the ML decoder has to consider significantly less options,
which is likely to be computationally feasible and considerably faster than considering all (
) combinations, as discussed in the complexity analysis in
Section 4.5. It is important to note that (
) options are considered only during the second stage of the first step (the ML stage). In the second step, the infection levels are produced only once (using either the LS option or the iterative detection option), after the most likely set of infected items have been generated.
Algorithm 1 requires two conditions to hold. First, the number of PD items, i.e., , should be relatively small, preferably close to k, so that the ML algorithm is feasible. Furthermore, for the LS recovery method to have a unique solution, it should hold that . If the solution is not unique, the algorithm may produce an undesirable output of different decisions when applied to the same pooled measurements. Both these requirements are numerically asserted to hold in the experimental scenarios considered in the paper, and their validity is also theoretically studied in the following section.
We note that in this paper, we assume that the binary quantization of the tests
does not have any error. If this does not hold, one can use the noisy variants of the DND and the ML algorithms of Step 1, as known in the GT literature [
41,
55]. We leave this case for future work.
In our numerical evaluations of multi-level GT, we demonstrate its ability to achieve error-free detection when the number of measurements
m satisfies the
upper bound on this quantity for which the computationally prohibitive ML GT is guaranteed to approach zero error, i.e.,
for some
[
55]. Furthermore, in the case where it is possible to identify the
number of infected subjects in a pool-test by observing its measurement, i.e., when one can separately tell for each element of
how many defective patients are mixed into it (though not necessarily their infection level), we notice empirically that it is possible to achieve the
lower bound number of pool-tests suggested for GT using the ML algorithm, i.e.,
[
41,
55] (for further details, see
Section 5). We leave the analysis of this case for future work.
3.3. Unknown k
It is commonly assumed in the GT literature that the number of infected items
k is known in advance [
62], or that it can be estimated correctly within
tests [
53,
54]. We mainly study this scenario in this paper to focus on the novel aspects of our proposed method. However, in practical scenarios, let alone during a pandemic, this assumption can be restrictive, and the estimations may suffer from errors, especially in the pooled tests that are noisy. In this section, we assume that the estimated number of infected items
is different than the actual number of infected items
k. We propose an adaptation to the proposed method that can handle this mismatch. Recall that our method is comprised of two main steps: designing the pooling procedure and executing the recovery algorithm.
Designing the pooling procedure boils down to designing the pooling matrix . We propose to use the estimate when constructing the pooling matrix, i.e., . We break the analysis for the two possible mismatches.
: In this case, the DND step will result in fewer PD items than if we designed the pooling matrix with the correct estimation for
k. In the non-asymptotic regime discussed in this paper, such mismatch is likely to result in the DND step correctly identifying all the infected items, as the number of tests required to correctly identify all infected items using the DND algorithm is
[
41,
58].
: In this case, the DND step is likely to generate more PD items since this estimation is equivalent to lowering
in the number of tests
m. Assuming the error in the estimation is not severe, the modified ML step can identify the correct infected items with manageable complexity, as we next discuss. For a numerical example, see the discussion in
Section 4.1.
It is worth mentioning that once the pooling matrix is set, the DND step does not require knowledge of the number of infected items. Hence, the DND step remains unchanged.
The ML step in Algorithm 1 assumes that the number of infected items is known in advance. If k is not known in advance, one can alter the ML procedure of Algorithm 1 (starting at Line 28) to look for any set of defective items. The objective is to find the smallest number of infected items that would explain the output results . This changes the ML procedure to the one given in Algorithm 2. The adapted version of the algorithm is still able to find the infected items correctly with high probability. This is also supported by our simulations.
We do note that in the case where the number of infected items
k is known in advance, declaring that
k items are infected is equivalent to declaring that the rest
items are not infected. However, in the case where
k is unknown, the ML step identifies the smallest set of items that are declared infected
among the
PD items. In this case, we cannot say for certain that the rest of
are not infected. We also empirically show that the infected items are identified correctly with high probability, and there may be some false positives among the non-infected items. Recall that in the context of COVID-19, falsely identifying non-infected items as infected is not as severe an error as false negatives, where infected items are declared as not infected.
Algorithm 2 Maximum Likelihood (ML) over PD subjects , unknown k |
procedure ML() for k=0, 1, … do if explains then return end if end for end procedure
|
4. Theoretical Analysis
In this section, we provide a theoretical analysis of the proposed multi-level GT scheme. In particular, we characterize two key components of Algorithm 1: the expected number of PD subjects, which dominates the computational complexity of the algorithm, and guarantees for LS recovery to have a unique solution. We then proceed to analyze the complexity of Algorithm 1, and show it is computationally efficient in the regimes of interest in the context of COVID-19. Our theoretical results are given as a function of the system parameters, i.e., the overall number of patients
n and the number of infected ones
k (assumed in this section to be the true number of infected patients and not a bound on it), as well as using those of the pooling pattern, namely, the number of measurements
m and the Bernoulli distribution parameter
p. While our theoretical measures are given using the true number of infected patients, we recall that Algorithm 1 can be applied using an approximation or an upper bound on this quantity. For the analysis in this section, we assume that
is drawn according to an i.i.d. Bernoulli distribution with parameter
p. It is known in the literature that there is approximately a 10% performance gap between picking a typical
and an i.i.d. one, so this can serve as a bound on the performance [
64]. We also empirically demonstrate this result in
Section 5.2.
In our analysis of the expected number of PD subjects, we consider different models of the noise in Step 1. Following conventional GT terminology [
55], the cases we treat are as follows:
Noiseless case—here, the noise induced by the measurement process does not affect the ability to identify whether or not a pool contains a defected subject. Namely, .
Additive noise—here, the presence of noise may result in a pool not containing any defected subject being measured as an infected pool. This implies that for each .
Dilution noise—dilution implies that pools containing infected subjects may be falsely measured as non-defective. In this case, for each .
These cases are studied individually in
Section 4.1,
Section 4.2 and
Section 4.3, respectively, for arbitrary settings of the parameters
; conditions for uniqueness of the LS solution are stated in
Section 4.4, and the complexity of Algorithm 1 is characterized in
Section 4.5.
4.1. Noiseless Case
We first calculate how many items are declared PD after running DND in Algorithm 1 in the noiseless setting. Recall that in this stage, we use the DND algorithm to declare the PD subjects. In the noiseless setting, all defective k items are declared PD, and there is some probability that each non-defective item is declared PD by the DND algorithm. For Algorithm 1 to run in reasonable time, it is essential that is small enough, so the number of options to be considered by the ML decoder is not exponentially large. In the noiseless setting, the expected number of PD subjects is stated in the following theorem:
Theorem 1. The expected number of items declared PD by the DND algorithm (first stage of the first step of Algorithm 1) in the noiseless setting is given by Theorem 1 characterizes the exact expected number of PD items declared by DND in the noiseless setting. This characterization holds regardless of the fact that the outputs of the DND method are later used for identifying infection levels. Note that the second part of (5) is similar to a known result on the DND algorithm, e.g., [
65] (Equation (8)). The main difference is that it is traditionally used to bound the probability of error of the algorithm using a union bound, whereas we use this expression to calculate the exact expected number of PD items.
To affirm and numerically evaluate Theorem 1, we depict in
Figure 2a the expected number of items declared PD, computed via (5), plotted as the dashed lines, as a function of the number of defective items
k. This theoretical quantity is compared to the empirical expected number of items declared PD, averaged over
Monte-Carlo simulations. The number of tests are chosen to match the upper ML bound
, as described in
Section 5. We observe in
Figure 2a that the simulation results agree with the theoretical values dictated by Theorem 1. The figure also shows that in the non-asymptotic regime, which is of interest for applications such as COVID-19 pooled testing, the expected number of PD items is significantly smaller than the number of patients
n, and in particular, often lies in
. This means that the second part of Step 1 of Algorithm 1, which uses an ML decoder, has to consider much less combinations compared to that needed without the prior DND subject identification. Similar observations are reported in
Figure 3a, which considers the same comparison as in
Figure 2a, only with a significantly larger number of items
n. The figure shows that the average number of PD items after DND is substantially smaller than
n, indicating that the gains of using DND in reducing the computational burden are more dominant as the number of patients grows. In
Figure 2b and
Figure 3b, we empirically show that, using the results of [
66], the probability that the number of PD items is significantly larger than the calculated expected value is small. Namely, we show that
is small for
, as we show in
Figure 2a and
Figure 3a, where
g is a selected threshold. This ensures that the complexity associated with carrying out the ML step is manageable. The complexity can be controlled and further reduced by increasing the number of pooled tests
m.
This paper studies GT-based methods to identify patients infected with COVID-19 using the RT-qPCR method. As such, the regime in mind is a non-asymptotic regime (see Assumption 5). However, in the GT literature, the analysis is often conducted in the asymptotic regime, where
are large and
. In this regime, the number of non-defective items declared PD by the DND step goes to zero for
(see
Appendix A.2 for further details). This demonstrates the trade off between the desired complexity of the ML step and the number of tests
m.
Figure 2a,b also show that it is feasible to perform the ML step in the case of a mismatch between the actual number of infected items and the estimated number of infected items (as elaborated in
Section 3.3). In
Figure 2a, the number of pooled tests
m is designed for the case where there are
infected items and
. We see that when the true number of infected items
k exceeds 5, the expected number of PD items
is small enough, and
Figure 2b shows it will not be too far from
with high probability. For example, when the actual number of infected items is
k = 7, the DND step returns
= 16 PD items on average. The number of combinations considered by the ML step, assuming all the infected items are correctly identified, is equal to
, which is a feasible number of combinations to consider by the ML step.
Figure 2b shows that the probability to exceed
= 16 items is bounded above by 0.05.
4.2. Additive Noise
In the additive noise model, a test may be positive even if no defective items participate in it [
55]. This case reflects the possibility of false positives in traditional binary GT. This happens with probability
q, i.i.d. between different tests. That is, we define a set of i.i.d. random variables
, such that
obeys a Bernoulli distribution with parameter
q. The binarized test results of the
i-th test are here modelled as a Boolean OR function of the test result without additive noise and
i.e.,
. It follows from [
55] that the DND algorithm can still guarantee accurate recovery of the PD subjects when the number of tests is increased by a factor of at least
with respect to the noiseless scenario. We expand the results of Theorem 1 in the following theorem:
Theorem 2. Let q be the probability that a test without infected items comes out positive. The expected number of items declared PD by DND algorithm (first stage of the first step of Algorithm 1) in the additive noise setting is given by Theorem 2 extends Theorem 1 to the case of additive noise. Note that when , i.e., the effect of the additive noise vanishes, we obtain the corresponding expected value derived in Theorem 1 for the noiseless case.
To exemplify the difference between the expected number of PD subjects for noiseless and noisy setups, we next evaluate (6) with the parameters used in
Figure 2a for the noiseless case. Specifically, we set
, and focus on
defective items. In the presence of additive noise with Bernoulli parameter
, we obtain that
, as opposed to
in the noiseless case. This demonstrates the importance of adjusting the number of tests in the additive noise scenario: when we use a number of tests that corresponds to the noiseless case, DND, on average, outputs that most of the items are PD.
4.3. Dilution Noise
In the dilution noise model, a test may come negative even if defective items participate it [
55]. This case reflects the possibility of false negatives in traditional binary GT, which are often more severe than false positives, especially in the context of COVID-19. In particular, if a defective item is declared DND in Step 1 of Algorithm 1, then it will not be detected as infected by the algorithm.
To model the presence of dilution noise, we let each defective item be diluted with probability
u, independently of other defective items in the test. In order to handle the presence of dilution noise, the DND algorithm has to be adjusted. In the modified DND algorithm, a patient is declared PD if it participates in at least
tests that come out positive [
41,
65]. As a result, the number of positive tests that an item participates in is distributed according to a binomial distribution with
m trials. The probability to participate in a test depends on whether the item is infected or not. We denote the probability that an infected/non-infected item participates in a positive result by
, respectively. Consequently, the number of positive tests an infected/non infected item participates in is distributed according to
,
, respectively. Here,
is the binomial distribution with
a trials and probability of success
b.
In the presence of dilution noise, we are no longer guaranteed that a defective item would always be declared as PD. We define as the set of all defective/non-defective items that are marked PD, respectively, i.e., while is an empty set. The extension of Theorem 1 to the presence of dilution noise is given in the following theorem:
Theorem 3. Let u denote the dilution probability of a defective item in each test, and τ be the threshold used by the modified DND algorithm. Then, the set subjects declared PD is comprised of the distinct sets which satisfy ,
and thus
.
Theorem 3 characterizes the overall number of PD items, as performed in Theorems 1 and 2 for the noiseless and additive noise setups, respectively. In addition to expressing , Theorem 3 also identifies the expected number of patients which are not defective but are identified as PD by the DND algorithm. Unlike the models used in deriving Theorems 1 and 2, under the dilution noise model, we are no longer guaranteed that a defective item is declared DD. Thus, there is a non-zero probability that some of the k infected patients are not assigned into by the DND algorithm, and thus .
4.4. Uniqueness of LS Recovery
Next, we identify sufficient conditions under which the LS solution is unique. Step 2 recovers the infection levels based on the output of Step 1, i.e., . Therefore, to provide a lower bound on the probability that this output enables recovery of infection levels via LS followed by discretization of the values, in the following, we focus on the case where all the infected items have been identified by the first step of the algorithm.
Once the set of infected items, whose indices are denoted by
, has been identified by the GT procedure in Step 1, the infection levels of those subjects are recovered by quantizing the LS solution to (
3). Notice that since we are left with
k subjects to resolve, it holds that
, and
. When solving (
3), we treat
as a real matrix rather than a matrix over the binary field. A unique solution exists to the LS problem in this scenario if
is of full rank, namely
. When this holds, the unique solution to the LS procedure is used to estimate viral loads. Our proposed Algorithm 1 recovers the infection levels by quantizing these estimates, which is numerically demonstrated to yield accurate identification of the infection levels in
Section 5. The following theorem bounds the probability that
is rank-deficient:
Theorem 4. Let the number of tests satisfy for some , and the Bernoulli parameter p satisfy . Then, the probability that the solution of (3) is not unique is bounded as Theorem 4 guarantees that Step 2 of Algorithm 1 uses a unique LS estimate with high probability, assuming that Step 1 of the algorithm has successfully identified the infected items. The resulting probability depends on the number of pools
m via the parameter
. In
Section 5, we assess in simulation that the high probability of
having a full rank in Step 2 results in identifying the correct infection levels.
4.5. Complexity Analysis
One of the key features of the proposed multi-level GT methods is its reduced complexity compared to ML-based GT. To fully analyze this aspect of Algorithm 1, we give the following proposition which characterizes its complexity, focusing on the implementation with LS recovery.
Proposition 1. Let be a design parameter. The overall complexity of Algorithm 1, when used with , is given by Proof. The complexity of DND is
[
56]. Once PD items are identified, the complexity of the ML algorithm is
, as we have to compute a Boolean OR of
k vectors of size
m, and compare it with
. The complexity of LS depends on the dimensions of
, but also on the specific implementation. Since
, LS involves
operations [
67]. We note that it is likely that the matrix
is sparse [
39], and it becomes more sparse as
p decreases. Efficient algorithms for solving LS are known in the literature when the matrix is sparse, e.g., LSQR [
68]. Substituting
as the number of tests, we obtain (
8), thus proving the proposition. □
The complexity expression in Proposition 1 includes three terms, as the algorithm has three main components: the DND algorithm, which identifies the subset of PD items
; the ML search in (
2) on
; and the recovery of the infection levels using LS. The computational complexity of the ML stage is dictated by the set
, which is a random quantity depending on the observations. As shown in the previous subsections, the expected cardinality of the PD set can be computed in closed form depending on the assumed noise model.
To analyze the dominant part of the algorithm in terms of complexity, we consider the values
that are used in
Section 5, which are representative values when testing for COVID-19 using pooled RT-qPCR measurements [
21]. In this regime, it holds that
, and so the third component (representing the LS complexity) is negligible. To quantify the complexity associated with ML, which is dictated by the random set
, we adopt here the noiseless model for the binary measurements. In this case, when the number of tests is set to
, which corresponds to the upper bound on the number of tests required for ML decoding, we have that
(see
Figure 2a). Under these settings and when replacing
with its expected value, the ML decoder has to consider
combinations. This overhead is much lower than that required in order to apply ML directly to the measurements, which involves a search over
different combinations. As
is greater than
n, the dominant complexity factor corresponds to that required by the ML decoder. This complexity term is not prohibitive in the considered regime, and is of the same order as low complexity CS-based recovery, such as the Orthogonal Matching Pursuit algorithm, which typically requires
operations (after plugging in the number of tests
m) [
69].