Abstract
Specimens are collected from N different sources. Each specimen has probability p of being contaminated (in the case of a disease, e.g., p is the prevalence rate), independently of the other specimens. Suppose we can apply group testing, namely take small portions from several specimens, mix them together, and test the mixture for contamination, so that if the test turns positive, then at least one of the samples in the mixture is contaminated. In this paper we give a detailed probabilistic analysis of a binary search scheme, we propose, for determining all contaminated specimens. More precisely, we study the number T(N) of tests required in order to find all the contaminated specimens, if this search scheme is applied. We derive recursive and, in some cases, explicit formulas for the expectation, the variance, and the characteristic function of T(N). Also, we determine the asymptotic behavior of the moments of T(N) as and from that we obtain the limiting distribution of T(N) (appropriately normalized), which turns out to be normal.
Keywords: Prevalence (rate), Adaptive group testing, Binary search scheme, Linear regime, Probabilistic testing, Average-case aspect ratio, Characteristic function, Moments, Limiting distribution, Normal distribution
Introduction
Consider N containers containing samples from N different sources (e.g., clinical specimens from N different patients, water samples from N different lakes, etc.). For each sample we assume that the probability of being contaminated (by, say, a virus, a toxic substance, etc.) is p and the probability that it is not contaminated is , independently of the other samples. All N samples must undergo a screening procedure (say a molecular or antibody test in the case of a viral contamination, a radiation measurement in the case of a radioactive contamination, etc.) in order to determine exactly which are the contaminated ones.
One obvious way to identify all contaminated specimens is the individual testing, namely to test the specimens one by one individually (of course, this approach requires N tests). In this work we analyze a group testing (or pool testing) approach, which can be called “binary search scheme” and requires a random number of tests. In particular, we will see that if p is not too big, typically if , then by following this scheme, the expected number of tests required in order to determine all contaminated samples can be made strictly less than N. There is one requirement, though, for implementing this scheme, namely that each sample can undergo many tests (or that the sample quantity contained in each container suffices for many tests).
Group testing procedures have broad applications in areas ranging from medicine and engineering to airport security control, and they have attracted the attention of researchers for over seventy-five years (see, e.g., Dorfman’s seminal paper (Dorfman 1943), as well as Hwang 1972; Sobel and Groll 1959; Ungar 1960, etc.). In fact, recently, group testing has received some special attention, partly due to the COVID-19 crisis (see, e.g., Aldridge 2019, 2020; Armendáriz et al. 2020; Golliery and Gossnerz 2020; Malinovsky 2019; Mallapaty 2020, and the references therein).
The contribution of the present work is a detailed quantitative analysis of a precise group testing procedure which is quite natural and easy to implement. The scheme goes as follows: First we take samples from each of the N containers, mix them together and test the mixture. If the mixture is not contaminated, we know that none of the samples is contaminated. If the mixture is contaminated, we split the containers in two groups, where the first group consists of the first containers and the second group consists of the remaining containers. Then we take samples from the containers of the first group, mix them together and test the mixture. If the mixture is not contaminated, then we know that none of the samples of the first group is contaminated. If the mixture is contaminated, we split the containers of the first group into two subgroups (in the same way we did with the N containers) and continue the procedure. We also apply the same procedure to the second group (consisting of containers).
The main quantity of interest in the present paper is the number T(N) of tests required in order to determine all contaminated samples, if the above binary search scheme is applied.
Following the terminology of Aldridge (2019) we can characterize the aforementioned procedure as an adaptive probabilistic group testing in the linear regime (“linear” since the average number of contaminated specimens is pN, hence grows linearly with N). “Adaptive” refers to the fact that the way we perform a stage of the testing procedure depends on the outcome of the previous stages. “Probabilistic" refers to the fact that each specimen has probability p of being contaminated, independently of the other specimens and, furthermore, that the required number of tests, T(N), is a random variable. Let us be a bit more specific about the latter. Suppose is a sequence of independent Bernoulli random variables of a probability space , with and , , so that is the event that the k-th specimen is contaminated. Then, for each N, the quantity T(N) of our binary search scheme is a specific function of , say,
1.1 |
In particular, is not hard to see that
1.2 |
i.e. if none of the contents of the N containers is contaminated, then , whereas if all N containers contain contaminated samples, then by easy induction on N we can see that (observe that if , then , while for we have ). Thus, T(N) can become bigger than N, while the deterministic way of checking of the samples one by one requires N tests (i.e. in the case of individual testing we would have ).
Let us mention that, for a given p, the optimal testing procedure, namely the procedure which minimizes the expectation of the number of tests required in order to determine (with zero-error) all contaminated specimens in an adaptive probabilistic group testing, remains unknown (Malinovsky 2019). However, Ungar (1960) has shown that if
1.3 |
then the optimal procedure is individual testing. In practice, though, p is usually quite small, much smaller than .
We do not claim that our proposed procedure is optimal. However, for small values of p the numerical evidence we have suggests (see Tables 1 and 2 in Sects. 2.1 and 3.1 respectively) that, in comparison with the optimal results obtained by Aldridge (2019), Aldridge (2020) via simulation, our binary scheme is near-optimal.
Table 1.
Values of for certain choices of and p
1 | 1 | 1 | 1 | 1 | 1 | |
.505 | .515 | .575 | .645 | .715 | .780 | |
.265 | .280 | .395 | .528 | .653 | .768 | |
.148 | .169 | .335 | .518 | .679 | .820 | |
.092 | .121 | .328 | .542 | .719 | .871 | |
.068 | .103 | .340 | .566 | .748 | .901 | |
.059 | .099 | .352 | .581 | .763 | .917 | |
.057 | .101 | .359 | .589 | .771 | .924 | |
.058 | .103 | .363 | .593 | .775 | .928 | |
.059 | .105 | .365 | .595 | .777 | .930 | |
.060 | .106 | .366 | .596 | .778 | .931 |
Table 2.
Values of for certain choices of and p
.345 | .357 | .447 | .554 | .654 | .749 | |
.217 | .234 | .362 | .510 | .645 | .768 | |
.186 | .205 | .348 | .510 | .656 | .787 | |
.163 | .184 | .337 | .510 | .663 | .800 | |
.110 | .136 | .325 | .526 | .696 | .843 | |
.061 | .099 | .345 | .571 | .751 | .902 | |
.057 | .099 | .354 | .582 | .762 | .912 | |
.057 | .101 | .358 | .585 | .765 | .917 | |
.058 | .104 | .362 | .589 | .769 | .920 | |
.059 | .105 | .363 | .591 | .771 | .922 | |
.060 | .106 | .365 | .594 | .776 | .929 |
Let us also mention that our “binary search scheme” is different from the “binary splitting” procedures of Sobel and Groll (1959) (see, also, Aldridge (2019) and Hwang (1972)) where at each stage a group is splitted in two subgroups whose sizes may not be equal or even nearly equal. This feature makes these binary splitting procedures very hard to analyze theoretically, while the advantage of our proposed procedure is that, due to its simplicity it allows us to calculate certain relevant quantities (e.g., the expectation and the variance ) explicitly. We even managed to determine the limiting distribution of T(N) (appropriately normalized), which turns out to be normal.
Finally, let us mention that in Armendáriz et al. (2020) the authors consider optimality in the case where the pool sizes are powers of 3 (except possibly for an extra factor of 4 for the first pool). In the case where and their average number of tests in order to determine the contaminated specimens is 20, thus their average-case aspect ratio is 0.2. Tables 1 and 2 below, indicate that our proposed strategy gives comparable, if not better, results.
In Sect. 2 we study the special case where . Here, the random variable is denoted by . We present explicit formulas for the expectation (Theorem 1) and the variance (Corollary 1) of . In addition, in formulas (2.12) and (2.42) we give the asymptotic behavior of and respectively, as . Then, we determine the leading asymptotics of all the moments of (Theorem 3) and from that we conclude (Corollary 2) that, under an appropriate normalization converges in distribution to a normal random variable. Finally, at the end of the section, in Corollary 3 and in formula (2.75) we give a couple of Law-of-Large-Numbers-type results for .
Section 3 studies the case of an arbitrary N. In Theorem 4, Theorem 5, and Corollary 5 respectively we derive recursive formulas for , , and , while Corollary 4 gives an explicit, albeit rather messy, formula for . We also demonstrate (see Remark 3) the nonexistence of the limit of , as , which is in contrast to the special case , where the limit exists. In Sect. 3.2 we show that the moments of converge to the moments of the standard normal variable Z as . An immediate consequence is (Corollary 7) that Y(N) converges to Z in distribution.
Here, let us point out that the determination of the limiting behavior of a random quantity is a fundamental issue in many probabilistic models. Thus, our result that converges in distribution to the standard normal variable Z, and that the convergence is good to the point that we also have convergence of all moments to the corresponding moments of Z, gives much more information about the testing procedure than the information we obtain from the behavior of . And since the number N can be in the order of or, even , and the explicit results can get very messy for such big values of N, the asymptotic behavior of T(N) enables us to obtain valuable information without much effort.
At the end of the paper we have included a brief appendix (Sect. 4) containing a lemma and two corollaries, which are used in the proofs of some of the results of Sects. 2 and 3.
The case
We first consider the case where the number of containers is a power of 2, namely
2.1 |
As we have said in the introduction, the first step is to test a pool containing samples from all containers. If this pool is not contaminated, then none of the contents of the N containers is contaminated and we are done. If the pool is contaminated, we make two subpools, one containing samples of the first containers and the other containing samples of the remaining containers. We continue by testing the first of those subpools. If it is contaminated we split it again into two subpools of samples each and keep going. We also repeat the same procedure for the second subpool of the samples.
One important detail here is that if the first subpool of the samples turns out not contaminated, then we are sure that the second subpool of the remaining samples must be contaminated, hence we can save one test and immediately proceed with the splitting of the second subpool into two subpools of samples each. Likewise, suppose that at some step of the procedure a subpool of samples is found contaminated. Then this subpool is splitted further into two other subpools, each containing samples. If the first of these subpools is found not contaminated, then we automatically know that the other is contaminated and, consequently, we can save one test.
Let be the number of tests required to find all contaminated samples by following the above procedure. Thus by (1.2) we have
2.2 |
(in the trivial case , i.e , we, of course, have ).
Example 1
Suppose that (hence ). In this case can take any value between 1 and 7, except for 2. For instance in the case where the samples of the first three containers are not contaminated, while the content of the fourth container is contaminated, we have . On the other hand, in the case where the samples of the first and the third container are contaminated, while the samples of the second and fourth container are not contaminated, we have .
The expectation of
Theorem 1
Let and as above. Then
2.3 |
(in the trivial case the sum is empty, i.e. 0), where, as it is mentioned in the introduction, q is the probability that a sample is not contaminated.
Proof
Assume and let be the event that none of the samples is contaminated. Then
and hence
2.4 |
where for typographical convenience we have set
2.5 |
In order to find a recursive formula for let us first consider the event that in the group of the containers none of the first contain contaminated samples. Clearly, and
2.6 |
Likewise, if is the event that in the group of the containers none of the last contains contaminated samples, then
2.7 |
Let us also notice that , hence .
Now, (i) given and we have that , where the notation signifies that the random variables X and Y have the same distribution. To better justify this distributional equality we observe that, in the case , formula (1.1) takes the form and . Thus, . Now, and . Hence, the given event means that only some of the variables may differ from 0. Consequently, given we have that .
(ii) Similarly, given and we have that .
Finally, (iii) given and we have that , where, is an independent copy of . Thus (given ) by conditioning on the events , and we get, in view of (2.5), (2.6), and (2.7),
or
2.8 |
By replacing n by in (2.4) we get
2.9 |
Therefore, by using (2.8) in (2.4) and (2.9) we can eliminate , and obtain
2.10 |
Finally, (2.3) follows easily from (2.10) and the fact that .
For instance, in the cases , , and formula (2.3) becomes
2.11 |
respectively. In the extreme case we know that , while in the other extreme case we know that , and these values agree with the ones given by (2.3).
Let us also notice that formula (2.3) implies that for any given we have
2.12 |
where
2.13 |
It is clear that is a power series about with radius of convergence equal to 1 (actually, it is well known that the unit circle is the natural boundary of this power series), which is strictly decreasing on [0, 1] with and . Furthermore, it is not hard to see that it satisfies the functional equation
2.14 |
Also, by using the inequality
and by estimating the integral as one can show that
2.15 |
Remark 1
By dividing both sides of (2.3) by and comparing with (2.13) we can see that as
2.16 |
Actually, the convergence is much stronger, since for every the m-th derivative of with respect to q converges to the m-th derivative of uniformly on compact subets of [0, 1).
The expectation and, consequently, the average-case aspect ratio are getting smaller as q approaches 1 (equivalently, as p approaches 0). As an illustration, by employing (2.3) we have created the following table:
(by comparing the above values with the graphs, given in Aldridge (2019) and Aldridge (2020), of the optimal average-case aspect ratio as a function of p, we can see that, if , our procedure is near-optimal).
If q is such that
2.17 |
then, by applying the aforementioned testing strategy, the number of tests required to determine all containers with contaminated content is, on the average, less than N, namely less than the number of tests required to check the containers one by one.
Let us set
2.18 |
It is clear from (2.3) that for the quantity is a polynomial in q of degree , strictly decreasing on [0, 1], with and . Thus, there is a unique such that
2.19 |
and (2.17) holds if or, equivalently, if . Therefore, it only makes sense to consider the procedure if . Alternatively, if q is given, one may try to find the optimal value, say m, of n which minimizes and, then, split N into subgroups having size m.
As an example, let us give the (approximate) values of for :
where is a convention. Notice that , where is as in (1.3). Thus, is the threshold found by Ungar (1960) below which group testing does not improve on individual testing.
In view of (2.3) and (2.18), formula (2.19) becomes
2.20 |
which by letting yields
2.21 |
being the function defined in (2.13). In fact, . Furthermore, by formula (2.20) we have
which implies that the sequence is strictly increasing. Therefore, if , equivalently if , then for any .
The behavior of as
We start with a recursive formula for the generating function of .
Theorem 2
Let
2.22 |
Then
2.23 |
(clearly, ).
Proof
With as in the proof of Theorem 1 we have
thus
2.24 |
where
2.25 |
Let and be the events introduced in the proof of Theorem 1. Then, given and we have that ; given and we have that ; finally given and we have that , where is an independent copy of . Thus (given ) by conditioning on the events , and we get, in view of (2.25), (2.6), and (2.7),
2.26 |
By replacing n by in (2.24) we get
2.27 |
and (2.23) follows by eliminating and from (2.24), (2.26), and (2.27).
By setting in (2.23) it follows that the characteristic function
2.28 |
of satisfies the recursion
2.29 |
with, of course, .
For instance, for formula (2.29) yields
2.30 |
Corollary 1
Let
2.31 |
Then
2.32 |
(in the trivial case all the above sums are empty, i.e. 0).
Proof
Differentiating (2.23) twice with respect to z and then setting yields
2.33 |
where, in view of (2.22),
2.34 |
Thus,
2.35 |
and by using (2.10) and (2.35) in (2.33) we obtain
2.36 |
Finally, formula (2.36) together with the fact that imply
2.37 |
from which (2.32) follows by invoking (2.3).
As we have mentioned, in the extreme cases where or the variable becomes deterministic and, consequently, , which is in agreement with (2.32). E.g., setting and in formula (2.32) yields
2.38 |
and
2.39 |
Notice that the maximum of on [0, 1] is attained at . Actually, both and have one (simple) zero in [0, 1]. The maximum of on [0, 1] is attained at , while both and have one simple zero in [0, 1].
Let us also notice that, since
2.40 |
formula (2.32) can be also written as
2.41 |
From formula (2.41) it follows that for any given we have
2.42 |
where
2.43 |
An equivalent way to write is
2.44 |
(notice that , for , and ; furthermore, is a power series about with radius of convergence equal to 1).
We would like to understand the limiting behavior of , as , for any fixed . Formulas (2.12) and (2.42) suggest that we should work with the normalized variables
2.45 |
Let us consider the characteristic functions
2.46 |
In view of (2.28) we have
2.47 |
and using (2.47) in (2.29) yields
2.48 |
with .
From formula (2.48) one can guess the limiting distribution of : Assuming that for each the sequence has a limit, say , we can let in (2.48) and conclude that . Actually, if the limit exists, then formulas (2.49) and (2.50) below imply that and exists (in ). From the existence of these derivatives and the functional equation of we get that the function is continuous at 0 and satisfies the equation . Therefore, for all and, hence for some . Consequently, converges in distribution to a normal random variable, say X, with zero mean. As for the variance of X, formulas (2.42) and (2.45) suggest that , where is given by (2.43)–(2.44).
The above argument is not rigorous since it assumes the existence of the (pointwise) limit for every . It is worth, however, since it points out what we should look for.
From (2.12), (2.42), and (2.45) we get that for any fixed
2.49 |
thus
2.50 |
One important consequence of (2.50) is Durrett (2005) that the sequence , , where is the distribution function of , is tight.
Theorem 3
For and we have
2.51 |
where is given by (2.43)–(2.44) and Z is a standard normal random variable. In other words
2.52 |
Proof
By differentiating both sides of (2.48) with respect to t we get
2.53 |
We continue by differentiating times with respect to t both sides of (2.53), where . This yields
2.54 |
where
2.55 |
2.56 |
2.57 |
2.58 |
We will prove the theorem by induction. For and the truth of (2.51) follows immediately from (2.49) and (2.50). The inductive hypothesis is that the limit satisfies (2.51) or, equivalently, (2.52) for . Then, for (where ):
-
(i)Formula (2.55) implies
2.59 -
(ii)formula (2.56) implies
2.60 -
(iii)since , formula (2.57) implies
2.61 - (iv)
Let us set
2.65 |
Then, the inductive hypothesis together with (2.64) imply that the limit exists in . We can, therefore, apply Corollary A2 of the Appendix to (2.63) with , and as in (2.65) to conclude that the limit exists in . This allows us to take limits in (2.63) and obtain
or
2.66 |
where for typographical convenience we have set
2.67 |
We have, thus, shown that formula (2.66) holds for all . If , the sum in the right-hand side of (2.66) is empty, i.e. zero, and, hence we get what we already knew, namely that . If then (2.66) becomes vacuous (0 = 0), but we know that . Thus, formula (2.66) gives recursively the values of for every , and by using the induction hypothesis again, namely that for , it is straightforward to show that , as given by (2.66), equals for every m.
A remarkable consequence of Theorem 3 is the following corollary whose proof uses standard arguments, but we decided to include it for the sake of completeness.
Corollary 2
Let be the quantity given in (2.43)–(2.44). Then, for any fixed we have
2.68 |
where Z is a standard normal random variable and the symbol denotes convergence in distribution.
Proof
As we have seen, the sequence of the distribution functions , , is tight.
Suppose that , where , , is a subsequence of . Then (Durrett 2005) there is a sequence of random variables , converging a.s. to a random variable Y, such that for each the variables and have the same distribution (hence the limits Y and X also have the same distribution, since a.s. convergence implies convergence in distribution). Thus, by Theorem 3 we have
2.69 |
In particular, for we get from (2.69) that for all and, consequently (Chung 2001), that the sequence , , is uniformly integrable for every . Therefore in , for all (i.e. as , since by we denote the space of all real-valued random variables of our probability space , mentioned in the introduction, with a finite -norm), and (2.69) yields
2.70 |
It is well known (and not hard to check) that the moments of the normal distribution satisfy the Carleman’s condition and, consequently, a normal distribution is uniquely determined from its moments (Chung 2001; Durrett 2005) (alternatively, since the characteristic function of a normal variable is entire, it is uniquely determine by the moments). Therefore, it follows from (2.70) that X and have the same distribution. Hence, every subsequence of which converges in distribution, converges to and (2.68) is established.
From formula (2.45) we have
2.71 |
hence Theorem 3 has the following immediate corollary:
Corollary 3
For any and we have
2.72 |
Finally, let us observe that for any given we have (in view of (2.71) and Chebyshev’s inequality)
2.73 |
hence (2.50) yields
2.74 |
from which it follows by a standard application of the 1st Borel-Cantelli Lemma that if , , are considered random variables of the space , then, for any
2.75 |
The case of a general N
Let us now discuss the case of a general N, namely the case where N is not necessarily a power of 2. As we have described in the introduction, the first step of the binary search scheme is to test a pool containing samples from all N containers. If this pool is not contaminated, then none of the contents of the N containers is contaminated and we are done. If the pool is contaminated, we form two subpools, one containing samples of the first containers and the other containing samples of the remaining containers (recall that ). We continue by testing the first of those subpools. If it is contaminated we split it again into two subpools of and samples respectively and keep going. We also apply the same procedure to the second subpool of the samples.
Suppose is the number of tests required to find all contaminated samples by following the above procedure (thus , where is the random variable studied in the previous section). Then, as in formula (1.2),
3.1 |
In the extreme cases and the quantity T(N) becomes deterministic and we have respectively
3.2 |
Evidently, , where denotes the usual stochastic ordering (recall that means that for all ). In other words, T(N) is stochastically increasing in N. In particular
3.3 |
where is the logarithm to the base 2. Also, it follows easily from a coupling argument that if , then .
The expectation, the generating function, and the variance of T(N)
Theorem 4
Let us set
3.4 |
Then satisfies the recursion
3.5 |
(of course, ).
Proof
We adapt the proof of Theorem 1. Assume and let be the event that none of the N samples is contaminated. Then
and hence
3.6 |
where for typographical convenience we have set
3.7 |
In order to find a recursive formula for u(N) let us first consider the event that in the group of the N containers none of the first contain contaminated samples. Clearly and
3.8 |
Likewise, if is the event that in the group of the N containers none of the last contains contaminated samples, then
3.9 |
Let us also notice that , hence .
Now, (i) given and we have that , where has the same distribution as ; (ii) given and we have that ; finally (iii) given and we have that , where, is independent of and has the same distribution as . Thus (given ) by conditioning on the events , and we get, in view of (3.7), (3.8), and (3.9),
or
3.10 |
from which, in view of (3.6), formula (3.5) follows immediately.
Remark 2
By easy induction formula (3.5) implies that, for the quantity is a polynomial in q of degree N whose coefficients are , except for the constant term which is equal to . The leading term of is . In the special case where these properties of follow trivially from (2.3).
For instance,
3.11 |
while , , and are given by (2.11), since , , and .
With the help of (3.5) one can obtain an extension of formula (2.3) valid for any N. We first need to introduce some convenient notation:
3.12 |
so that , while we also have the standard convention (we furthermore have ). For example, if and, as in (3.3), , then
3.13 |
Corollary 4
Let be as in (3.4) and
3.14 |
(the last equality follows from the fact that ). Then
3.15 |
(if or , then the double sum in the right-hand side is 0).
Proof
By setting
3.16 |
and by recalling that and , formula (3.5) implies (in view of (3.12) and (3.14))
3.17 |
From (3.17) we obtain
or, in view of (3.13),
3.18 |
where (in the case , formula (3.18) is trivially true, since the sum in the right-hand side is empty). Consequently, (3.18) becomes
3.19 |
from which (3.15) follows immediately.
With the help of (3.15) we have created the following table for the average-case aspect ratio:
(as in Table 1, if we compare the above values with the graphs, given in Aldridge (2019) and Aldridge (2020), of the optimal average-case aspect ratio as a function of p, we can see that, if , our procedure is near-optimal).
Remark 3
Let us look at the case , where M is a fixed odd integer. We set
3.20 |
so that
3.21 |
Then (3.5) becomes
3.22 |
From (3.21) and (3.22) it follows that
3.23 |
thus
3.24 |
for all . It follows that, as ,
3.25 |
where
3.26 |
(as in the case of Remark 1, the convergence in (3.25) is much stronger, since for every the m-th derivative of with respect to q converges to the m-th derivative of uniformly on for any ).
From (3.26) it is obvious that if , then , except for at most countably many values of q (notice, e.g., that and for all M). Therefore, it follows from (3.25) that for
3.27 |
Comment
The reason we consider the case , where M is odd, is mainly theoretical: We use these values of N and the corresponding subsequences in order to demonstrate the non-convergence of T(N)/N. We do not claim any practical value of this choice of N.
Open Question
Is it true that ?
Next, we consider the generating function of T(N). Using the approach of the proofs of Theorems 2 and 4 we can derive the following result:
Theorem 5
Let
3.28 |
Then g(z; N) satisfies the recursion
3.29 |
for (of course, ).
Notice that in the case where , formula (3.29) reduces to (2.23).
By setting in (3.29) it follows that the characteristic function
3.30 |
of satisfies the recursion
3.31 |
with, of course, . For example, for formula (3.31) confirms the value of given in (2.30).
By differentiating (3.29) twice with respect to z and then setting and using (3.5) we get the following corollary:
Corollary 5
Let
3.32 |
Then satisfies the recursion
3.33 |
(of course, ).
In the case formula (3.33) reduces to (2.36) (e.g., if , then (3.33) agrees with (2.38)).
As we have mentioned, in the extreme cases where or the variable T(N) becomes deterministic and, consequently, , which is in agreement with (3.33).
Corollary 6
For there are constants , depending on q, such that satisfies
3.34 |
Proof
For we have , which implies that . Hence
3.35 |
Using (3.35) in (3.33) implies
3.36 |
Now, as we have already seen, , while by (2.38) we have . We can, thus, choose
3.37 |
Then, (3.37) implies that and from (3.36) we have , i.e. the first inequality in (3.34) is valid for and . Finally, the inequality for every follows easily from (3.36) by induction.
To establish the second inequality of (3.34) we set
3.38 |
Then, formula (3.33) implies (in view of (3.12))
3.39 |
where
3.40 |
Observe that by (3.40) and (3.1) we have that
3.41 |
Now, as in the case of (3.18) we have, in view of (3.39),
3.42 |
where . From (3.42) and (3.41) we get that is bounded. Therefore, the second inequality of (3.34) follows immediately from (3.38).
Of course, in the case where we have formula (2.42), which is much more precise than (3.34).
Let us also notice that with the help of (3.38), (3.42), and (3.15) we can get a messy, yet explicit formula for , extending (2.32) to the case of a general N.
The behavior of T(N) as
We start with a Lemma.
Lemma 1
Let as in the previous subsection. Then for any fixed we have
3.43 |
Proof
As we have seen in the proof of Corollary 6, is bounded, hence
3.44 |
Now, if we divide (3.33) by and invoke (3.34) we get
3.45 |
Therefore, by using (3.44) in (3.45) and invoking again (3.34) we obtain
3.46 |
which is equivalent to the first equality of (3.43). The second equality follows immediately.
In order to determine the limiting behavior of T(N), as , for any fixed , it is natural to work with the normalized variables
3.47 |
(Y(1) is not defined). Obviously,
3.48 |
thus (Durrett 2005) the sequence of the distribution functions of Y(N) is tight.
Let us, also, introduce the characteristic functions
3.49 |
In view of (3.47) and (3.30) we have
3.50 |
and by using (3.50) in (3.31) and then invoking (3.5) we obtain the recursion
3.51 |
with and taken from (3.49). Actually, if we define , then (3.51) is valid for .
We are now ready to present a general result, which can be viewed as an extension of Theorem 3 to the case of an arbitrary N.
Theorem 6
For and we have
3.52 |
where Y(N) is given by (3.47) and Z is a standard normal random variable. In other words
3.53 |
Proof
We will follow the approach of the proof of Theorem 3.
We apply induction. By (3.48) we know that (3.53) is valid in the special cases and . The inductive hypothesis is that the limit (where denotes the k-th derivative of with respect to t) satisfies (3.53) for . Then, for (where ) formula (3.51) implies
3.54 |
where
3.55 |
Let us set
3.56 |
Then, the inductive hypothesis together with Lemma 1 and (3.55) imply that the limit exists in . We can, therefore, apply Corollary A1 of the Appendix to (3.54) with , , (so that, in view of Lemma 1 and (4.6), ), and b(N) as in (3.56), to conclude that
3.57 |
Thus by taking limits in (3.54) we obtain (in view of (3.57) and Lemma 1)
or
3.58 |
We have, thus, shown that formula (3.58) holds for all . If , the sum in the right-hand side of (3.58) is empty, i.e. zero, and, hence we get what we already knew, namely that . If then (3.58) becomes vacuous (0 = 0), but we know that . Thus, formula (3.58) gives recursively the values of for every , and by using the induction hypothesis again, namely that for , it is straightforward to show that , as given by (3.58), equals for every m.
Theorem 6 together with the fact that the sequence Y(N) is tight imply the following corollary which can be considered an extension of Corollary 2. Its proof is omitted since it is just a repetition of the proof of Corollary 2.
Corollary 7
For any fixed we have
3.59 |
where Z is a standard normal random variable.
Finally, let us mention that, since , a rather trivial consequence of Theorem 6 is that for any given we have that
3.60 |
for any and any .
It follows that if , , is a sequence of integers (with ) such that the limit exists (e.g., or ), then
3.61 |
in the -sense, for every (the case is treated by Corollary 3).
Also, by Chebyshev’s inequality (applied to a sufficiently high power of ) and the 1st Borel-Cantelli Lemma it follows that
3.62 |
for any and any given . In this case, again, if exists for some sequence , then
3.63 |
Acknowledgements
The author wishes to thank the two anonymous referrees, as well as Professor Tom Britton, associate editor of JOMB, for their constructive comments and suggestions.
Appendix
Lemma A1
Let , , be a sequence of nonnegative real numbers such that there exists an integer for which
4.1 |
where , are nonnegative constants satisfying and . Then,
4.2 |
Proof
We have that . Thus, by taking in both sides of (4.1) we get
i.e.
4.3 |
Since , to finish the proof it suffices to show that is bounded , and we can, e.g., see that as follows. Let and
4.4 |
Then (noticing that (4.4) implies that ), easy induction on N implies that for all .
If , then (4.2) does not hold. For example, if for , then for all N.
Corollary A1
Let z(N), , be a sequence of complex numbers satisfying the recursion
4.5 |
where , , and b(N) are complex sequences such that
4.6 |
with . Then
4.7 |
Proof
Pick so that and then choose so that and for all . Then (4.5) implies
4.8 |
where , and the argument at the end of the proof of Lemma A1 applies to (4.8) and implies that the sequence z(N) is bounded.
Next, we write (4.5) as
4.9 |
where
4.10 |
Notice that our assumptions for , , and b(N), together with the fact that z(N) is bounded, imply that . Thus, if we take absolute values in (4.9) and set
4.11 |
we obtain
4.12 |
hence (since ) Lemma A1 implies that .
Finally, by using the above arguments we can also show the following simpler result:
Corollary A2
Let , , be a sequence of complex numbers satisfying the recursion
4.13 |
where and . Then,
4.14 |
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- Aldridge M (2019) Rates of adaptive group testing in the linear regime. In: IEEE International Symposium on Information Theory (ISIT), Paris, France, pp 236–240. 10.1109/ISIT.2019.8849712
- Aldridge M (2020) Conservative two-stage group testing, arXiv:2005.06617v1 [Stat.AP] (6)
- Armendáriz I, Ferrari PA, Fraiman D, Ponce Dawson S (2020) Group testing with nested pools, arXiv:2005.13650v2 [math.ST] (6)
- Chung KL. A course in probability theory. 3. San Diego: Academic Press; 2001. [Google Scholar]
- Dorfman R. The detection of defective members of large populations. Ann Math Stat. 1943;14(4):436–440. doi: 10.1214/aoms/1177731363. [DOI] [Google Scholar]
- Durrett R (2005) Probability: theory and examples, third edition. Duxbury Advanced Series, Brooks/Cole–Thomson Learning, Belmont
- Golliery C, Gossnerz O. Group Testing against Covid-19. Covid Econ. 2020;1(2):32–42. doi: 10.18231/j.ijirm.2020.023. [DOI] [Google Scholar]
- Hwang FK. A method for detecting all defective members in a population by group testing. J Am Stat Assoc. 1972;67(339):605–608. doi: 10.1080/01621459.1972.10481257. [DOI] [Google Scholar]
- Malinovsky Y. Sterrett procedure for the generalized group testing problem. Methodol Comput Appl Probab. 2019;21:829–840. doi: 10.1007/s11009-017-9601-4. [DOI] [Google Scholar]
- Mallapaty S. The mathematical strategy that could transform coronavirus testing. Nature. 2020;583:504–505. doi: 10.1038/d41586-020-02053-6. [DOI] [PubMed] [Google Scholar]
- Sobel M, Groll PA. Group testing to eliminate efficiently all defectives in a binomial sample. Bell Labs Tech J. 1959;38(5):1179–1252. doi: 10.1002/j.1538-7305.1959.tb03914.x. [DOI] [Google Scholar]
- Ungar P. The cut off point for group testing. Commun Pure Appl Math. 1960;13(1):49–54. doi: 10.1002/cpa.3160130105. [DOI] [Google Scholar]