Article

Group testing: Revisiting the ideas

Skorniakova Viktor
Vilnius University, Lituania
Leipus Remigijus
Vilnius University, Lituania
Juzeliunas Gediminas
Vilnius University, Lituania
K˛estutis Staliunas
Vilnius University, España

Group testing: Revisiting the ideas

Nonlinear Analysis: Modelling and Control, vol. 26, núm. 3, pp. 534-549, 2021

Vilniaus Universitetas

Recepción: 10 Julio 2020

Revisado: 21 Febrero 2021

Publicación: 01 Mayo 2021

Abstract: The task of identification of randomly scattered “bad” items in a fixed set of objects is a frequent one, and there are many ways to deal with it. “Group testing” (GT) refers to the testing strategy aiming to effectively replace the inspection of single objects by the inspection of groups spanning more than one object. First announced by Dorfman in 1943, the methodology has underwent vigorous development, and though many related research still take place, the ground ideas remain the same. In the present paper, we revisit two classical GT algorithms: the Dorfman’s algorithm and the halving algorithm. Our fresh treatment of the latter and expository comparison of the two is devoted to dissemination of GT ideas, which are so important in the current COVID-19 induced pandemic situation.

Keywords: group testing, quick sort algorithm, COVID-19.

1 Introduction

The task of identification of bad items in a given set of objects arises quite often. For example, consider identification of: (i) the infected patients in a fixed cohort or (ii) the defective items in the production batch. Usually, this identification task is a composite problem and spans many subtasks. One of such subtasks can be described as “an efficient utilization of resources devoted to testing of investigated objects”. It turns out that, among plenitude of context dependent methods designed for the solution of this subtask, the appropriately chosen testing plan plays an exceptional role since it alone can reduce the testing costs substantially. This is the contextual target of the present paper. To be more precise, we focus on testing strategies widely known under the name of Group (or Pooled Sample) Testing (in what follows, we make use of an abbreviation GT). The core idea underlying GT strategy is an observation that, in many cases, the testing of single items can be replaced by the testing of a group spanning more than one item. Though it is difficult to trace back the exact date and inventor of this cornerstone idea (for a good historical account, see [13, Chap. 1]), without doubt, much of the credit goes to the pioneering work of Dorfman [12]. In that paper, the blood testing problem was described, and the following scheme was suggested. Given . individual blood samples, pool them and test for the presence of an infection in the pooled sample; in case of the negative test – finish; in case of the positive test – retest each single patient. The rationale behind is clear: if the prevalence of the infection is low, one usually ends up with a single test applied to the pool instead of . tests applied individually.

Since appearance of Dorfman’s work [12] in 1943, GT ideas were evolving in many directions and found important applications in molecular biology, quality control, com- puter science and other fields. Digging into the literature, one can observe that it is indeed very widespread across different disciplines. Because of this reason, some developments were overlapping and rediscovered by researchers working in the different fields. Our personal familiarity with the field also underwent this route: attracted by potential appli- cations in the context of COVID-19 epidemics, we have rediscovered some well-known facts. Nonetheless, the attained experience and understanding of the importance of the tool inspired us to write a promotional paper on the topic. This is the main intent of the paper: we believe that, in the current pandemic situation, the spread of GT ideas and attraction of other researchers to the field is an important and meaningful task. We do not propose novel GT schemes or methodological improvements. Our presentation is primarily devoted to those unfamiliar with the subject aiming to provide a quick lightweight introduction “by example” without delving into details yet giving a flavor of the topic as a whole. Choosing a mathematical journal, we, first of all, were interested in the dissemination within the mathematically oriented community. Secondly, while getting familiar with the topic, we have encountered a lot of papers, where the subject was treated without sufficient mathematical rigorousness. We therefore felt that our rigorous treatment of the GT scheme H (see Section 2), unseen (or at least unobserved) by us, was a missing item in the existing literature. Finally, after submission of the initial version of the paper, we have discovered that our Proposition 2 adds some new information to what is known about classical Dorfman’s scheme (see the comments in Section 2).

The remaining part of the paper is organized as follows. In Section 2, we provide some preliminaries, then describe and contrast two classical GT schemes. In Section 3, we give an accompanying discussion highlighting some relevant issues and skim through the related literature. Appendices A and B contain some mathematical derivations and tables.

Because of COVID and the exemplary nature, we attach the whole presentation to the biomedical context

2 Two classical GT schemes

Consider the following setup. Assume that the prevalence of some disease (the fraction of the infected individuals) is equal to . (0, 1) in an infinite (or large enough) population. A cohort spanning N independent individuals has to be tested, and infected patients have to be identified. To achieve the goal, samples are collected from each individual. The applied test performs equally well for individual and for pooled samples: a situation might occur, e.g., when the test indicates the presence of the infection in the blood sample and there is no difference whether the latter is obtained from a single individual or from a pooled cohort of samples. For the situation described, physicians can choose different testing strategies. Let us assume that the following are three possible choices2.

Scheme A: Test each patient’s sample.

Scheme D: Conduct testing of the pooled sample. Test each member of the cohort sepa- rately only in case of detected infection in the pooled sample.

Scheme H:

Step 1. Test pooled sample of the whole cohort. Proceed to Step 2.

Step 2. If the test is positive, proceed to Step 3, otherwise, finish testing cohort.

Step 3. Divide the cohort into two parts consisting of the first and second halves, respectively. Apply the whole algorithm to the two obtained parts recursively.

Although it is not obvious at the first glance, Schemes D and H can be much more ef- ficient as compared to Scheme A, provided prevalence . is low enough. To give a rigorous justification (along with the concept of efficiency), let us formally define the underlying model.

Consider the sample of N individuals. Put Xi= 1, provided the test of i th individual is positive, and Xi = 0 otherwise. Let S = SN= X1 + ...+ XNbe the total number of infected individuals in the sample, and let T = T. be the total number of tests applied to the cohort

We start with Scheme D. The test is applied once if the result is negative, and it is further applied to each of N individuals otherwise, i.e.,

The above implies that X1, . . . , XNare independent identically distributed (i.i.d.) ran- dom variables each having Bernoulli distribution Be(p). Therefore, . has the binomial distribution Bin(N, p). Consequently, an average number of tests per cohort is

where q := 1 − p. An average number of tests per individual, say t = t(N), is

(1)

Consider a function given in (1). By equating its derivative to 0 wesee that the stationary points solve equation

(2)

which is a fixed point equation for and hence, can be easily solved iteratively. It is further not difficult to prove that, for . in the region enclosing (0, 0.2), there exists a unique solution Np > 0 of (2), which is a minimizer of t(N) (see Proposition 2 below). Then, turning back to economic/biomedical interpreta- tion, we conclude that, having a cohort of [Np] (here and in the sequel, [y] stands for an integer part of) individuals, Scheme D results in a lowest average number of tests per person, which is possible when applying scheme of this type for a population having prevalence .. Scheme A, in contrast, always has a constant number of tests 1 per person. Therefore, an average (absolute) gain attained applying Scheme D instead of Scheme A is given by the difference

Right panel in Fig. 1 shows the graph of which is an average gain measured by the number of tests saved per 100 individuals. The corresponding values are provided in Table B1 (see Appendix B). An accompanying graph of (see the left panel of Fig. 1) demonstrates dependence of an optimal sample size on .. To obtain a fast numerical evidence, assume that . is bounded away from zero and . Then from (2) it follows that the optimal sample size satisfies

Hence, assuming that . is small enough for pN ≈ 0 to hold, the above implies that

For example, if p = 0.01, then we have , i.e., an approximate average gain is 80% or so

Now let us switch to the Scheme H. Its main features are summarized in the following proposition (for the proof, see Appendix A).

Proposition 1.

Assume the Scheme .. Then

(i) an average number of tests per person is given by

(3)

(ii) an average number of tests per person in the case of an infinitely large cohort is

(iii) for a fixed , function admits at most two mini- mizers Np. the value N = Np corresponding to optimal sample size is either .

Inspection of the results in the statement of the proposition leads to a quick compari- son of Scheme H with A and D. Indeed, consider first the limit in (ii). Obviously,

Hence, for (or alternatively The latter means that, when the prevalence is low, this scheme always outperforms common sequential Scheme A. Again, to gain a quick quantitative insight, assume that p is small enough for to hold. Then turning to (iii) and taking a “continuous” (undiscretized) version of Npequal to ln yields relationships (see Remark A1, Eq. (A.3))

(4)

Therefore, an approximation to an average gainTaking, e.g., p = 0.01 results in G0.01 0.867. Considering analogous example given for Scheme D, we see that the gain has an increase close to 7%. In fact, this is not surprising (for a visual comparison of Schemes D and H on the linear and the log–log scales, see Fig. 1, and, for the numerical one, see Tables B1 and B2 in Appendix B) since for Scheme D, we had Equality (4), however, exhibits some magic flavor. To see this, note that, for . ≈ 0, entropy Ipof . X∼ Be(p) is asymptotically equivalent to p log2p since

Consequently, (4) means that the optimal average number of tests per one individual scales like entropy of the prevalence of the infection. Keeping in a view the above relationship, it is not surprising that the significant number of works [1, 8, 19] have approached the testing problem from the information theory perspective. In the next section, we provide additional comments regarding connections with the information theory. Here we end up with the previously mentioned Proposition 2, which is proved in Appendix A.

Proposition 2.

be fixed. Consider function It admits a unique fixed point N., which minimizes t(N)given by (1).

Scheme H (red) vs. Scheme D (black) on the linear and the log–log scale.
Fig. 1
Scheme H (red) vs. Scheme D (black) on the linear and the log–log scale.

The above proposition can be viewed as a counterpart of Proposition 1. Note that it does not contain analytical expression of an optimal sample size. The latter was given by Samuels [35] and is either 1 + p.1.2 or 2 + p.1.2 .Of most importance is that Samuels [35] not only provided the analytical expression of the optimal sample size but has also shown that, for the case of Scheme D, the optimal sample size equals to 1 for p > 1 − (1.3)1/3 ≈ 0.31. This, in turn, is in agre√e.ent with the fundamental fact of the GT theory discovered by Ungar [42]: if p ≥ (3 − 5).2 ≈ 0.38, then there does not exist an algorithm that is better than individual one-by-one testing.

An interesting detail here is that our proof given in Appendix A differs from that of Samuels and leads to an exact analytical expression for the range of . (the set . above), where gp(N) has a unique minimizer.

3 Discussion

Since its appearance, the Dorfman’s scheme D was rigorously investigated by many authors (we refer to [23, 35, 39, 40] to name a few). Talking about Scheme H, the situation is a bit different. To our best knowledge, the reference [46] is the only work close to ours both in nature of investigations and results. However, in that paper, the authors focus on the treatment of an asymptotic regime of Scheme H when Majority of other references encountered by us provide instructions suitable for the practical application of Scheme H with a brief and nonrigorous theoretical background. For example, in the present context, it was currently afresh discussed by Gollier and Gossner [18], Mentus et al. [29] and Shani-Narkiss et al. [37]. For an older reference discussing the case of nonhomogeneous population (i.e., the one in which the probability of being infected p may vary across individuals) and containing quite a large body of applied literature on halving algorithm (i.e., Scheme H), we refer to [4].

One should have noticed that halving, constituting the core of the Scheme H, yields another link to the information and algorithm theory in addition to the one already mentioned3 in Section 2. Namely, in its essence, Scheme H is nothing more than the quick sort (QS) algorithm designed to sort a set containing keys of two types (bad and good ones). It is well known that QS yields the best (up to the constant multiple) possible average performance among comparison-based algorithms: to sort an array having . nonconstant (i.e., random) items, the smallest average number of comparisons is of the order . ln . [10], and all randomized “divide-and-conquer” type algorithms (with QS being one among the rest) have expected time asymptotically equivalent to QS, which randomly splits sorted set into two equal subsets [11]. Our formula (3) is just a confir- mation of the well-known fact. To see this, note that, in the context of sorting task, (3) presents an average number of comparisons per item. Though the order is correct, we are inclined to think that the multiplier. can be improved by making use of QS modification (or another comparison-based algorithm) designed to sort items with a small number of possible values (in our case, there are just two values: “sick” and “healthy”). On the other hand, as already mentioned above, the order is optimal since though there are algorithms, which can beat QS when sorting integers, e.g., [2, 41], they operate under different, i.e., noncomparison-based, mode. In our case, however, comparison is predefined by the setting of the problem at hand: we assume that biomedical tests can only be carried out by making use of comparison.

Though biomedical context is very frequent in applications, there are many others including engineering, environmental sciences, information theory, etc. (see [6, 15, 21, 22, 24, 26, 28, 30, 32]). This “real life” contextual diversity brings many constraints to take into account despite the fact that the standard binomial setting, considered in Section 2, quite often can be regarded as a good starting approximation. To get a full understanding of the matter touched, below is a short list of key issues with a briefdescription of each.

*Heterogeneity of population. The prevalence of disease may depend on other factors (e.g., age and gender).

*Imperfectness of the test. The test can have sensitivity and/or specificity below 1.

*Dilution effect. Pooling can reduce testing accuracy substantially. If this is the case, it is necessary to impose upper bound on the number of pooled samples.

*Implementation costs. In Section 2, we silently assumed that implementation of the considered schemes only involves retesting related costs. However, it may involve others as well.

*Dependence. It can happen that tested individuals are somehow related.

All these underpinnings have to be addressed carefully. Take, for example, the last one. From results presented in Section 2 one can infer that the application of the GT procedures is most effective when the prevalence is low. In such case, under classical assumption pN λ > 0, the number of infected individuals S. can be well approximated by the Poisson distribution Pois(pN ), and the approximation remains quite accurate irrespectively of the nature of the dependence exhibited by summands (see [3, 7, 43] and references therein for results of this kind with possible extensions beyond the classical setting). It is therefore reasonable to assume that, after switching to Poisson approximation, at least some of the existing schemes can be carried over to the dependent case. Clearly, additional restrictions call for new theoretical investigations.

The set of directions of such investigations can be significantly appended by including other methods and GT related tasks. More concretely, the schemes considered in Section 2 broadly fall into the class of probabilistic GT schemes. Another widely adopted paradigm is called combinatorial approach. Within its framework, one does not assume any random mechanism and tries to make use of combinatorial methods in order to identify . bad items in the given group of N ≥ d objects (see monographs [13, 14]). Speaking about tasks, up to now, we have focused only on the identification of bad items (or infected patients) under assumption that the prevalence . is known. In addition to the literature devoted to this task, there is a huge body of literature dealing with an estimation (both point and interval) of p from pooled samples observations as well as testing issues (see, e.g., [16, 20, 34] and references therein).

We hope that our discussion complies well with our initial goal stated in the intro- duction. To emphasize the relevance of similar promotional discussions in the present context, we point out a huge burst of papers devoted to similar problems (see, e.g., [5, 17, 31, 33, 36, 38, 44, 45]). Besides that, we also note that some countries have already successfully applied pooling methodology for testing of the SARS-CoV-2 virus4.

Acknowledgments

We would like to thank two anonymous referees for insightful andconstructive comments, which helped to improve the preliminary version of the paper.

References

1. M. Aldridge, O. Johnson, J. Scarlett, Group testing: An information theory perspective, Found. Trends Commun. Inf. Theory, 15(3–4):196–392, 2019, https://doi.org/10.1561/ 0100000099.

2. A. Andersson, T. Hagerup, S. Nilsson, R. Raman, Sorting in linear time?, J. Comput. Syst. Sci., 57(1):74–93, 1998, https://doi.org/10.1006/jcss.1998.1580.

3. A.D. Barbour, L. Holst, S. Janson, Poisson Approximation, Clarendon Press, Oxford, 1992.

4. M.S. Black, C.R. Bilder, J.M. Tebbs, Group testing in heterogeneous populations by using halving algorithms, J. R. Stat. Soc., Ser. C, Appl. Stat., 61(2):277–290, 2012, https:// doi.org/10.1111/j.1467-9876.2011.01008.x.

5. A.Z. Broder, R. Kumar, A note on double pooling tests, 2020, arXiv:2004.01684.

6. T. Bui, M. Kuribayashi, M. Cheraghchi, I. Echizen, Efficiently decodable non-adaptive threshold group testing, IEEE Trans. Inf. Theory, 65:5519–5528, 2019, https://doi. org/10.1109/TIT.2019.2907990.

7. L.H.Y. Chen, Poisson approximation for dependent trials, Ann. Probab., .:534–545, 1975, https://doi.org/10.1214/aop/1176996359.

8. P. Chen, L. Hsu, M. Sobel, Entropy-based optimal group-testing procedures, Probab. Eng. Inf. Sci., .:497–509, 1987, https://doi.org/10.1017/S0269964800000541.

9. Wikipedia contributors, COVID-19 testing, 2021, https://en.wikipedia.org/ wiki/COVID-19_testing.

10. T.H. Cormen, C.E. Leiserson, R.L. Rivest, C. Stein, Introduction to Algorithms, 3rd ed., MIT Press, Cambridge, MA, 2009.

11. B.C. Dean, A simple expected running time analysis for randomized “divide and conquer” algorithms, Discrete Appl. Math., 154(1):1–5, 2006, https://doi.org/10.1016/j. dam.2005.07.005.

12. R. Dorfman, The detection of defective members of large populations, Ann. Math. Stat., 14(4):436–440, 1943, https://doi.org/10.1214/aoms/1177731363.

13. D. Du, F. Hwang, Combinatorial Group Testing and its Applications, 2nd ed., Ser. Appl. Math., Vol. 12, World Scientific, Singapore, 2000, https://doi.org/10.1142/4252.

14. D. Du, F. Hwang, Pooling Designs and Nonadaptive Group Testing: Important Tools for DNA Sequencing, Ser. Appl. Math., Vol. 48, World Scientific, Singapore, 2006, https://doi.org/10.1142/6122.

15. J. W. Fahey, P. J. Ourisson, F. H. Degnan, Pathogen detection, testing, and control in fresh broccoli sprouts, Nutr. J., .(13):1–6, 2006, https://doi.org/10.1186/1475-2891- 5-13.

16. European Centre for Desease Prevention and Control, Methodology for estimating point prevalence of SARS-CoV-2 infection by pooled RT-PCR testing, 2020.

17. C. Gollier, Optimal group testing to exit the Covid confinement, preprint, Toulouse School of Economics, Toulouse Cedex, 2020, https://www.tse-fr.eu/optimal-group- testing-exit-covid-confinement.

18. C. Gollier, O. Gossner, Group testing against Covid-19, Covid Economics, .:32–42, 2020.

19. L. Hsu, New procedures for group-testing based on the Huffman lower bound and Shannon entropy criteria, in N. Flournoy, W.F. Rosenberger (Eds.), Adaptive Designs, IMS Lect. Notes, Monogr. Ser., Vol. 25, IMS, Hayward, CA, 1995, pp. 249–262, https://doi.org/10. 1214/lnms/1215451490.

20. S.-H. Huang, M.-N.L. Huang, K. Shedden, W.K. Wong, Optimal group testing designs for estimating prevalence with uncertain testing errors, J. R. Stat. Soc., Ser. B, Stat. Methodol., 79(5):1547–1563, 2017, https://doi.org/10.1111/rssb.12223.

21. N. Johnson, S. Kotz, R. Rodriguez, Statistical effects of imperfect inspection sampling III. Screening (group testing), J. Qual. Technol., 20:98–124, 1988, https://doi.org/10. 1080/00224065.1988.11979092.

22. N. Johnson, S. Kotz, R. Rodriguez, Statistical effects of imperfect inspection sampling IV. Modified Dorfman screening procedures, J. Qual. Technol., 22:128–137, 1990, https://doi.org/10.1080/00224065.1990.11979224.

23. J.-K. Lee, M. Sobel, Dorfman and .1-type procedures for a generalized group-testing problem, Math. Biosci.,15:317–340, 1972, https://doi.org/10.1016/0025-5564(72) 90040-5.

24. J.T. Lennon, Diversity and metabolism of marine bacteria cultivated on dissolved DNA, Appl. Environ. Microbiol.,73(9):2799–2805, 2007, https://doi.org/10.1128/AEM. 02674-06.

25. K. Li, D. Precup, T.J. Perkins, Pooled screening for synergistic interactions subject to blocking and noise, PLoS One, .(1), 2014, https://doi.org/10.1371/journal.pone. 0085864.

26. E. Litvak, X. M. Tu, M. Pagano, Screening for the presence of a disease by pooling sera samples, J. Am. Stat. Assoc., 89(426):424–434, 1994, https://doi.org/10.1080/ 01621459.1994.10476764.

27. Y. Malinovsky, P.S. Albert, Revisiting nested group testing procedures: New results, comparisons, and robustness, Am. Stat., 73(2):117–125, 2019, https://doi.org/10. 1080/00031305.2017.1366367.

28. S. May, A. Gamst, R. Haubrich, C. Benson, D. M. Smith, Pooled nucleic acid testing to identify antiretroviral treatment failure during HIV infection, JAIDS, 53(2):194–201, 2010, https://doi.org/10.1097/QAI.0b013e3181ba37a7.

29. C. Mentus, M. Romeo, C. DiPaola, Analysis and applications of adaptive group testing method for COVID-19, 2020, https://www.medrxiv.org/content/10.1101/2020.04. 05.20050245v2.

30. O.T. Monzon, F.J. Paladin, E. Dimaandal, A.M. Balis, C. Samson, S. Mitchell, Relevance of antibody content and test format in HIV testing of pooled sera, AIDS, .:43–48, 1992, https://doi.org/10.1097/00002030-199201000-00005.

31. L. Mutesa, P. Ndishimye, Y. Butera, J. Souopgui, A. Uwineza, R. Rutayisire, E.L. Ndoricim- paye, E. Musoni, N. Rujeni, T. Nyatanyi, E. Ntagwabira, M. Semakula, C. Musanabaganwa, D. Nyamwasa, M. Ndashimye, E. Ujeneza, I.E. Mwikarago, C.M. Muvunyi, J.B. Mazarati, S. Nsanzimana, N. Turok, W. Ndifon, A pooled testing strategy for identifying SARS-CoV-2 at low prevalence, Nature, 589:276–280, 2021, https://doi.org/10.1038/s41586- 020-2885-5.

32. M.S. Nagi, L.G. Raggi, Importance to “airsac” disease of water supplies contaminated with pathogenic Escherichia coli, Avian Dis., 16(4):718–723, 1972, https://doi.org/10. 2307/1588749.

33. K. R. Narayanan, A. Heidarzadeh, R. Laxminarayan, On accelerated testing for COVID-19 using group testing, 2020, arXiv:2004.01684.

34. N.A. Pritchard, J.M. Tebbs, Estimating disease prevalence using inverse binomial pooled testing, J. Agric. Biol. Environ. Stat., 16(1):70–87, 2011, https://doi.org/10.1007/ s13253-010-0036-4.

35. S.M. Samuels, The exact solution to the two-stage group-testing problem, Technometrics, 20: 497–500, 1978, https://doi.org/10.1080/00401706.1978.10489706.

36. M. Schmidt, S. Hoehl, A. Berger, H. Zeichhardt, K. Hourfar, S. Ciesek, E. Seifried, Novel multiple swab method enables high efficiency in SARS-CoV-2 screenings without loss of sensitivity for screening of a complete population, Transfusion, 60:2441–2447, 2020, https://doi.org/10.1111/trf.15973.

37. H. Shani-Narkiss, O.D. Gilday, N. Yayon, I.D. Landau, Efficient and practical sample pooling for High-Throughput PCR diagnosis of COVID-19, 2020, https://www.medrxiv.org/ content/10.1101/2020.04.06.20052159v2.

38. N. Sinnott-Armstrong, D.L. Klein, B. Hickey, Evaluation of group testing for SARS-CoV-2 RNA, 2020, https://www.medrxiv.org/content/10.1101/2020.03.27. 20043968v1.

39. M. Sobel, Optimal group testing, in A. Rényi (Ed.), Proceedings of the Colloquium on Information Theory held at the University L. Kossuth in Debrecen (Hungary) from 19 to 24 September 1967, János Bolyai Mathematical Society, Budapest, 1967, pp. 411–488.

40. M. Sobel, P.A. Groll, Group testing to eliminate efficiently all defectives in a binomial sample, Bell Syst. Tech. J., 38:1179–1252, 1959, https://doi.org/10.1002/j.1538-7305. 1959.tb03914.x.

41. M. Thorup, Randomized sorting in .(. log log .) time and linear space using addition, shift, and bit-wise Boolean operations, J. Algorithms, 42(2):205–230, 2002, https://doi.org/ 10.1006/jagm.2002.1211.

42. P. Ungar, The cutoff point for group testing, Commun. Pure Appl. Math., 13(1):49–54, 1960, https://doi.org/10.1002/cpa.3160130105.

43. V. Cˇ ekanavicˇius, Approximation Methods in Probability Theory, Springer, Switzerland, 2016, https://doi.org/10.1007/978-3-319-34072-2.

44. J. Žilinskas, A. Lancˇinskas, M.R. Guarracino, Pooled testing with replication as a mass testing strategy for the COVID-19 pandemics, Sci. Rep., 11:3459, 2021, https://doi.org/10. 1038/s41598-021-83104-4.

45. I. Yelin, N. Aharony, E. Shaer Tamar, A. Argoetti, E. Messer, D. Berenbaum, E. Shafran,Kuzli, N. Gandali, O. Shkedi, T. Hashimshony, Y. Mandel-Gutfreund, M. Halberthal, Y. Geffen, M. Szwarcwort-Cohen, R. Kishony, Evaluation of COVID-19 RT-qPCR test in multi sample pools, Clin. Infect. Dis., 71:2073–2078, 2020, https://doi.org/10.1093/ cid/ciaa531.

46. N. Zaman, N. Pippenger, Asymptotic analysis of optimal nested group-testing procedures, Probab. Eng. Inf. Sci., 30:547–552, 2016, https://doi.org/10.1017/ S0269964816000267.

Appendix

A: Technical details

Proof of Proposition 1. (i) For 1 ≤ i ≤ j ≤ N = 2n, let Mij = {Xi, Xi+1, . . . , Xj}, and let S(i, j) = Σj Xk be the number of infected individuals in the cohort Mij. Let 1i,j denote an indicator function taking value 1 if there is at least one infected individual in the group Mi,j, i.e.,

Also, let T (i, j) be the total number of tests applied to the cohort Mij after the initial pooled test. By the description of the testing Scheme H, applying recursive equations, we have

Taking expectations yields

(A.1)

Hence, the first equality in (i). For the second one, take the last sum above and continue as follows

(ii) By (A.1)

(A.2)

(iii) Since N = N (n) = 2n, by the second equality in (A.2),

Clearly, q2N 0 as N . Therefore, there exist no more than two Np N such that ∆n ≤ 0 for all N ≤ Np and ∆n ≥ 0 for all N ≥ Np, and t(Np) attains its minimal value at Np. To find Np, we first solve 1 − 2q2N = 0 with respect to N , and then choose from the two nearest integers (i.e., [N ♩, [N ♩ + 1) the one which minimizes tN .

Remark A1. Note that if N ≥ 1 and pN → 0, then for t(N ) in (3), it holds

(A.3)

To see this, it suffices to use the following bounds:

Proof of Proposition 2. Step 1 (fixed points). Define

Then equation f′(N ) = 1 − e2N/v/√v = 0 is equivalent to N √= (v ln √v)/2. Note that f′(N ) → −∞ as N → ∞. Moreover, f′(0) > 0 since 1 − (1/ v) > 0 ⇔ p < 1 − e−4,

which is satisfied for any p ∈ A. Therefore, f attains maximal value at

(A.4)

And

since ln √v − 1 > 0 ⇔ p < 1 − e−4/e2 . On the other hand, f (0) = −√v/2 < 0 and f (N ) . Consequently, f has two zeroes: Np (0, Nmax) and Np (Nmax, ). The latter means that g has two fixed points.

Step 2 (minimizer). In this step, we show that Np from Step 1 is the minimizer for

t(N ) given in (1). By (2),

Hence,

(A.5)

From Step 1 (see (A.4)) it follows that Nmax/(v/2) = ln √v > 1, i.e., v/√2 ∈ (0, Nmax) Therefore, (A.5) holds if and only if f (v/2) > 0. The latter reads as (v e v)/2 > 0 and is equivalent to p < 1 − e−4/e showing that Np (being the critical point of t) is indeed the announced minimizer. Finally, note that the above analysis also implies that Np from Step 1 is the maximizer of t(N ), which affirms the uniqueness of the minimizer.

Remark A2. One can also show that p Np is strictly decreasing and continuous on A. However, the latter properties seem to be of less importance, and we omit the details.

B: Tables

In the tables below, the following information is provided:

  1. Column “Np” shows an optimal sample size corresponding to p ranging in an interval given in the column “Range of p”.

    Column “Range of 100Gp” shows an average gain (as defined in the main body of the paper) per 100 individuals corresponding to values of p and Np given in the two leading columns. The highest gain corresponds to the lowest p in the corresponding interval. For example, in Table B1, the first line should be interpreted as follows: for p 2 [0:1865; 0:2000], optimal sample size Np is equal to 2; if p = 0:2000, then average gain per 100 individuals 100Gp is equal to 16.1782; if p = 0:1865, then 100Gp = 14:000; for intermediate values of p, the value of 100Gp lies in [14:0000; 16:1782].

Table B2
Performance of Scheme D.
Performance of Scheme D.

Table B2.
Performance of Scheme H
Performance of Scheme H

Notes

1 The author is supported by grant S-COV-20-4 from the Research Council of Lithuania.
2 The notations for a short reference of GT schemes come from Dorfman (D); Halving (H). Scheme A reflects the most naive and straightforward option.
3 3The discussed appearance of entropy in formula (4), in fact, is a simple conclusion following fromShannon’s coding theory; a bit more on the connections with that theory can be found in Appendix H of theSupplementary Material of the reference [27].
4 4According to Wikipedia [9], “In Israel, researchers at Technion and Rambam Hospital developed a methodfor testing samples from 64 patients simultaneously, by pooling the samples and only testing further if thecombined was positive. Pool testing was then adopted in Israel, Germany, Ghana, South Korea, Nebraska,China and the Indian states Uttar Pradesh,West Bengal, Punjab, Chhattisgarh, and Maharashtra.” Also see “Listof countries implementing pool testing strategy against COVID-19” therein.
HTML generado a partir de XML-JATS4R por