 Technical advance
 Open Access
 Published:
Group testing performance evaluation for SARSCoV2 massive scale screening and testing
BMC Medical Research Methodology volume 20, Article number: 176 (2020)
Abstract
Background
The capacity of the current molecular testing convention does not allow highthroughput and community level scans of COVID19 infections. The diameter in the current paradigm of shallow tracing is unlikely to reach the silent clusters that might be as important as the symptomatic cases in the spread of the disease. Group testing is a feasible and promising approach when the resources are scarce and when a relatively low prevalence regime is observed on the population.
Methods
We employed group testing with a sparse random pooling scheme and conventional group test decoding algorithms both for exact and inexact recovery.
Results
Our simulations showed that significant reduction in per case test numbers (or expansion in total test numbers preserving the number of actual tests conducted) for very sparse prevalence regimes is available. Currently proposed COVID19 group testing schemes offer a gain up to 15X20X scaleup. There is a good probability that the required scale up to achieve massive scale testing might be greater in certain scenarios. We investigated if further improvement is available, especially in sparse prevalence occurrence where outbreaks are needed to be avoided by population scans.
Conclusion
Our simulations show that sparse random pooling can provide improved efficiency gains compared to conventional group testing or ReedSolomon error correcting codes. Therefore, we propose that special designs for different scenarios could be available and it is possible to scale up testing capabilities significantly.
Background
The first half of the year 2020 has been in the midst of the first wave of COVID19 pandemic. It is expected that, with the strict social distancing applications, the spread of SARSCoV2 is going to be stabilized or sustained at several geographies [1]. However, second wave epidemics and resurgences are highly likely to occur [2]. Moreover, different geographies, due to segregation by administrative borders, asynchronously experience epidemic growths. While social distancing and quarantine measures have been the primary factor to contain SARSCoV2, widespread testing and aggressive contact tracing are observed to be effective to isolate the spreaders from vulnerable populations. According to this, widespread scanning of populations or subpopulations is crucial to locate and isolate epidemic clusters, especially by detecting asymptomatic carriers. Excessive numbers of asymptomatic carriers appear to be especially important as they might be the most important factor in the difficulty of lowering the reproductive numbers. Unbiased estimation of the ratio of asymptomatic or presymptomatic spreaders might be difficult to assess. Yet, statistics from small or mediumsized cohorts and case studies indicate that they might be as abundant as 10%50% [3, 4]. These ratios possess special importance regarding the case studies observing that asymptomatic/presymptomatic carriers are likely to infect their contacts [5, 6].
Currently, PCRbased molecular testing primarily serves as the standard diagnostics in the majority of health systems. Since the infrastructure and financial limitations dictate the allocation of resources for higher priority cases, the capacity of the current molecular testing convention does not allow highthroughput and community level scans. Therefore, the focus of testing is on hospitalized symptomatic cases and subsets of their contacts reached via tracing. The diameter in the current paradigm of shallow tracing is unlikely to reach the silent clusters that might be as important as the symptomatic cases in the spread of the disease.
Group testing is a feasible approach when the resources are scarce and when a relatively low prevalence regime is observed on the population [7]. It allows scanning large populations by pooling samples and conducting orders of magnitude lower number of tests while being able to locate the positive cases. Starting from its initial practical use in US Army in the 1940s [8], screening recruits for syphilis, group testing has been popularly employed in several different fields where pooling/mixing and testing was available to detect the defective elements. In general, group testing algorithms seek for \(T= \mathcal {O}(k \log n)\) measurements in order to correctly detect k positive cases in n testing samples. Here, \(\mathcal {O}(.)\) denotes that the measurements are in the order of magnitude of the argument. Practical algorithms are shown to achieve \(T\approx 2.72 \frac {\alpha }{1\alpha }\ln {{n}\choose {k}}\) while k scales sublinearly as \(k = \mathcal {O} \left (n^{\alpha }\right)\) [9]. With an assumption that early phases of epidemic waves or screening of asymptotic carriers would form populations of relatively low prevalence, conventional group testing schemes can potentially provide significant gains in testing capability.
Complying with the idea of employing group testing in SARSCoV2 screening, SinnottArmstrong et al. [10] recently proposed an adaptive scheme, pooling rows and columns of PCR wellplates, and they showed that up to 10X efficiency gain could be achieved in 0.1%5% prevalence band. Eberhardt et al. [11] used multistage group testing with two or three rounds. Their results imply even greater efficiency gains of 15X20X around low prevalence bounds. While these are very practical and simple schemes eliminating the requirements of robotic pipetting needs and reducing the pipetting errors, more sophisticated pooling schemes that can improve the efficiency gains are available. Indeed, Shental et al. [12] used sampling 48 pools out of 384 wellplate by ReedSolomon coding, and showed 8X efficiency gain around the band of 1% prevalence. These schemes offer very significant scaleups in testing capacity, especially around midscale sparsity bands (i.e. 1%5% prevalence). However, in potential scenarios of achieving widespread or community level scans with lower expected prevalence, further efficiency gains would be required. Therefore, designing group testing schemes aiming sparsity is beneficial to explore the practical capabilities.
We employed group testing with a sparse random pooling scheme and conventional group test decoding algorithms both for exact and inexact recovery. Our simulations showed that significant reduction in per case test numbers (or expansion in total test numbers preserving the number of actual tests conducted) for very sparse prevalence regimes is available.
Methods
We designed a randompooling based group testing scheme for different laboratory setups including 96, 384, and 1536well plate options. Two different approaches were considered i) twopass adaptive testing for exact recovery. ii) singlepass nonadaptive testing for approximate recovery of test results. For both approaches, we use the same random pooling design.
Random pooling
Random pooling procedure samples a pooling procedure from the space of m pools out of n samples by adding each sample exactly to k pools. The distribution to pools is selected without replicating the same pattern to be able to distinguish samples. Also, the sampled pooling matrices are intended to have low coherence, as low coherence improves the rate (i.e. efficiency gain in this setting) in sparse recovery problems [13]. The coherence of a measurement matrix denotes the maximum inner product of its column vector pairs, and it is associated with the maximum number of pools a couple of samples coexist in our case. We achieve low coherence by Metropolis sampling algorithm [14] with a modification that acceptance is driven by coherence, rather than density. According to that, the pooling matrix is iteratively perturbed via changing the pools that a given sample is contributing to, with a small probability p. Perturbed matrices are accepted with a probability proportional to the decrease in coherence as suggested by Metropolis sampling. This MonteCarlo random walk in the pooling matrix space is expected to design low coherence random pooling procedures. The random pooling algorithm is provided below, where m by nΦ denotes the binary pooling matrix. Φ_{(i,j)} indicates inclusion of sample j in the i^{th} pool. Φ_{(:,i)} denotes the i^{th} column of Φ and α is inversely related to the coherence of Φ (i.e. it attains k in minimum coherence and 0 in maximum coherence).
Figure 1 illustrates a toy example of the sampling scheme for the set of 12 samples gathered in 6 pools, where each sample is included in two pools (i.e. n=12, m=6, k=2). The pooling is represented as a bipartite graph, and the corresponding pooling matrix Φ is shown. The pooling schemes proposed by SinnottArmstrong et al. and Eberhardt et al. could be considered as specific instances of the proposed pooling scheme where k=2 and k=1 respectively.
Twopass testing
This scheme concentrates on two rounds of testing. The first test round employs random pooling and testing of pooled samples. Confidently negative samples are eliminated using a possible positive detection algorithm. At this point, a subset of the samples is determined as possible positives for further testing, and no false negatives are eliminated. Testing the possible positives directly for eliminating the false positives concludes the testing process without false calls. Note that, perfect recovery in this sense assumes single tests are accurate.
We have employed the straightforward “Definite Defectives” algorithm [15] to detect the possible positives in the first round. Once the possible positives are subject to the second round, the remaining procedure includes the conventional “one testone sample” approach. The procedure is given as follows.
Here, we refer to the binary vector y as the pooled test results where 1s are the positive and 0s are the negative pool results. \(\Omega _{i}^{row}\) represents the support set of the i^{th} row in Φ pooling matrix determined by random pooling scheme, and T sets represent the set of positive candidates.
Singlepass testing
Singlepass testing includes only the approximate decoding of pooled samples in one round. We considered Sequential Combinatorial Orthogonal Matching Pursuit (COMP) algorithm [15] for decoding. According to this, among the set of possible defectives, the minimal set explaining the pooling test results is considered using a greedy forward addition. This set is the predicted positives for the entire test population. Unlike the previous approach, singlepass testing is vulnerable to false positives and negatives, whenever the pooling scheme is underpowered to explain the actual population uniquely.
Here, similarly, we refer to the binary vector y as the pooled test results. ω^{y} represents the support set of y, \(\Omega _{i}^{row}\) represents the support set of i^{th} row in Φ, \(\Omega _{i}^{col}\) represents the support set of i^{th} column in Φ, and T sets represent the set of positive candidates.
Since the pooling matrix design procedure is stochastic, it is possible to design several different pooling schemes. The decoding algorithms are driven by pooling matrices, therefore twopass and singlepass testing instances may vary physically.
Simulations
During our experiments, we conducted the simulations under the following conditions. Random pooling was performed with p=0.05 perturbation probability applied to each samplepool pairing independently for 100.000 iterations. For each pooling setup, sample sets were simulated assuming a Bernoulli process where the probability of a positive sample draw was estimated to be the prevalence. For twopass testing, the mean efficiency gain is reported at each prevalence sampled over a number of independent simulations. In order to attain the detection performance of singlepass testing, the number of pools were incremented with new pooling designs until the perfect recovery of single tests was achieved at each prevalence instance. Similarly, the mean performances of a number of simulations were reported.
Results
In order to observe the performance of group testing under different regimes, simulations under Bernoulli sampling were repeated and reported. The prevalence range between 0.0005 and 0.2 was chosen as the simulation operating region. Efficiency gain, defined as the ratio of tested subjects to the number of actual tests performed in pools was used as the performance metric. Three pooling strategies, each corresponding to a different plate operation were considered: 96well plate, 384well plate, and 1536well plate.
Twopass testing
In case of exact recovery of one sampleone test naive testing, which would fix the sensitivity and specificity of group testing approximately to that of PCR technical limitations, the twopass procedure explained in the methods, referred to as “Definitive defectives” was considered. For 96well plate set up, 10 pools are considered for the first round group testing and each sample is represented exactly in three different pools. The preferred number of pools was 15 with each sample being distributed to exactly 4 pools for 384well plate setting, also 20 pools with each sample being distributed to exactly 4 different pools were considered for 1536well plate set up. The corresponding random pooling strategies construct sparse mixing matrices which are expected to result in good performance of decoding algorithms (i.e. detecting a small number of possible positives with ideal specificity). Consequently, greater efficiency gain values are expected compared to conventional grouping.
In order to observe how this pooling strategy scales for low prevalence scenarios compared to previously proposed techniques, we considered the group testing scheme suggested by SinnottArmstrong et al. (which we refer to as “rowcolumn grouping”) for the same setup. Multistage testing strategy proposed by Eberhardt et al. [11] was also implemented in the same setup. From several two or threestage testing schemes, pools of 16 with three stages (named as “Multistage P16S3”) and pools of 32 with two stages (named as “Multistage P32S2”) were observed to be the most efficient strategies throughout the prevalence spectrum. Therefore, we included these two in our comparison. Finally, a wellknown group testing strategy proposed by Phatarfod and Sudbury [16], which is an adaptive version of “rowcolumn grouping” was considered as a reference. While underperforming in the small plate setup (96well plate), the proposed group testing resulted in higher efficiency in sparse prevalence regimes (i.e. <0.5%) for medium (384well plate) and large plate (1536well plate) setups (Fig. 2). Especially, for the large plate setting, the efficiency gain was observed to achieve more than twice the rates of other schemes. Each observed value is the recorded mean of 1000 independent simulations. This came with the tradeoff of underperforming in denser areas (0.2% to 1% regions).
Singlepass testing
We have considered the scenarios in which the preservation of naive test accuracy is not the primary concern. A widespread scanning concept might be focusing on determining the hotspots within a population rather than diagnosing each individual for COVID19. In such a case, test accuracy could be traded off in return of gains in testing capacity. Since singlepass group testing schemes are underdetermined systems analogous to compressed sensing, reduction in the number of pools to be tested will translate to reduction in testing accuracy, which corresponds to the tradeoff operation between efficiency gain and test accuracy. To observe these characteristics we conducted the simulations in similar prevalence ranges as the twopass process. However, in this case, random pools were decoded to sample test results in a single pass using sCOMP decoding as described in the methods section. The accuracy increase was recorded with respect to increasing number of random pools at each prevalence instance. Starting from the same setting with twopass testing pooling scheme, new pooling procedures were sampled at each pooling expansion. The parameter k, which corresponds to the number of pools that a sample contributes to, was incremented by one at the introduction of every 10 new pools.
As the performance metric, we chose sensitivity to reflect the operation success of group testing scheme. As expected, the specificity measures were close to 1 throughout the operational regions (Fig. 4). Figure 3 shows the efficiency gain operations, indicating operation at each sensitivity level as separate contours. Following similar characteristics with twopass testing, pooling from larger well plates yields greater gains, especially for very sparse regimes. Typically, the trend of equalsensitivity contours exhibits greater slopes as the prevalence gets smaller, indicating better efficiency gain boosts for lower sensitivity tradeoff.
Accuracy of singlepass testing
Sequential COMP decoding employed for singlepass decoding of group testing by random pooling operates on a greedy algorithm, selecting a suboptimal minimal subset satisfying the pooled test results. Clearly, the solutions are not unique, and in case several linear combinations cover the same support set of pools, the decoding is vulnerable to errors. Lowering the number of random pools drives the system to be more underdetermined, expecting a greater number of errors on average. This can be observed from our simulations. Observing the empirical results, the specificity has never been experienced to be lower than 0.97, which is expected since the negative/positive distribution is unbalanced. This is also an indication that the decoding is not overexploring the positives. A more rigorous or liberal decoding scheme, allowing relaxations such as linear programming might be further investigated to observe if any improvement in decoding accuracy is achievable empirically.
Singlepass vs. twopass testing
Although singlepass and twopass schemes presented are not directly comparable since the former is a partial recovery method and the latter is an exact recovery method, it is interesting to compare their efficiency gains where singlepass testing performs close to exact recovery. According to our simulations, operation characteristics of singlepass testing converges to rowcolumn group testing as the prevalence increases, and it operates close to the proposed twopass testing in lowprevalence regions (Fig. 5). It might be possible to perform singlepass testing with appropriate random pooling schemes even in high prevalence regimes.
Figure 4 provides the mean efficiency gain performance of the two proposed schemes. Roughly, the single pass scheme approximates the gains of random pooling and rowcolumn pooling schemes at both sides of the prevalence spectrum. The single pass scheme appears to perform similar to twopass complete recovery around near errorless operation (e.g. sensitivity >0.99). While this is practically preferable, as singlepass testing requires only one round of testing being more efficient in time requirements, it might be more difficult to design a singlepass test. It should be noted that each efficiency gain requires different random pooling numbers for different prevalence regimes and a correct guess of prevalence should be predicted beforehand. In case prevalence distribution attains high variance, twopass testing might be a safer approach.
Figure 5 depicts the expansion rates, showing how many folds the testing capacity can be expanded to include presumably negative dominant scan population (e.g. extending the tracing to not only symptomatic but also to asymptomatic contacts) conducting the same number of tests. As expected, it would be possible to achieve larger expansions at low prevalences, especially with large plate setups. On the other hand, still, room for a fewfold expansion seems to be achievable around relatively large prevalences such as 10%. This might be an indicator that group testing could be a feasible operation at a large spectrum.
Singlepass testing vs. PBEST testing
PBEST testing was shown to perform accurately around the band 0.2%0.13% with 8X efficiency gain. While PBEST incorporates a specific ReedSolomon errorcorrecting code scheme for pooling, we ran simulations around this band with similar aspects (i.e. each sample contributes to exactly 6 pools in 384well plate setting) to observe the efficiency gain rates around perfect recovery operating points. According to the averages over 3000 simulations, random pooling performs similarly around 1% prevalence (Table 1). Note that this observation is not based on practical experiments neglecting any physical noise source. In order to validate the results and draw a confident conclusion, further studies and wetlab experiments are needed.
Discussion
Conducting an effective test, trace & isolate solution might be the key for containing SARSCoV2 and an elemental part of fighting against the current COVID19 pandemic. Perhaps the most significant and challenging issue is performing massivescale tests. Asymptomatic cases, being the possible silent culprits of undetected infection clusters, are off the radar of conventional testing procedures. In order to expand the radius of contact tracing to subpopulation scan level or performing community scans, given the current testing infrastructures and resources, efficient group testing strategies are needed. Currently proposed COVID19 group testing schemes offer a gain up to no more than 10X and 25X scaleup for medium (0.5%10%) and low (<0.5%) prevalence respectively [10–12, 16]. There is a good probability that the required scale up to achieve massive scale testing might be greater in certain scenarios. We investigated if further improvement is available, especially in sparse prevalence occurrence where outbreaks are needed to be avoided by population scans.
We have tested our proposed pooling schemes for standard 96, 384, and 1536well plate settings. The reason why testing schemes were restricted to standard plate settings is the practical concern that manual or robotic laboratory handling conventions generally follow standard well setups and integration/adoption in these circumstances is more feasible. Our simulations indicate that the greatest efficiency gains promised by the proposed pooling schemes are achieved at very low prevalence rates for 384 and 1536well plate settings. This observation leads to the fact that such a highefficiency gain approach should be limited to certain testing scenarios. The proposed group testing, along with the other approaches reviewed previously, would not be helpful in the diagnostic use for symptomatic cases around outbreak peaks, for instance when the prevalence in a tested population is greater than 10%. On the other hand, a few cases at the early phases of an outbreak are worth to detect, as they might be the seeds for exponential growth of epidemic curves. Widespread testing for early detection of emerging clusters therefore might be a powerful preventive approach. Our results claim that random pooling might be a preferable group testing strategy in similar scenarios. Another use case might be scanning for asymptomatic cases, separate from diagnostic testing on symptomatic individuals. This might be either as extending contact tracing to be able to test greater number of case contacts or as a periodic scan of a population (e.g. healthcare workers and other risk groups). In case of community scans in which individuallevel sensitivity can be sacrificed in return of spanning larger populations, singlepass group tests might also be useful. We have observed that for singlepass group testing, a small decrease in sensitivity could enable large efficiency gains at low prevalence groups. Such a highthroughput group testing on relatively large populations could be devised as a tool for locating infection hot spots.
Assume that contact tracing in a relatively low prevalence cluster is being conducted using naive testing. It might be useful to extend tracing by testing severalfold more individuals but with the same number of tests, using group testing. The extended subpopulation may be very low prevalence, however, detecting any asymptomatic case in it could be very valuable. We ran simulations to investigate what should be the initial prevalence for such an expansion opportunity, and how many folds of an extension would be available in the proposed group testing scheme. We observed that up to 60X expansion is available around 2%3% prevalence band (Fig. 6). This result might be an implication that large scale contact tracing might be possible at early forming clusters.
It should be kept in mind that our experiments are simulations designed on two major assumptions. First, we assumed that pooling will not attenuate the detection efficiency of molecular detection. Lohse et al. [17] and Yelin et al. [18] independently showed that pooling a single positive SARSCoV2 sample with tens of negative samples does not harm detectability of positives in case of RTqPCR testing. The pooling size in our simulations does not exceed the experimented limits, so we believe the proposed pooling schemes should act similarly in practical settings. Our second assumption is that the molecular tests are accurate enough to detect the samples without errors. Therefore, all simulations are performed for a noiseless scenario. While the actual picture might not comply with this assumption, the experiments ran with PBEST approach indicate that the imperfection of RTqPCR tests does not have a negative effect on group testing. In fact, the encoding procedures of mixing matrices might even work as errorcorrecting codes for low prevalence. However, further studies and wetlaboratory experiments should be conducted to support this claim.
Another aspect of group testing is that the efficiency gains mentioned in this and previous studies do not translate as linear gains in SARSCoV2 testing economy. Group testing only amplifies the number of tests, however previous steps of sampling, logistics, and any sample preparation steps remain unscaled. Feasibility of prepooling operations is out of the scope where it might be another serious concern. Moreover, group testing might be subject to greater use of consumables (e.g. more pipetting), imposing the risk of mixing errors or the need for robotics systems.
Although the current popular approach of COVID19 testing is RTqPCR based molecular tests, clearly group testing is not limited to PCR based molecular techniques. Any testing scheme preserving positivity as well as lack of signals in the absence of positives while combining samples in pools can be trivially integrated to group testing approaches. IgC/IgM based serological tests or any prospective test strategies (e.g. CRISPRCas12 based testing [19]) should be considered for large scale testing/scanning with group testing under appropriate circumstances. It should be noted that the current concept of group testing relies on recovering the actual sample profile from binary outcomes of pooled tests. Testing schemes allowing quantitative outputs might be possible and further quantitative information might enable unprecedented efficiency gains. This opportunity can be hypothesized following the results of compressed sensing literature. For example, for the case of Sudocodes [20], at a prevalence of 0.1%, 1 Million subjects can be scanned by performing under 10.000 tests even with perfect recovery. Of course, this requires very precise quantitative and unique measures for each test subject. It might be a promising research direction to investigate such testing options. For example, shallow shotgun sequencing allowing positive signals while pronouncing specific sequences belonging to each sample could be a candidate for ultrahigh throughput group testing. Proposed group testing algorithms assume and simulate under the circumstances that each sampled instance is independent. On the contrary, we know that COVID19 sampling is mainly conducted on populations with social contacts, violating the independence assumptions. Sophisticated decoding algorithms informed by metadata might perform superior, allowing greater efficiency gains, while complicating the overall testing procedure in practice. Nevertheless, we believe that even in primitive forms, group testing is a promising direction for efficient allocation of testing resources that should be considered for practical use.
Conclusions
Our simulations show that sparse random pooling can provide improved efficiency gains compared to conventional group testing or ReedSolomon errorcorrecting codes. Therefore, we propose that special designs for different scenarios could be available and it is possible to scale up testing capabilities significantly.
Availability of data and materials
The codes, protocols, and datasets used during the current study are available from the corresponding author on reasonable request.
Abbreviations
 COMP:

Combinatorial orthogonal matching pursuit
 COVID19:

Coronavirus disease 2019
 PCR:

Polymerase chain reaction
 PBEST:

Poolingbased efficient SARSCoV2 testing
 RTqPCR:

Realtime quantitative polymerase chain reaction
 SARSCoV2:

Severe acute respiratory syndrome coronavirus 2
 sCOMP:

Sequential combinatorial orthogonal matching pursuit
References
Anderson RM, Heesterbeek H, Klinkenberg D, Hollingsworth TD. How will countrybased mitigation measures influence the course of the COVID19 epidemic?Lancet. 2020; 395(10228):331–4.
Colbourn T. COVID19: extending or relaxing distancing control measures. Lancet Public Health. 2020; 5(5):236–7. https://doi.org/10.1016/S24682667(20)300724.
Mizumoto K, Kagaya K, Zarebski A, Chowell G. Estimating the asymptomatic proportion of coronavirus disease 2019 (COVID19) cases on board the Diamond Princess cruise ship, Yokohama, Japan, 2020. Eurosurveillance. 2020; 25(10):2000180.
Nishiura H, Kobayashi T, Miyama T, Suzuki A, Jung S, Hayashi K, Kinoshita R, Yang Y, Yuan B, Akhmetzhanov AR, et al. Estimation of the asymptomatic ratio of novel coronavirus infections (COVID19). Int J Infect Dis. 2020; 94:154–5.
Bai Y, Yao L, Wei T, Tian F, Jin DY, Chen L, Wang M. Presumed asymptomatic carrier transmission of COVID19. Jama. 2020; 323(14):1406–7.
Gandhi M, Yokoe DS, Havlir DV. Asymptomatic transmission, the Achilles’ heel of current strategies to control COVID19. N Engl J Med. 2020.
Du D, Hwang F. Combinatorial group testing and its applications, Vol. 12. Farrer Road: World Scientific; 2016.
Dorfman R. The detection of defective members of large populations. Ann Math Stat. 1943; 14(4):436–40.
Aldridge M, Johnson O, Scarlett J, et al. Group testing: an information theory perspective. Found Trends® Commun Inform Theory. 2019; 15(34):196–392.
SinnottArmstrong N, Klein D, Hickey B. Evaluation of group testing for SARSCoV2 RNA. medRxiv. 2020. https://doi.org/10.1101/2020.03.27.20043968.
Eberhardt JN, Breuckmann NP, Eberhardt CS. Multistage group testing improves efficiency of largescale COVID19 screening. J Clin Virol. 2020; 128,:104382.
Shental N, Levy S, Skorniakov S, Wuvshet V, ShemerAvni Y, Porgador A, Hertz T. Efficient high throughput SARSCoV2 testing to detect asymptomatic carriers. medRxiv. 2020. https://doi.org/10.1101/2020.04.14.20064618.
Andryani NAC, Pradnyana KD, Gunawan D. The Critical Study of Mutual Coherence Properties on Compressive Sensing Framework for Sparse Reconstruction Performance: Compression vs Measurement System. J Phys Conf Ser. 2019; 1196(1):012074.
Metropolis N, Rosenbluth AW, Rosenbluth MN, Teller AH, Teller E. Equation of state calculations by fast computing machines. J Chem Phys. 1953; 21(6):1087–92.
Aldridge M, Baldassini L, Johnson O. Group testing algorithms: Bounds and simulations. IEEE Trans Inf Theory. 2014; 60(6):3671–87.
Phatarfod R, Sudbury A. The use of a square array scheme in blood testing. Stat Med. 1994; 13(22):2337–43.
Lohse S, Pfuhl T, BerkóGöttel B, Rissland J, Geißler T, Gärtner B, Becker SL, Schneitler S, Smola S. Pooling of samples for testing for SARSCoV2 in asymptomatic people. Lancet Infect Dis. 2020. https://doi.org/10.1016/s14733099(20)303625.
Yelin I, Aharony N, ShaerTamar E, Argoetti A, Messer E, Berenbaum D, Shafran E, Kuzli A, Gandali N, Shkedi O, Hashimshony T, MandelGutfreund Y, Halberthal M, Geffen Y, SzwarcwortCohen M, Kishony R, et al. Evaluation of COVID19 RTqPCR test in multisample pools. Clin Infect Dis. 2020. https://doi.org/10.1093/cid/ciaa531.
Broughton JP, Deng X, Yu G, Fasching CL, Servellita V, Singh J, Miao X, Streithorst JA, Granados A, SotomayorGonzalez A, et al. CRISPR–Cas12based detection of SARSCoV2. Nat Biotechnol. 2020:1–5. https://doi.org/10.1038/s4158702005134.
Sarvotham S, Baron D, Baraniuk RG. Sudocodes fast measurement and reconstruction of sparse signals. In: 2006 IEEE International Symposium on Information Theory. Seattle: IEEE: 2006. p. 2804–8.
Acknowledgements
The author thanks Aycan Gundogdu, Ph.D., whose comments clearly helped in improving the overall quality of the manuscript.
Funding
Not applicable.
Author information
Authors and Affiliations
Contributions
OUN conceptualized and designed the study, developed the algorithms, ran the simulations, drafted and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
No ethical approval was required for this study, since it used only simulated data.
Consent for publication
Not applicable.
Competing interests
The author(s) declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Nalbantoglu, O. Group testing performance evaluation for SARSCoV2 massive scale screening and testing. BMC Med Res Methodol 20, 176 (2020). https://doi.org/10.1186/s12874020010481
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12874020010481
Keywords
 COVID19 testing
 Group testing
 Sparse recovery algorithms