We’re sorry, something doesn't seem to be working properly.

Please try refreshing the page. If that doesn't work, please contact us so we can address the problem.

# Latent class bivariate model for the meta-analysis of diagnostic test accuracy studies

- Paolo Eusebi
^{1, 2}Email author, - Johannes B Reitsma
^{3}and - Jeroen K Vermunt
^{4}

**14**:88

https://doi.org/10.1186/1471-2288-14-88

© Eusebi et al.; licensee BioMed Central Ltd. 2014

**Received: **18 December 2013

**Accepted: **8 May 2014

**Published: **11 July 2014

## Abstract

### Background

Several types of statistical methods are currently available for the meta-analysis of studies on diagnostic test accuracy. One of these methods is the Bivariate Model which involves a simultaneous analysis of the sensitivity and specificity from a set of studies. In this paper, we review the characteristics of the Bivariate Model and demonstrate how it can be extended with a discrete latent variable. The resulting clustering of studies yields additional insight into the accuracy of the test of interest.

### Methods

A Latent Class Bivariate Model is proposed. This model captures the between-study variability in sensitivity and specificity by assuming that studies belong to one of a small number of latent classes. This yields both an easier to interpret and a more precise description of the heterogeneity between studies. Latent classes may not only differ with respect to the average sensitivity and specificity, but also with respect to the correlation between sensitivity and specificity.

### Results

The Latent Class Bivariate Model identifies clusters of studies with their own estimates of sensitivity and specificity. Our simulation study demonstrated excellent parameter recovery and good performance of the model selection statistics typically used in latent class analysis. Application in a real data example on coronary artery disease showed that the inclusion of latent classes yields interesting additional information.

### Conclusions

Our proposed new meta-analysis method can lead to a better fit of the data set of interest, less biased estimates and more reliable confidence intervals for sensitivities and specificities. But even more important, it may serve as an exploratory tool for subsequent sub-group meta-analyses.

## Keywords

## Background

There is an increasing interest in meta-analyses of data from diagnostic accuracy studies [1–4]. Typically, the data from each of the primary studies are summarized in a 2-by-2 table cross-tabulating the dichotomized test result against the true disease status, from which familiar measures such as sensitivity and specificity can be derived [5]. Several statistical methods for meta-analysis of data from diagnostic test accuracy studies have been proposed [6–13]. Generally, we expect that such data show a negative correlation between sensitivity and specificity because of explicit or implicit variations in test-thresholds [1, 7], as well as contain a certain amount of heterogeneity [14]. Our research is motivated by the need to explore and explain sources of heterogeneity in systematic reviews of diagnostic tests in a more careful manner. In fact, methods for the meta-analysis of sensitivity and specificity are still an active field of research and debate. One frequently used method involves generating a ROC curve using simple linear regression [6, 7]. However, the assumptions of the underlying linear regression model are not always met, and as a consequence the produced statistics, in particular standard errors and p-values, may be invalid. There is also uncertainty as to the most appropriate weighting of studies to be used in the regression analysis [15]. Rutter and Gatsonis [9] suggested that in the presence of a substantial amount of heterogeneity, the results of meta-analyses should be presented as Summary ROC curves.

Reitsma and others [12] proposed the direct analysis of sensitivity and specificity estimates using a bivariate model BM, which yields a rigorous method for the meta-analysis of data on diagnostic test accuracy, in particular when the studies are selected based on a common threshold [16]. Chu and Cole [17] extended this bivariate normal model by describing the within-study variability with a binomial distribution rather than with a normal approximation of transformed observed sensitivities and specificities. Though the BM may work well with the normal approximation, Hamza [18] and others suggested that the binomial distribution is to be preferred, especially when only few studies with a small size are available. An additional advantage of the binomial approach is that it does not require a continuity correction. When using a bivariate generalized linear mixed model to jointly model the sensitivities and specificities, different monotone link functions can be implemented, such as logit, probit, and complementary log-log transformations [19]. Chu and others [20] also discussed a trivariate nonlinear random-effects model for jointly modeling disease prevalence, sensitivity and specificity, as well as an alternative parameterization for jointly modeling prevalence and predictive values.

Bayesian modeling approach to BM are gaining popularity, by allowing the structural distribution of the random effects to depend on multiple sources of variability and providing the predictive posterior distributions for sensitivity and specificity [21]. In order to avoid the Markov chain Monte Carlo sampling also a deterministic Bayesian approach using integrated nested Laplace approximations have been proposed [22]. BM can be seen within a unified framework which includes also the Hierarchical Summary ROC model [23]. Arends and others [24] showed that the bivariate random-effects approach not only extends the Summary ROC approach but also provides a unifying framework for other approaches. Rücker and Schumacher [13] proposed an alternative approach for defining a Summary ROC curve based on a weighted Youden index.

The Latent Class Model has been introduced in the literature as a tool for evaluating the accuracy of a new test when there is no gold standard against which to compare it [25, 26]. A probabilistic model is assumed for the relationship between the new diagnostic test, one or more imperfect reference tests, and the unobserved, or latent, disease status. This provides estimates of the sensitivity and specificity of the new diagnostic test. This application of latent class analysis has received considerable attention in the context of primary (individual) diagnostic accuracy studies, but its use in meta-analysis is rare [27].

In this paper we propose using a latent class approach for a different purpose. More specifically, we use it as a tool for clustering the studies involved in the meta-analysis [28]. For this purpose, the BM based on using a binomial distribution is expanded to include a discrete latent variable. The resulting Latent Class Bivariate Model (LCBM) allows obtaining more reliable estimates of sensitivity and specificity, as well as the estimation of a different between-study correlation between sensitivity and specificity per latent class. While this correlation is usually assumed to be negative, this is not always correct in real-data applications, probably due to varying accuracy levels and differences in test performance. In a simulation study, the LCBM is compared to the standard BM in terms of bias, power, and confidence. Moreover, it is applied to a well-known dataset on the diagnostic performance of multislice computed tomography and magnetic resonance imaging for the diagnosis of coronary artery disease [16, 29].

The remaining of the article is organized as follows. We first describe the BM and the LCBM, as well as discuss computational issues and the setup of the simulation study. Then, attention is paid to the results of the simulation study and the application of the LCBM to the coronary artery disease data. Next, we discuss the implications of our study and present some conclusions. Software code for estimating the LCBM is provided as Additional files.

## Methods

### The bivariate model

*x*

_{1i }and

*n*

_{1i }be the number of subjects with a positive test result and the total number of subjects with the disease in study

*i*, respectively, and

*x*

_{0i }and

*n*

_{0i }, analogously, be the number of subjects with a negative test result and the total number of subjects without the disease. Then, the observed sensitivity and specificity is

*x*

_{1i }/

*n*

_{1i }and

*x*

_{0i }/

*n*

_{0i }. The corresponding true values are denoted by

*η*

_{ i }and

*ξ*

_{ i }, respectively. The BM can be specified as follow [17]:

Here, **X**
_{
i
} and **W**
_{
i
} are (possibly overlapping) vectors of covariates related to sensitivity and specificity, ${\sigma}_{\eta}^{2}$ and ${\sigma}_{\xi}^{2}$ are the between-study variances of sensitivity and specificity, and *ρ* is their correlation. Thus, the parameters of the covariance matrix quantify the amount of heterogeneity present across studies together with how strongly the sensitivity and specificity of a study are related.

The hierarchical model specified by (1), (2) and (3) can be fitted using the generalized linear mixed model procedures in several standard statistical packages. Other kinds of link functions can be used instead of the logit one [19]. In simulation experiments, it has been shown that, in general, it is better to work with such a binomial model rather than with a normal model for transformed observed proportions [18].

### The latent class bivariate model

*K*latent classes [28]. This implies expanding the BM model with a discrete latent variable with categories denoted by

*c*. The basic structure of the resulting LCBM is:

*P*(

*c*) is the probability of a study to belong to latent class

*c*, and

*P*(

*x*

_{1i },

*x*

_{0i }|

*c*) is the joint sensitivity and specificity distribution within latent class

*c*. For

*P*(

*x*

_{1i },

*x*

_{0i }|

*c*) we specify a BM model with all parameters (including the variances and the correlation) varying across classes. More specifically, in the LCBM the binomial probabilities have the following form:

*η*

_{ i|c }and

*ξ*

_{ i|c }conditional on study

*i*belonging to class

*c*are assumed to have a bivariate normal distribution with parameters that vary across classes:

*ρ*

_{ c }is the class-specific between-study correlation. Finally, sensitivity and specificity are assumed to be mutually independent given the latent class memberships and the random effects. This yields the following expression for

*P*(

*x*

_{1i },

*x*

_{0i }|

*c*):

**Z**

_{ i }to predict the latent class membership in the LCBM, which yields:

Here, *P*(*c*|**Z**
_{
i
}) is the probability that a study belongs to latent class *c* given the covariate set **Z**
_{
i
} and where, as above, *P*(*x*
_{1i
},*x*
_{0i
}|*c*) denotes the class-specific sensitivity and specificity distribution. As can be seen, covariates **Z**
_{
i
} affect the latent classes but have no direct effects either on the true sensitivity and specificity or the random effects.

*P*(

*c*|

**Z**

_{ i }) is typically parameterized as follows:

Here *δ*
_{0} is an intercept term and *δ*
_{
p
} is the effect of covariate *p* on the class membership probability.

### Estimation techniques

Both the BM and LCBM can be estimated with the Latent GOLD 4.5 software package for latent class analysis [30]. To find the Maximum Likelihood estimates for the model parameters, a combination an EM and a Newton-Raphson algorithm is used; that is, the estimation process starts with a number of EM iterations and when close enough to the final solution, the program switches to Newton-Raphson. A well-known problem in LC analysis is the occurrence of local maxima. To prevent ending up with a local solution, multiple sets of starting values are used. The user can specify the number of start sets and the number of EM iterations to be performed per set.

### Simulation study

The performance of the LCBM was evaluated using a simulation study. More specifically, we evaluated Bias (difference between the mean estimate and the true value of the parameter for both LCBM and BM), Power (the proportion of replications in which the LCBM is preferred over the BM when data are generated from a true latent class structure), and Confidence (the proportion of replications in which the BM is preferred over the LCBM when data are generated from a single class structure). We were interested in the effects of the number of studies included in the meta-analysis, within-study sample size, and true mean sensitivity and specificity on the performance of the BM and the LCBM.

The settings used in our study were taken from the simulation study by Hamza et al. [18]. The disease prevalence was fixed to 50%. Four conditions were considered for the number of studies included in the meta-analysis; that is, 10, 25, 50, and 100 studies. The study size was generated from a normal distribution and rounded to the nearest integer. Two different distributions were considered, *N*(40,30^{2}) and *N*(500,450^{2}). The minimum study size was set to be 10, meaning that if the generated study size was less than 10, it was set to 10. Consequently, 40 and 500 are no longer the means for the simulated study sizes, but the medians, and the realized standard deviations will be slightly smaller than 30 and 450, respectively.

**The different scenarios used in the simulation study for assessing power**

Class1 | Class2 | |||
---|---|---|---|---|

Sensitivity | Specificity | Sensitivity | Specificity | |

1-8 | 90% | 75% | 75% | 75% |

9-16 | 90% | 75% | 90% | 90% |

17-24 | 90% | 75% | 75% | 90% |

25-32 | 90% | 60% | 60% | 60% |

33-40 | 90% | 60% | 90% | 90% |

41-48 | 90% | 60% | 60% | 90% |

**The different scenarios used in the simulation study for assessing confidence**

Sensitivity | Specificity | |
---|---|---|

1-8 | 90% | 60% |

9-16 | 90% | 75% |

17-24 | 90% | 90% |

Note that the criteria differ with respect to the weighting of parsimony in terms of number of parameters. Because the log of the sample size is usually larger than 3, BIC tends to select a model with fewer latent classes than AIC and AIC3. AIC has been shown to be a superior fit-index when dealing with complex models combined with small sample sizes [31].

## Results

### Simulation study

#### Bias

#### Power

#### Confidence

### Real data example: coronary artery disease data

We illustrate the use of the LCBM approach by re-analyzing data from Schuetz and others [29]. This well-known dataset is presented (Additional file 1) in the Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy [16]. In this meta-analysis, the diagnostic performances of multislice computed tomography (CT) and magnetic resonance imaging (MRI) for the diagnosis of coronary artery disease (CAD) are compared. Prospective studies that evaluated either CT or MRI (or both), used conventional coronary angiography (CAG) as the reference standard, and used the same threshold for clinically significant coronary artery stenosis (a diameter reduction of 50% or more) were included in the review. A total of 103 studies provided a 2-by-2 table for one or both tests and were included in the meta-analysis: 84 studies evaluated only CT, 14 evaluated only MRI, and 5 evaluated both CT and MRI. Because the studies were selected based on a common threshold for clinically significant coronary artery stenosis, BM was used for data synthesis and test comparison.

Using the Latent GOLD software version 4.5 [30], we estimated both the BM and LCBM with test type as a covariate (the software code is provided in the Additional file 2). For model selection, we used the AIC. The 2-class LCBM gave a lower AIC value (AIC = 961.9) than the standard BM (AIC = 963.7), which is in fact a 1-class LCBM.

**Coronary hearth disease data: point estimates and confidence intervals of sensitivity and specificity in BM and LCBM both for CT and MRI studies**

CT | MRI | ||
---|---|---|---|

BM | Sensitivity | 95.0% (94.0%-96.0%) | 86.2% (81.4%-91.0%) |

Specificity | 82.4% (80.4%-84.4%) | 71.0% (64.5%-77.6%) | |

LCBM | Sensitivity | 95.7% (94.6%-96.8%) | 86.9% (84.7%-89.1%) |

Specificity | 82.6% (80.1%-85.0%) | 69.5% (62.2%-76.7%) |

Looking at the model estimates (Table 3), we notice LCBM yields slightly different confidence intervals for sensitivity and specificity in MRI than BM.

## Discussion

In the simulation study we have seen that when sensitivity or specificity differs between latent classes, BM leads to biased estimates of sensitivity and specificity. In the real data example, we obtained slightly different confidence intervals for sensitivity and specificity in MRI with LCBM.

The disadvantage of using the LCBM is the considerable increase in the number of parameters to estimate compared to the BM, implying that the number of primary studies available may become an issue. As we can see from the results of the simulation study, even when there is a strong latent class structure, to obtain a reasonable power, the number of primary studies need to be about 25. While the AIC fit criterion has to be preferred in order to achieve a reasonable power, it can lead to false positive identification of latent classes. However, false positive identification of clusters by LCBM will not end-up in biased estimates but in inflated standard errors.

In the reported simulation study, we fixed the disease prevalence to 50%. However, prevalence can take on quite different values in diagnostic accuracy studies [32]. We also simulated scenarios with a lower disease prevalence, which had little impact on the results (data not shown).

Meta-analysis of diagnostic studies with bivariate mixed-effects models can sometimes end up with non-convergence. In the LCBM, an additional discrete latent variable is added to the standard BM, which increases number of parameters and makes the computational aspects even more challenging. However, the implementation of BM and LCBM in the Latent GOLD computer program turned out to be very stable.

LCBM could be programmed in R or implemented in a SAS macro by applying PROC NLMIXED. However, the computational approach used in Latent GOLD is very stable and offers good performance in terms finding the maximum likelihood solution and convergence. Additional advantages of the Latent GOLD implementation include that it allows expanding the model in various possible ways and that several useful outputs are readily available, which would be quite complicated to program from scratch in other languages.

## Conclusions

The proposed LCBM framework provides us with a tool for assessing whether the study heterogeneity can be explained in a more careful way. It yields a clustering of studies in diagnostic test accuracy reviews that can be used for explanatory purposes.

In the real data example, we saw how LCBM can improve the understanding of the relationship between sensitivity and specificity. The LCBM is able to identify subgroups of studies that are separated in ROC space. What is added by the LCBM framework is that it provides an explanatory and confirmatory tool for investigating and testing different patterns of performance across studies. In particular, in the real data example, we tested the equivalence in diagnostic performance between CT and MRI. Moving from BM to LCBM we obtained a clear picture of two clusters of studies, and obtained more reliable confidence intervals for the sensitivity and specificity.

We can conclude that the LCBM yields a statistically rigorous, flexible, and data-driven approach to meta-analysis. LCBM generates useful results when subgroups of studies can be related to meaningful design or clinical characteristics and it provides us with a model-based starting point for subgroup meta-analyses.

Future work will include the implementation of the LCBM in an R package, the Bayesian estimation and testing of the model, and the investigation of its application in Hierarchical Summary ROC.

## Declarations

### Acknowledgements

The work was funded by Regional Health Authority of Umbria. PE thanks Kathryn Mary Mahan, Iosief Abraha, Alessandro Montedori and Gianni Giovannini for their support.

## Authors’ Affiliations

## References

- Deeks JJ: Systematic reviews in health care: systematic reviews of evaluations of diagnostic and screening tests. BMJ. 2001, 323 (7305): 157-162.View ArticlePubMedPubMed CentralGoogle Scholar
- Bossuyt PM, Bruns DE, Reitsma JB, Gatsonis CA, Glasziou PP, Irwig LM, Lijmer JG, Moher D, Rennie D, De Vet HCW: Towards complete and accurate reporting of studies of diagnostic accuracy: the stard initiative. BMJ. 2003, 326: 41-44.View ArticlePubMedPubMed CentralGoogle Scholar
- Tatsioni A, Zarin DA, Aronson N, Samson DJ, Flamm CR, Schmid C, Lau J: Challenges in systematic reviews of diagnostic technologies. Ann Intern Med. 2005, 142 (12): 1048-1055.View ArticlePubMedGoogle Scholar
- Gluud C, Gluud LL: Evidence based diagnostics. BMJ. 2005, 330 (7493): 724-726.View ArticlePubMedPubMed CentralGoogle Scholar
- Eusebi P: Diagnostic accuracy measures. Cerebrovasc Dis. 2013, 36 (4): 267-272.View ArticlePubMedGoogle Scholar
- Littenberg B, Moses LE: Estimating diagnostic accuracy from multiple conflicting reports: a new meta-analytic method. Med Decis Making. 1993, 13 (4): 313-321.View ArticlePubMedGoogle Scholar
- Moses LE, Shapiro D, Littenberg B: Combining independent studies of a diagnostic test into a summary roc curve: data-analytic approaches and some additional considerations. Stat Med. 1993, 12 (14): 1293-1316.View ArticlePubMedGoogle Scholar
- Van Houwelingen HC, Zwinderman KH, Stijnen T: A bivariate approach to meta-analysis. Stat Med. 1993, 12 (24): 2273-2284.View ArticlePubMedGoogle Scholar
- Rutter CM, Gatsonis CA: A hierarchical regression approach to meta-analysis of diagnostic test accuracy evaluations. Stat Med. 2001, 20 (19): 2865-2884.View ArticlePubMedGoogle Scholar
- Dukic V, Gatsonis CA: Meta-analysis of diagnostic test accuracy assessment studies with varying number of thresholdss. Biometrics. 2003, 59 (4): 936-946.View ArticlePubMedGoogle Scholar
- Siadaty M, Shu J: Proportional odds ratio model for comparison of diagnostic tests in meta-analysis. BMC Med Res Methodol. 2004, 4 (1): 27-View ArticlePubMedPubMed CentralGoogle Scholar
- Reitsma JB, Glas AS, Rutjes AWS, Scholten RJ, Bossuyt PM, Zwinderman AH: Bivariate analysis of sensitivity and specificity produces informative summary measures in diagnostic reviews. J Clin Epidemiol. 2005, 58 (10): 982-990.View ArticlePubMedGoogle Scholar
- Rücker G, Schumacher M: Summary roc curve based on the weighted youden index for selecting an optimal cutpoint in meta-analysis of diagnostic accuracy. Stat Med. 2010, 29 (30): 3069-3078.View ArticlePubMedGoogle Scholar
- Lijmer JG, Bossuyt PM, Heisterkamp SH: Exploring sources of heterogeneity in systematic reviews of diagnostic tests. Stat Med. 2002, 21 (11): 1525-1537.View ArticlePubMedGoogle Scholar
- Walter SD: Properties of the summary receiver operating characteristic (sroc) curve for diagnostic test data. Stat Med. 2002, 21 (9): 1237-1256.View ArticlePubMedGoogle Scholar
- Macaskill P, Gatsonis C, Deeks J, Harbord R, Takwoingi Y: Chapter: Analysing and presenting results. Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy. Edited by: Deeks J, Bossuyt P, Gatsonis C. 2010, New York: The Cochrane Collaboration, 1-61.Google Scholar
- Chu H, Cole S: Bivariate meta-analysis of sensitivity and specificity with sparse data: a generalized linear mixed model approach. J Clin Epidemiol. 2006, 59 (12): 1331-1332.View ArticlePubMedGoogle Scholar
- Hamza TH, van Houwelingen HC, Stijnen T: Random effects meta analysis of proportions: The binomial distribution should be used to model the within study variability. J Clin Epidemiol. 2008, 61 (1): 41-51.View ArticlePubMedGoogle Scholar
- Chu H, Nie L, Cole S, Poole C: Bivariate random effects meta-analysis of diagnostic studies using generalized linear mixed models. Med Decis Making. 2010, 30 (4): 499-508.View ArticlePubMedGoogle Scholar
- Chu H, Nie L, Cole S, Poole C: Meta-analysis of diagnostic accuracy studies accounting for disease prevalence: Alternative parameterizations and model selection. Stat Med. 2009, 28 (18): 2384-2399.View ArticlePubMedGoogle Scholar
- Verde PE: Meta-analysis of diagnostic test data: a bivariate bayesian modeling approach. Stat Med. 2010, 29 (30): 3088-3102.View ArticlePubMedGoogle Scholar
- Paul M, Riebler A, Bachmann LM, Rue H, Held L: Bayesian bivariate meta-analysis of diagnostic test studies using integrated nested laplace approximations. Stat Med. 2010, 29 (12): 1325-1339.View ArticlePubMedGoogle Scholar
- Harbord RM, Deeks JJ, Egger M, Whiting P, Sterne JAC: A unification of models for meta-analysis of diagnostic accuracy studies biostatistics. Biostatistics. 2007, 8 (2): 239-251.View ArticlePubMedGoogle Scholar
- Arends LR, Hamza TH, van Houwelingen JC, Heijenbrok-Kal MH, Hunink MG, Stijnen T: Bivariate random effects meta-analysis of roc curves. Med Decis Making. 2008, 28 (5): 621-638.View ArticlePubMedGoogle Scholar
- Goetghebeur E, Liinev J, Boelaert M, Van der Stuyft P: Diagnostic test analyses in search of their gold standard: latent class analyses with random effects. Stat Methods Med Res. 2000, 9 (3): 231-248.View ArticlePubMedGoogle Scholar
- Pepe P, Janes H: Insights into latent class analysis of diagnostic test performance. Biostatistics. 2007, 8 (2): 474-484.View ArticlePubMedGoogle Scholar
- van Smeden M, Naaktgeboren CA, Reitsma JB, Moons KG: Latent class models in diagnostic studies when there is no reference standard - a systematic review. Am J Epidemiol. 2014, 179 (4): 423-431.View ArticlePubMedGoogle Scholar
- Goodman LA: Exploratory latent structure analysis using both identifiable and unidentifiable models. Biometrika. 1998, 61 (2): 215-231.View ArticleGoogle Scholar
- Schuetz GM, Zacharopoulou NM, Schlattmann P, Dewey M: Meta-analysis: noninvasive coronary angiography using computed tomography versus magnetic resonance imaging. Ann Intern Med. 2010, 152 (3): 167-177.View ArticlePubMedGoogle Scholar
- Vermunt JK, Magidson J: LG-Syntax user’s guide: Manual for Latent GOLD 4.5 Syntax module. Technical Report. 2008, Belmont, MA: Statistical InnovationsGoogle Scholar
- Lin TH, Dayton CM: Model selection information criteria for non-nested latent class models. J Educ Behav Stat. 1997, 22 (3): 249-264.View ArticleGoogle Scholar
- Leeflang MM, Rutjes AW, Reitsma JB, Hooft L, Bossuyt PM: Variation of a test’s sensitivity and specificity with disease prevalence. CMAJ. 2013, 185 (11): 537-544.View ArticleGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2288/14/88/prepub

### Pre-publication history

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

## Comments

View archived comments (1)