Abstract
It is difficult to estimate sensitivity and specificity of diagnostic tests when there is no gold standard. Latent class models have been proposed as a potential solution as they provide estimates without the need for a gold standard. Using a motivating example of the evaluation of point of care tests for leptospirosis in Tanzania, we show how a realistic violation of assumptions underpinning the latent class model can lead directly to substantial bias in the estimates of the parameters of interest. In particular, we consider the robustness of estimates of sensitivity, specificity, and prevalence, to the presence of additional latent states when fitting a two-state latent class model. The violation is minor in the sense that it cannot be routinely detected with goodness-of-fit procedures, but is major with regard to the resulting bias.
Keywords: Bayes, leptospirosis, model sensitivity, sensitivity, specificity
1 |. INTRODUCTION
Diagnostic tests are a critical component of effective clinical management and disease surveillance. Useful tests need to be simple, affordable and timely, and must accurately discriminate individuals with and without the disease in question. For many infectious diseases, the standard “best” available tests are either expensive, require specialist expertise, or involve considerable delay, for example, tests based on culture or requiring acute and convalescent samples.1 Recent advances in biology and technology have lead to a proliferation of new, rapid, point of care tests for infectious diseases. While these novel tests are often simple, affordable and timely, their diagnostic accuracy requires evaluation. Traditional approaches have estimated the sensitivity and specificity of novel assays by comparing them to a reference standard. For many diseases, however, the reference test has imperfect sensitivity or specificity. This has, in turn, hampered evaluation of the novel tests.
Two main approaches are used to overcome the deficiencies of an imperfect reference test: (i) defining a composite reference standard (CRS)2 and (ii) use of a latent class model (LCM).3,4 With the CRS approach, a combination of tests are used to develop rules for the determination of presence or absence of disease; a new test is then compared against this reference standard. Finding an agreed CRS that leads to minimal bias is difficult, if not impossible.
Latent class models have been widely used in many areas of research. Essentially, the models specify a probabilistic relationship between a set of measured variables and a hypothesized unmeasured latent class variable. In diagnostic testing, the measured variables are the outcomes from two or more diagnostic tests. The simplest LCM used is a two class model, with the two classes treated as “disease” and “disease free,” with the various tests assumed to be independent conditional on the disease status. Model parameters are interpreted as sensitivity, specificity, and disease prevalence. For simplicity, we use these labels for the remainder of the manuscript, while recognizing that use of these labels without a gold standard reference test can be problematic.5
In the last decade, there has been steady growth in the use of latent class analysis to evaluate new diagnostic tests, particularly in the area of infectious diseases.4 Three concerns with LCMs have been outlined in the statistics literature. First, the LCM is geometrically complex, and identifiability is not guaranteed even when the degrees of freedom are greater than the number of parameters.6,7 Second, parameter estimates can be biased if the dependence relationship between tests is misspecified.2,8 It is possible to have two or more dependence models that fit the data equally as well but lead to different conclusions regarding sensitivity and specificity.8 The importance of the assumed dependence relationship was highlighted by Pepe and Janes who provided explicit maximum likelihood estimators for the case of three conditionally independent tests with two classes.9 They showed that the estimated sensitivities and specificities rely on observed associations between the tests. The third concern is the validity of the interpretation of the latent classes. The classes are inferred in order to provide an optimal fit to the data. There is no reason why these inferred classes need to represent disease status, or in fact any realistic variable.2 In discussing latent variable models Agresti,10 pg 544, states
A danger with latent variable models … is the temptation to interpret latent variables too literally… One should realize the tentative nature of the latent variable. Be careful not to make the error of reification – treating an abstract construction as if it has actual existance.
With this in mind, the remainder of the manuscript explores the robustness of LCMs to a particular violation of assumptions. This was motivated by a study of leptospirosis in northern Tanzania in which 225 patients presenting to hospital with fever were tested for leptospirosis using a standard test and three new point of care tests.11 We fit a Bayesian two-state latent class model to explore the properties of the new tests; the results obtained did not conform with our understanding about the scientific basis of the tests in question. We present full details of the model in Section 2 and the motivating example in Section 2.3. It is possible that our expectations were incorrect and the latent class model has provided insight into the performance of these tests that would be otherwise hidden. Another possibility is that unidentified model misspecification has led to substantial bias in the estimates of sensitivity and specificity. We explore one such model misspecification in Section 3 that even for the simplest LCM can lead to substantial bias and is difficult to detect in practice. The remainder of the manuscript explores the implications of this form of misspecification.
2 |. BACKGROUND
2.1 |. Data description
The data are denoted by the N × T matrix y, where N is the number of individuals in the study and T is the number of tests. The entry yij denotes the result of the jth test for the ith individual. The value yij = 1 indicates a positive test, with yij = 0 a negative test. These data can be summarized by counting the number of times each possible combination of test results was seen. The S = 2T possible combinations of test results are specified in the S × T matrix x, with the number of times each combination of test results was seen recorded in the S-vector n. The ith entry of n is the number of individuals who have the combination of test results given in the ith row of x. An example of y, x and n for a hypothetical T = 2 study with N = 6 is
The data (x and n) for our motivating example is shown in Table 1, where T = 4 and N = 225.
TABLE 1.
MAT | Leptorapide | Leptocheck | Test-It | Count |
---|---|---|---|---|
1 | 1 | 1 | 1 | 3 |
1 | 1 | 1 | 0 | 0 |
1 | 1 | 0 | 1 | 0 |
1 | 1 | 0 | 0 | 1 |
1 | 0 | 1 | 1 | 3 |
1 | 0 | 1 | 0 | 4 |
1 | 0 | 0 | 1 | 0 |
1 | 0 | 0 | 0 | 12 |
0 | 1 | 1 | 1 | 6 |
0 | 1 | 1 | 0 | 3 |
0 | 1 | 0 | 1 | 0 |
0 | 1 | 0 | 0 | 25 |
0 | 0 | 1 | 1 | 11 |
0 | 0 | 1 | 0 | 17 |
0 | 0 | 0 | 1 | 9 |
0 | 0 | 0 | 0 | 131 |
Note: The indicator 1 represents that the patient tested positive for the given test and 0 that the patient tested negative. The counts give the number of participants with each combination of test results.
2.2 |. Latent class model
We outline an M-state LCM where all tests are assumed independent. The LCM consists of two components. The first allocates a class (or state) c to each individual,
(1) |
where πj gives the probability of an individual being in class j, iid denotes an independent and identically distributed random variable and Cat(b1, … , bp) represents a categorical distribution with support 1, … , p and parameters b1, … , bp. Throughout, we use the terms class and state interchangeably.
The second model component describes the distribution of the observed data conditional on the class
(2) |
where pjc gives the probability that test j is positive for an individual in class c, ind denotes an independent random variable and Bern(b) represents a Bernoulli distribution with probability b.
When fitting in a Bayesian context, model specification is completed with prior distributions on π = (π1, … , πM) and p = (P11, … , PTM).
We can use (1) and (2) as a complete data model that requires the inclusion of the latent variable c = (c1, … , cN) directly in the model. Despite being intuitive, this approach is slow to fit when using Markov chain Monte Carlo (MCMC) as each entry in c needs to be updated. A more computationally efficient strategy is to explicitly sum over c. The resulting marginal model is
(3) |
where MN(a, b) represents a multinomial distribution with index a and vector of probabilities b, and q = (q1, … , qs), with
(4) |
2.3 |. Motivating example: Leptospirosis in Tanzania
Leptospiriosis is a zoonotic disease, and is transmitted from animals to humans through contact with infected animal urine, blood or other bodily fluids, or through contaminated water or soil. Leptospirosis is a leading cause of morbidity and mortality among zoonotic diseases, particularly in resource-poor countries.12 Humans present with symptoms common to many infectious diseases, so laboratory testing is required for diagnosis.13
The best available diagnostic test for leptospirosis is a serological test, the microscopic agglutination test (MAT). In the MAT live antigens representing different serovars are reacted with patient serum samples to measure immunoglobulin M (IgM) and immunoglobulin G (IgG) antibody levels. Because antibodies can remain detectable for months after an acute infection, a second MAT test identifying either seroconversion or a rise in antibody titres is required to determine current infection. The specificity of the MAT is thought to be near perfect, but sensitivity has been estimated as approximately 0.8 with at least 11 days after the onset of illness among patients with positive blood cultures.14 The diminished sensitivity is thought due to an incomplete panel of serovars, a subjective degree of agglutination used as the cut-off for interpretation, and some patients failing to mount an antibody response to Leptospira infection.
As outcome following leptospirosis infection is improved by early treatment with antibiotics,13 there is a need for a diagnostic test which provides a quick, accurate result early in the illness. A number of point-of-care tests for acute leptospirosis disease have been developed which aim to detect anti-Leptospira IgM. The three tests carried out in our study were Test-It Leptospira IgM Lateral Flow Assay (Lifeassay Diagnostics (Pty) Ltd. Cape Town, South Africa), Leptocheck-WB (Zephyr Biomedicals, Goa, India) and Leptorapide (Linnodee Ltd., Doagh, Northern Ireland). Test-It Leptospira and Leptocheck WB are similar in that they both use lateral flow technology and antigen from whole-cell boiled Leptospira biflexa serovar Patoc. In our systematic review of lateral flow assays we found sensitivity ranged from 53% to 95%, but noted the variation may reflect different reference tests or different stages of the illness.15 As all assays detect antibodies, which take between one and four weeks after illness onset to develop, lateral flow assays performed on serum taken from early in the illness were expected to have lower sensitivity than MAT performed on serum from 4-6 weeks after the onset of illness.14
Participants were enrolled from August 2007 through September 2008 and from February 2012 through May 2014 as part of febrile illness surveillance studies at Kilimanjaro Christian Medical Centre (KCMC) and Mawenzi Regional Referral Hospital (MRRH) in Moshi, Tanzania. Blood samples were taken on the date of enrolment, and participants were requested to return 4-6 weeks later for a convalescent sample. The MAT test returned a positive leptospirosis diagnosis if a patient returned a single reciprocal titer ≥800, or a ≥4 fold rise in antibody titer between enrolment and convalescent samples.11,16 The analysis presented here included all patients enrolled who had results for all 4 tests (n = 225).
Ethical approval for the study was granted by the Kilimanjaro Christian Medical Centre Research Ethics Committee (# 295), the Tanzania National Institutes for Medical Research National Ethics Coordinating Committee (NIMR1HQ/R.8cNo1. 11/283), Duke UniversityMedical Center Institutional Review Board (IRB# Pro00016134), and the University of Otago Human Ethics Committee (Health) (H15/055).
The data are presented in Table 1. We fit a two-state Bayesian LCM as outlined in (3) and (4) in JAGS.17 We ran three Markov chains for 10 000 iterations following a burn-in/adaptive phase of 5000 iterations.
We assumed uniform prior distributions for π and p. We used the algorithm of Stephens18,19 to correct for label switching (see the supplementary materials for more information). We take the inferred state that has the lowest prevalence to be state 1, and refer to this as the “disease” group. The sensitivities are given by p11, … , pT1 and the specificities are given by 1 – p12, … ,1 – pT2.
Posterior summaries of the sensitivities and specificities are plotted in Figure 1. The posterior median for π1, the prevalence of disease, was 0.14. The results obtained did not conform to our expectations regarding the properties of the tests (Figure 1). In particular, the estimated sensitivity for Test-It Leptospira and Leptocheck were both considerably higher than that for the MAT (posterior medians 0.77 and 0.93 versus 0.27). If we extend the model in (2) to allow two-way dependence between Test-It Leptospira and Leptocheck WB, the estimate of the MAT test sensitivity remains lower than that for Test-It Leptospira and Leptocheck WB, however, there is increased posterior uncertainty. Details of the dependence model implemented are provided in the Supplementary Materials.
To assess goodness-of-fit in the latent class model we use posterior predictive assessment.20 We graphically compare the posterior predictive distribution for each of the S combinations of test results with the observed counts (Figure 2). In the Supplementary Materials, we also consider posterior predictive checks of the pairwise counts in the spirit of Qu et al.21 In both cases, the observed counts are consistent with draws from the posterior predictive distribution indicating that there is no evidence of substantive lack-of-fit in the model. If it were not for the results being the opposite of our expectations, there would be little cause for concern.
Our results are more extreme than those of Limmathurotsakul et al22 who estimated a sensitivity for the MAT test of approximately 50%, although we note that they assessed different tests and implement LCMs that differ from those we have used. The key finding of their article was that the true sensitivity of the MAT is low (pg. 328), making the MAT inappropriate for use as a gold standard. In our case, the estimated sensitivity is so low that we believe it to be implausible. Instead, we believe the results are due to model misspecification that is difficult to detect with standard goodness-of-fit assessment, yet has led to considerable bias in the estimates of interest.
3 |. MODEL VIOLATION
To assess the robustness of the LCM to underlying assumptions, we consider the possibility that there is a third state in the LCM. We replicate the motivating example with four tests (test 1 through test 4) and consider two situations.
In the first situation, the data generating process is a two-state LCM (M = 2). The two states correspond to diseased and non-diseased individuals. This provides a baseline where the generating and fitting model are the same.
In the second situation, we allow the data generating process to be a three-state LCM (M = 3). This provides a scenario where the data generating process is no longer the same as the fitting model. There are numerous possible definitions that we could assume for the three states. One such possibility that we refer to as the “alternate disease” definition is where we take the first two states to be the disease of interest and non-diseased individuals as in the M = 2 model. We allow the third state to represent individuals that have an alternate disease that can trigger a test response at rates in excess of those with no disease. Another possibility that we refer to as the “mixed disease” definition is where state 2 refers to non-diseased individuals, and states 1 and 3 reflect different biological aspects of the disease of interest.
Either of these situations are biologically plausible for our motivating leptospirosis example. One example where the alternate disease situation would be appropriate is if there is moderate prevalence of other diseases that can trigger an IgM response. IgM is a relatively non-specific immunoglobulin that is produced rapidly in response to illness. While its rapid response can be leveraged for testing purposes, the downside is that we may detect diseases other than that we are targeting. One example where the mixed disease situation would be appropriate is where tests detect one biological process (eg, IgG response) at a different rate from another process (eg, IgM response). We restrict our attention to three-state models, but note that the actual process is likely to be more complex.
The input values for the simulation were determined so that the expected counts for M = 2 and M = 3 are (i) similar to each other, and (ii) similar to the observed leptospirosis data (Tables 2 and 3).
TABLE 2.
Parameter | M = 2 | M = 3 |
---|---|---|
p 11 | 0.27 | 0.88 |
p 21 | 0.33 | 0.10 |
p 31 | 0.94 | 0.22 |
p 41 | 0.67 | 0.05 |
p 12 | 0.08 | 0.01 |
p 22 | 0.14 | 0.15 |
p 32 | 0.08 | 0.10 |
p 42 | 0.05 | 0.05 |
p 13 | — | 0.25 |
p 23 | — | 0.36 |
p 33 | — | 0.93 |
p 43 | — | 0.81 |
π 1 | 0.15 | 0.08 |
π 2 | 0.85 | 0.80 |
π 3 | — | 0.12 |
Note: The first two classes represent the disease of interest and no disease, respectively. The third class represents individuals with an alternate disease that triggers a response in some or all of the tests. The value pjc represents the probability of testing positive for test j when in class c and πc is the probability of being in class c.
TABLE 3.
Test 1 | Test 2 | Test 3 | Test 4 | Expected count |
|
---|---|---|---|---|---|
M = 2 | M = 3 | ||||
1 | 1 | 1 | 1 | 1.9 | 1.8 |
1 | 1 | 1 | 0 | 1.1 | 0.8 |
1 | 1 | 0 | 1 | 0.2 | 0.2 |
1 | 1 | 0 | 0 | 1.9 | 1.4 |
1 | 0 | 1 | 1 | 3.9 | 3.4 |
1 | 0 | 1 | 0 | 2.9 | 3.9 |
1 | 0 | 0 | 1 | 0.9 | 0.9 |
1 | 0 | 0 | 0 | 11.6 | 11.9 |
0 | 1 | 1 | 1 | 5.2 | 5.6 |
0 | 1 | 1 | 0 | 4.4 | 3.9 |
0 | 1 | 0 | 1 | 1.5 | 1.6 |
0 | 1 | 0 | 0 | 21.7 | 23.1 |
0 | 0 | 1 | 1 | 11.0 | 10.5 |
0 | 0 | 1 | 0 | 16.6 | 17.1 |
0 | 0 | 0 | 1 | 7.6 | 7.6 |
0 | 0 | 0 | 0 | 132.6 | 131.1 |
Note: The expected counts are similar between the two-state (M = 2) and three-state (M = 3) models.
The true sensitivities for the two-state model are given by p11, … ,p41 in Table 2, while the true specificities are given by 1 – p12, … , 1 – p42. The sensitivities and specificities for the three-state model depend on the definition that we have made regarding the three states. Initially we interpret the results assuming the “alternate disease” definition. In this situation, group 1 refers to disease and the true sensitivities are given by p11, … ,p41 in Table 2. As there are two groups that do not have the disease of interest, we combine the two classes “no disease” and “other disease” so that the true specificity for test i is
where pij and πj are given in Table 2. We consider the interpretation of the results using the “mixed disease” definition in Section 3.3.
3.1 |. Goodness-of-fit assessment
To assess model fit, we use posterior predictive model assessment with a Pearson-type test statistic
where nk is the observed count for combination of test results k, and Nqk is the expected count for combination of test results k, with qk given in (4).
In iteration t, we use the posterior sample (p(t), π(t)), t = 1, … , l to generate replicate data ~ MN(N, q(t)), where q(t) = (q(t)1, … ,q(t)S). The Bayesian p-value, denoted PB is proportion of posterior samples in which the test statistic evaluated using the replicated data T(, p(t), π(t)) exceeds the test statistic calculated using the observed data T(n, p(t), π(t)).
The use of Bayesian P-values for model assessment is often conservative.23 Under the true model, the distribution of PB is underdispersed relative to a uniform distribution for many choices of test statistic T. To overcome this, we use a calibrated Bayesian P-value,24 denoted . The calibrated value adjusts PB according to an estimate of its empirical distribution when the model is correct. Conveniently, we can estimate this distribution using values of PB from our baseline simulations with M = 2 as we are simulating and fitting under the true model. For a given simulation with Bayesian p-value of PB = X, we calculate the value as the proportion of PB values from the baseline simulation that were smaller than X.
3.2 |. Simulation results
We simulated 1000 datasets under each the following four scenarios:
True model has two-states; sample size is N = 225.
True model has two-states; sample size is N = 1000.
True model has three-states; sample size is N = 225.
True model has three-states; sample size is N = 1000.
The sample size of N = 225 corresponds to that observed in the motivating example. A sample size of N = 1000 allows us to assess the impact of a larger sample size on the properties of the model and estimation.
For each scenario, we fit a two-state latent class model in JAGS with three parallel Markov chains. The starting values for each of the chains are drawn at random from the prior distribution. In each chain we discard 5000 iterations for burn-in, before obtaining a posterior sample of 10 000 iterations, that is thinned by 10. This gives a posterior sample of 3000 across all chains. We then run the algorithm of Stephens18 to resolve label switching. The use of thinning is to mitigate the computational demands of the label switching algorithm in a simulation setting. We assess convergence using the Ȓ statistic.25 In each simulation we find the maximum Ȓ value across all parameters in the model. The maximum 1.02 in more than 99% of the simulations across all four scenarios (3984 simulations of 4000). We discard all simulations where 1.02.
For each simulated dataset, we obtain point and interval estimates for the sensitivity and specificity of each test in terms of the posterior median, and central 90% credible interval, respectively. We also find the values PB and to assess lack-of-fit.
For each simulation scenario we summarize the estimation of sensitivity and specifity for each of the four tests. In particular, we find (i) the bias of the point estimates (Figure 3 and Table 4); (ii) the coverage of the interval estimates (Table 4); and (iii) the proportion of models that were assessed as ill-fitting (Table 5).
TABLE 4.
M = 2 |
M = 3 |
|||||||
---|---|---|---|---|---|---|---|---|
Bias |
Coverage |
Bias |
Coverage |
|||||
N = 225 | N = 1000 | N = 225 | N = 1000 | N = 225 | N = 1000 | N = 225 | N = 1000 | |
sens1 | 0.005 | −0.004 | 0.92 | 0.90 | −0.613 | −0.617 | 0.00 | 0.00 |
sens2 | 0.007 | −0.002 | 0.92 | 0.92 | 0.235 | 0.227 | 0.06 | 0.00 |
sens3 | −0.092 | −0.041 | 0.94 | 0.92 | 0.623 | 0.677 | 0.00 | 0.00 |
sens4 | 0.011 | 0.006 | 0.98 | 0.91 | 0.628 | 0.629 | 0.00 | 0.00 |
spec1 | −0.001 | −0.000 | 0.89 | 0.90 | −0.042 | −0.040 | 0.31 | 0.01 |
spec2 | 0.000 | −0.000 | 0.90 | 0.89 | 0.034 | 0.033 | 0.68 | 0.17 |
spec3 | −0.007 | 0.000 | 0.97 | 0.92 | 0.118 | 0.123 | 0.03 | 0.00 |
spec4 | 0.005 | 0.005 | 0.95 | 0.91 | 0.106 | 0.103 | 0.00 | 0.00 |
π 1 | 0.012 | 0.009 | 0.97 | 0.92 | 0.082 | 0.076 | 0.26 | 0.01 |
Note: The value of M denotes the number of states in the true model. The labels sensi and speci represent the sensitivity and specificity of test i, respectively. The posterior median was used for estimation and the coverage is based off a 90% central credible interval.
TABLE 5.
M | N | Pr(PB < 0.05) | Pr |
---|---|---|---|
2 | 225 | 0.00 | 0.05 |
2 | 1000 | 0.00 | 0.06 |
3 | 225 | 0.00 | 0.07 |
3 | 1000 | 0.04 | 0.23 |
Note: The true model is indexed based on the number of latent states (M) and the sample size is given by N.
The estimates are close to unbiased and coverage near nominal when data are simulated and fitted under a two-state latent class model (Figure 3 and Table 4). The properties of the model are as expected when a two-state LCM is the true data generating process.
In contrast, many of the sensitivities and specificities are biased with very low coverage when data are simulated under the three-state model (Figure 3 and Table 4). The extent of the bias is concerning. In the three state model, test 1 has a true sensitivity of 0.88. The bias (when N = 1000) is −0.62 completely changing our interpretation. What is a highly sensitive test is estimated as only correctly identifying around one in four individuals who truly have the disease. Things are equally as worrying regarding the effect on the estimated sensitivity of test 3. What is in truth a poor test is estimated as having high sensitivity due to a bias of 0.68 (N = 1000).
It is common for statistical models to be biased when assumptions are violated. This need not be a concern if we can identify the model violation and adjust our model and/or interpretation accordingly. The model violation we implement cannot be easily identified with standard goodness-of-fit assessment (Table 5). When the sample size is N = 225 we identify lack-of-fit at the nominal (α = 0.05) rate. Even when the sample size is N = 1000 individuals, we only identify lack-of-fit in little more than 1 in 5 of the cases.
We note that similar results (and identical conclusions) are obtained if we fit the models using maximum likelihood instead of Bayesian inference. We have opted for Bayesian inference here for two reasons. The first is that Bayesian inference appears to be the preferred method of inference for those applying LCMs to diagnostic testing.22,26 The second is that using Bayesian inference avoids the need to rely on asymptotic approximations and combining cells when performing goodness-of-fit testing.
3.3 |. Interpretation of latent variable
The results above are interpreted assuming the “alternate disease” definition where state 1 refers to disease, and states 2 and 3 refer to no disease. If instead, we consider the “mixed disease” definition, then states 1 and 3 refer to disease and the true sensitivity of test i becomes
while the specificities are given by 1 – p12, … ,1 – p42.
We can consider this definition using the same simulation results as above; only the definition of “disease” differs when assessing bias and coverage. For this definition, the bias of the sensitivities and specificities are non-negligible and the coverage rates poor (Figure 4 and Table 6). When N = 1000, none of the coverage rates for sensitivity are over 55% and for two tests the coverage is no more than 2%. The inferred latent variable does not appear to correspond to either of the two definitions that we have considered (“alternate disease” or “mixed disease”).
TABLE 6.
Bias |
Coverage |
|||||
---|---|---|---|---|---|---|
N = 225 | N = 1000 | N = 225 | N = 1000 | |||
sens1 | −0.235 | −0.239 | 0.21 | 0.00 | ||
sens2 | 0.079 | 0.071 | 0.77 | 0.54 | ||
sens3 | 0.197 | 0.251 | 0.53 | 0.02 | ||
sens4 | 0.172 | 0.173 | 0.74 | 0.41 | ||
spec1 | −0.073 | −0.071 | 0.01 | 0.00 | ||
spec2 | 0.006 | 0.006 | 0.89 | 0.87 | ||
spec3 | 0.010 | 0.015 | 0.94 | 0.84 | ||
spec4 | 0.007 | 0.004 | 0.95 | 0.92 | ||
π 1 | −0.038 | −0.044 | 0.91 | 0.61 |
Note: The labels sensi and speci represent the sensitivity and specificity of test i, respectively. The posterior median was used for estimation and the coverage is based off a 90% central credible interval.
3.4 |. Further simulations
In the Supplementary Materials, we present a second simulation study with different input values. The broad conclusions are the same as the simulation above, that is, many of the estimates are biased, coverage is poor, and we are unable to detect lack-of-fit in the majority of simulations.
4 |. DISCUSSION
Our work was motivated by a real example where the tests detect disease at various times through different biological characteristics. For the best available diagnostic test (MAT), a two state LCM gave estimated sensitivity well below what was expected. Our simulations show that one possible explanation is that there was undetected misspecification in the two-state LCM. We considered two definitions for a three state model (“alternate disease” and “mixed disease”). The inferred latent “disease” state does not appear to correspond to either definition, with simulation results showing that the sensitivity and specificity estimated using the two-state model is heavily biased with poor coverage. This is a reminder that we must be careful with interpretation of latent variables, as they exist only in the hypothetical model in which we have defined them and they need not relate to any realistic quantity.
In our simulation studies, there was little power to detect lack-of-fit when the model was incorrect even when N = 1000. This is similar to the lack-of-power observed in LCMs when distinguishing between models for dependence.8 Evaluations of novel diagnostic tests often have far fewer than 1000 participants as data are challenging and expensive to collect. A requirement of thousands or tens of thousands of samples for the evaluation of diagnostic tests would challenge standard practice.
Our results show that model misspecification, even when it cannot be detected by lack-of-fit assessment, may result in the latent class variables no longer representing disease status. This can lead to substantial bias in estimates of sensitivity, specificity and prevalence. As such, we advise caution with the use of latent class models for evaluating novel diagnostic tests in the absence of a gold standard. Care is needed in determining both the number of latent states and what they represent. As it is the latent classes themselves that are of interest, it is critical that the assumptions underlying LCMs be justified when they are used for evaluating diagnostic tests. Such justification should be made by a cross-disciplinary team and account for biological underpinnings of the various tests and diseases under consideration.27,28 In some situations, gold standard tests (those with perfect sensitivity and specificity), or silver standard tests29 (those with perfect specificity) may be available but expensive or difficult to perform. If this is the case, we advise that such tests be used on a subset of patients.29–31 Having some form of ground truth for a subset of observations is advantageous; not only are the latent variables more likely to represent disease status, but the presence of known disease status individuals can aid lack-of-fit assessment.
The model violation we have considered could be overcome through extensions to the statistical model. In principle, we could fit a three-state LCM so that the data generating and fitting models once again coincide. We do not consider that here because our goal was to evaluate the performance of two-state LCMs, as they are highly used for evaluating diagnostic tests with no gold standard. Moreover, there was no evidence of lack-of-fit when fitting the two-state model to the motivating data, nor the majority of simulations, therefore no obvious remedial measures were required. We also note that fitting a three-state model need not solve the underlying problems. Fitting and interpreting three-state models is challenging. The three state model is not identifiable with only four tests.6 Moreover, the interpretation of a three state model is as difficult, if not more so, than the two-state models we focus on here. While more flexible, a three-state model is still unlikely to incorporate all aspects of the complexity of the underlying biological process, including that different tests may vary in the time at which they can detect the disease. Model misspecfication remains a potential issue for more complex models.
Latent variables models have become an essential tool in applied statistical modelling. We think they have tremendous value in uncovering hidden processes that cannot be directly observed. Often the application of latent variable models can generate hypotheses about scientific processes that can be tested using external data, leading to gains in scientific knowledge that might otherwise be impossible. Our objection is not to the use of latent variable models themselves, but to the way in which they are often used and interpreted. Care needs to be taken when the goal is to estimate parameters or to confirm hypotheses that depend on direct interpretation of the hidden process.
Supplementary Material
ACKNOWLEDGEMENTS
This work was supported by the joint US National Institutes of Health (NIH:www.nih.gov) and National Science Foundation (NSF:www.nsf.gov) Ecology of Infectious Disease program (R01 TW009237), and the Research Councils UK, Department for International Development (UK) and UK Biotechnology and Biological Sciences Research Council (BBSRC:www.bbsrc.ac.uk) (BB/J010367/1, BB/L018926, BB/L017679, BB/L018845). It is also funded in part by the Bill & Melinda Gates Foundation funded Typhoid Fever Surveillance in sub-Saharan Africa Program (TSAP) grant (OPPGH5231). MJM received support from University of Otago scholarships: the Frances G. Cotter Scholarship and the MacGibbon Travel Fellowship. MPR received support from the United States National Institutes of Health: the Fogarty International Center (R25 TW009343) Global Health Scholars Fellowship and the National Institute of Allergy & Infectious Diseases (K23 AI116869). MPR and JAC received support from a US National Institutes of Health National Institute for Allergy and Infectious grant (R01 AI121378). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
The findings and conclusions in this report are those of the authors and do not necessarily represent the official position of the Centers for Disease Control and Prevention
Funding information
Bill and Melinda Gates Foundation, Grant/Award Number: OPPGH5231; Biotechnology and Biological Sciences Research Council, Grant/Award Numbers: BB/J010367/1, BB/L017679, BB/L018845, BB/L018926; Fogarty International Center, Grant/Award Numbers: R01 TW009237, R25 TW009343; National Institute of Allergy and Infectious Diseases, Grant/Award Numbers: K23 AI116869, R01 AI121378
Footnotes
SUPPORTING INFORMATION
Additional supporting information may be found online in the Supporting Information section at the end of this article.
DATA AVAILABILITY STATEMENT
The motivating data used in manuscript are available in Table 1. R code to replicate the simulations are available upon request from the lead author.
REFERENCES
- 1.Chen H, Liu K, Li Z, Wang P. Point of care testing for infectious diseases. Clin Chim Acta. 2019;493:138–147. 10.1016/j.cca.2019.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Alonzo TA, Pepe MS. Using a combination of reference tests to assess the accuracy of a new diagnostic test. Stat Med. 1999;18:2987–3003. [DOI] [PubMed] [Google Scholar]
- 3.Garrett ES, Eaton WW, Zeger S. Methods for evaluating the performance of diagnostic tests in the absence of a gold standard: a latent class model approach. Stat Med. 2002;21:1289–1307. [DOI] [PubMed] [Google Scholar]
- 4.Smeden M, Naaktgeboren CA, Reitsma JB, Moons KGM, Groot JAH. Latent class models in diagnostic studies when there is no reference standard – a systematic review. Am J Epidemiol. 2013;179(4):423–431. [DOI] [PubMed] [Google Scholar]
- 5.FDA Statistical guidance on reporting results from studies evaluating diagnostic tests – guidance for industry and FDA staff. US Food and Drug Administration; 2007. [Google Scholar]
- 6.Goodman LA. Exploratory latent structure analysis using both identifiable and unidentifiable models. Biometrika. 1974;61(2):215–231. [Google Scholar]
- 7.Jones G, Johnson WO, Hanson TE, Christensen R. Identifiability of models for multiple diagnostic testing in the absence of a gold standard. Biometrics. 2010;66:855–863. [DOI] [PubMed] [Google Scholar]
- 8.Albert PS, Dodd LE. A cautionary note on the robustness of latent class models for estimating diagnostic error without a gold standard. Biometrics. 2004;60:427–435. [DOI] [PubMed] [Google Scholar]
- 9.Pepe MS, Janes H. Insights into latent class analysis of diagnostic test performance. Biostatistics. 2007;8:474–484. [DOI] [PubMed] [Google Scholar]
- 10.Agresti A Categorical Data Analysis. 2nd ed. Hoboken, NJ: Wiley; 2002. [Google Scholar]
- 11.Maze MJ The Impact of Leptospirosis in Northern Tanzania [PhD thesis]. University of Otago; 2019. http://hdl.handle.net/10523/8838. [Google Scholar]
- 12.Costa F, Hagan JE, Calcagno J, et al. Global morbidity and mortality of leptospirosis: a systematic review. PLoS Negl Trop Dis. 2015;9(9):e0003898. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Haake DA, Levett PN. Leptospirosis in Humans. New York, NY: Springer; 2015:65–97. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Goris MGA, Leeflang MMG, Boer KR, et al. Establishment of valid laboratory case definition for human leptospirosis. J Bacterial Parasitol. 2012;3:1–8. 10.4172/2155-9597.1000132. [DOI] [Google Scholar]
- 15.Maze MJ, Sharples KJ, Allan KJ, Rubach MP, Crump JA. Diagnostic accuracy of leptospirosis whole-cell lateral flow assays: a systematic review and meta-analysis. Clin Microbiol Infect. 2019;25(4):437–444. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Centers for disease control and prevention. Leptospirosis (Leptospira interrogans), 2013 case definition; 2013.
- 17.Plummer M. JAGS: a program for analysis of Bayesian graphical models using Gibbs sampling; 2003.
- 18.Stephens M Dealing with label switching in mixture models. J Royal Stat Soc Ser B (Stat Methodol). 2000;62:795–809. [Google Scholar]
- 19.Papastamoulis P Label.switching: an R package for dealing with the label switching problem in MCMC outputs. J Stat Softw. 2016;69:1–24. [Google Scholar]
- 20.Gelman A, Meng XL, Stern H. Posterior predictive assessment of model fitness via realized discrepancies. Stat Sin. 1996;6:733–759. [Google Scholar]
- 21.Qu Y, Tan M, Kutner MH. Random effects models in latent class analysis for evaluating accuracy of diagnostic tests. Biometrics. 1996;52:797–810. [PubMed] [Google Scholar]
- 22.Limmathurotsakul D, Turner EL, Wuthiekanun V, et al. Fool’s gold: why imperfect reference tests are undermining the evaluation of novel diagnostics: a reevaluation of 5 diagnostic tests for leptospirosis. Clin Infect Dis. 2012;55:322–331. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Hjort NL, Dahl FA, Steinbakk GH. Post-processing posterior predictive p values. J Am Stat Assoc. 2006;101(475):1157–1174. [Google Scholar]
- 24.Link WA, Schofield MR, Barker RJ, Sauer JR. On the robustness of N-mixture models. Ecology. 2018;99:1547–1551. [DOI] [PubMed] [Google Scholar]
- 25.Brooks SP, Gelman A. General methods for monitoring convergence of iterative simulations. J Comput Graph Stat. 1998;7(4):434–455. [Google Scholar]
- 26.Niloofa R, Fernando N, Silva NL, et al. Diagnosis of leptospirosis: comparison between microscopic agglutination test, IgM-ELISA and IgM rapid immunochromatography test. PLoS One. 2015;10(6):e0129236. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Dendukuri N, Hadgu A, Wang L. Modeling conditional dependence between diagnostic tests: a multiple latent variable model. Stat Med. 2009;28:441–461. [DOI] [PubMed] [Google Scholar]
- 28.Schumacher SG, Van Smeden M, Dendukuri N, et al. Diagnostic test accuracy in childhood pulmonary tuberculosis: a Bayesian latent class analysis. Am J Epidemiol. 2016;184:690–700. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Wu Z, Deloria-Knoll M, Hammitt LL, Zeger SL. Partially latent class models for case-control studies of childhood pneumonia aetiology. J Royal Stat Soc Ser C (Appl Stat). 2016;65:97–114. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Albert Paul S, Dodd LE. On estimating diagnostic accuracy from studies with multiple raters and partial gold standard evaluation. J Am Stat Assoc. 2008;103(481):61–73. 10.1198/016214507000000329. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Wu Z, Deloria-Knoll M, Zeger SL. Nested partially latent class models for dependent binary data; estimating disease etiology. Biostatistics. 2017;18:200–213. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The motivating data used in manuscript are available in Table 1. R code to replicate the simulations are available upon request from the lead author.