Abstract
Misclassification of outcomes or event types is common in health sciences research and can lead to serious bias when estimating the cumulative incidence functions in settings with competing risks. Recent work has shown how to estimate nonparametric cumulative incidence functions in the presence of nondifferential outcome misclassification when the misclassification probabilities are known. Here, we extend this approach to account for misclassification that is differential with respect to important predictors of the outcome using misclassification probabilities estimated from external validation data. Moreover, we propose a bootstrap approach in which the observations from both the main study data and the external validation study are resampled to allow the uncertainty in the misclassification probabilities to propagate through the analysis into the final confidence intervals, ensuring appropriate confidence interval coverage probabilities. The proposed estimator is shown to be uniformly consistent and simulation studies indicate that both the estimator and the standard error estimation approach perform well in finite samples. The methodology is applied to estimate the cumulative incidence of death and disengagement from HIV care in a large cohort of HIV infected individuals in sub-Saharan Africa, where a significant death underreporting issue leads to outcome misclassification. This analysis uses external validation data from a separate study conducted in the same country.
Keywords: competing risks, cumulative incidence, external validation data, misclassification
1 |. INTRODUCTION
One goal of health sciences research is to describe the natural history of a specific health outcome in a given target population. A useful parameter for describing this natural history is the risk1,2 or the set of cumulative incidence functions for the outcomes of interest. We can consistently estimate the cumulative incidence functions under a set of assumptions, including (1) that the time to the event is measured without error; (2) that, in settings in which more than one event type is possible, such as when competing events may occur, the event type is observed without error; and (3) that participants in the sample who are under observation at time t are exchangeable with participants in the target population not under observation at time t.
In some settings, the event times are thought to be known, but the event type is subject to misclassification. When an internal validation sample is available, the problem reduces to the problem of missing cause of failure in the competing risks model and the analysis can be performed using methods for missing data, such as multiple imputation.3–5 More recent proposals for the analysis of misclassified competing risks data and more general Markov processes with interval validation samples are based on maximum nonparametric pseudolikelihood estimation.6 In settings without internal validation data, Bakoyannis and Yiannoutsos7 explored the extent of bias in cumulative incidence functions due to nondifferential misclassification of event type and proposed a simple uniformly consistent estimator to correct for such missclassification when the misclassification probabilities are known.
However, the probability that event type is misclassified is usually not known with certainty. Rather, the probability of misclassification is often estimated from internal or external validation data. Ignoring the uncertainty of the estimated misclassification probabilities may lead to estimates that are overly precise, resulting in confidence interval (CI) coverage that is too low.8 Moreover, if the misclassification probabilities depend to one or more variables (ie, there is differential misclassification), applying average misclassification probabilities may produce biased estimates of the cumulative incidence of the event of interest. Here, we present an approach to account for both differential and nondifferential misclassification of event type when estimating nonparametric cumulative incidence functions in settings with external validation data. To account for the uncertainty in the misclassification probabilities that arises when estimating these probabilities from a finite external validation data set, we follow the works of Greenland8 and Chatterjee and Chatterjee9 and employ a bootstrap approach in which both the main study and the external validation data are resampled to compute the standard error of the risk estimates.
We illustrate the proposed approach to estimate the cumulative incidence of mortality in the Family AIDS Care and Education Services (FACES) cohort. In FACES, mortality is underascertained such that people who die are sometimes incorrectly classified as lost to care, leading to bias in estimates of the cumulative incidence of mortality and the cumulative incidence of loss to care. In this setting, an external validation data set, in which a study team traced a sample of patients lost to care in a nearby cohort to determine vital status, was available.
The outline of the remainder of this paper is as follows. Notation, parameters of interest, and the proposed estimator are defined in Section 2. In Section 3, we explore the finite-sample properties of the proposed estimator using a simulation experiment. Section 4 applies the proposed estimator to analysis of data from the motivating study outlined above. A discussion is provided in Section 5. The Appendix presents a proof of uniform consistency of the proposed estimators.
2 |. DATA AND METHOD
For simplicity, we consider two competing events throughout this paper. However, it is straightforward to generalize the methods presented here to the case of more than two competing outcomes. Let Ti and Ci denote the failure time and the true failure cause, respectively, for individual i, with Ci ∈ {1, 2}. Without loss of generality, assume that the cause of interest is Ci = 1. Let Ui denote the right censoring time. Xi represents the observed failure time such that Xi = min(Ti, Ui). Among censored individuals (eg, those for whom Xi = Ui), Ci = 0. The study duration interval is denoted by [0, τ], where τ is finite. The basic identifiable quantities from competing risks data are the cause-specific hazards and the cumulative incidence functions.10–12 The cause specific hazard function for event type j is defined as
The cumulative incidence function is defined as
where Λj(t) is the cumulative cause-specific hazard, . The cumulative incidence depends on the cause-specific hazards for both causes of failure through the overall survival function S(·). The cumulative incidence function can be estimated nonparametrically using the Aalen-Johansen estimator13
where and are the Kaplan-Meier and Nelson-Aalen estimators for the overall survival function and the cumulative cause-specific hazard function, respectively.
In settings where the outcome ascertainment procedure is imperfect, we observe in place of Ci, where is the measured, and potentially misclassified cause of failure. Misclassification occurs when for at least some study participants. We define misclassification probabilities a and b as
and
Here, as in our motivating example, we assume that right censoring is always correctly ascertained, such that if and only if Ci = 0. Misclassification can be addressed by (a) measuring the outcome type using a gold standard diagnostic procedure in a small random subsample of the participants to retrieve the true Ci (as in an internal validation sample), (b) using estimates of the misclassification probabilities based on an external validation study, or (c) drawing on prior knowledge of the misclassification probabilities. In studies where there are no internal validation data, but the misclassification probabilities are a priori known, the estimator by Bakoyannis and Yiannoutsos7 can be used to account for misclassification
| (1) |
and
where , j = 1, 2, is the Aalen-Johansen estimator based on the misclassified cause of failure C⋆. Note that the supremum in (1) ensures that will be monotonic; may not be monotonically increasing.
If a and b are known, then is uniformly consistent for F1(t).7 If and are estimated from an external validation sample, provides a reasonable estimator for F1(t), provided that and are consistent for a and b in the main study (or transportable (see the work of Ha and Tsodikov14p10) to the main study). More precisely, it can be shown that
as n → ∞ and n0 → ∞ (where n is the sample size in the main study and n0 is the number of events included in the external validation sample), which means that is uniformly consistent for F1(t).
2.1 |. Accounting for differential misclassification of event type
When misclassification depends on time T or participant characteristics Z (eg, gender), we can relax the misclassification probabilities a and b to be dependent on these factors such that and . In this setting, we generalize the nonparametric estimator of the cumulative cause-specific hazard by Ha and Tsodikov14 to write
| (2) |
and
where r1(t, z) = {1 − b(t, zi)}∕{1 − a(t, zi) − b(t, zi)}, r2(t, z) = b(t, zi)∕{1 − a(t, zi) − b(t, zi)}, , j = 1, 2, and Yi(t) = I(Ti ≥ t).
Estimates of the misclassification probabilities and may be obtained using maximum likelihood. If the corresponding parametric models are correctly specified and and are consistent estimators of a(t, z) and b(t, z) in the main study, a reasonable estimator for F1(t) is the plug-in estimator
| (3) |
More importantly,
as n → ∞ and n0 → ∞, which means that is uniformly consistent for F1(t) under differential (or nondifferential) misclassification. The proof of uniform consistency of this estimator is outlined in the Appendix.
Often, the choice between applying the estimator that allows for differential misclassification and the estimator that does not relies on investigator knowledge of the misclassification process. If uncertain about whether a and b differ by T or Z, one may perform a likelihood ratio test comparing parametric models for and with and without terms for T and Z, provided that validation data containing information on T and Z are available.
2.2 |. Accounting for uncertainty and estimated from external validation data
Suppose an external validation data set is available in which both the gold standard outcome measure Ci and the possibly misclassified measure of the outcome type are available for all participants. Information on the events occurring among those in the validation sample can be used to estimate and (if misclassification is nondifferential) or and if misclassification is differential. Note that, because we focus on accounting for misclassification of event type (ie, between Ci = 1 and Ci = 2), only events in the validation sample contribute to estimation of and . We will let n0 represent the number of events in the validation sample.
Assuming that misclassification is nondifferential with respect to predictors of the outcome, the misclassification probabilities a and b may be estimated in the external validation data as
and one may consistently estimate F1(t) by plugging and into (1).
As described above, we may relax the assumption that misclassification is nondifferential by allowing the misclassification probabilities to vary over time and strata of covariates z, provided that covariates z are available in both the main study data and the external validation data. In such cases, and may be estimated in the validation data using logistic regression such that
and
where w = (1, Ti, Zi)T. If the parametric models used to estimate and are correctly specified, one may consistently estimate F1(t) by plugging and into (2).
However, simply plugging values of and or and estimated from the external validation study into (1) or (2) as in the work of Bakoyannis and Yiannoutsos7 ignores the uncertainty in the estimated misclassification probabilities, which will lead to CIs with a low coverage probability. Because the misclassification probabilities are estimated using an external and independent sample, we can define the following nonparametric bootstrap approach for 1−α CI calculation, when the external study data set is available.
For k = 1 to M,
Draw bootstrap sample k by sampling n participants with replacement from the main study data;
Draw bootstrap sample k by sampling n0 events with replacement from the external validation data to estimate and or and ;
Apply and (or and ) to bootstrap sample k from the main study data to estimate the cumulative incidence using under an assumption of nondifferential misclassification (or under an assumption of differential misclassification).
The α/2 and 1 –α/2 percentiles of the distribution of the estimates over the M bootstrap samples at time t are the limits of the 1 − α CI for the estimated cumulative incidence at t for t ∈[0, τ]. This approach for CI calculation will have the correct nominal coverage as it incorporates the uncertainty in the misclassification probabilities.
However, in some settings, resampling from the external validation data may not yield ideal results. For example, in small validation samples, individual bootstrap samples from the validation sample may contain zero events of type C = 1 or C = 2, resulting in errors in estimating a or b, respectively. Moreover, in some situations, published summary measures and are available for the external validation data, but the validation data themselves are not available. In these settings, one may replace the nonparametric bootstrap procedure of step 2 with a parametric approach by drawing and for each bootstrap from the asymptotic distribution of and . For example, allowing for differential misclassification where w = (t, z)T, the simulated and for each individual in bootstrap sample k are
and
where and are random draws from and , respectively.
2.3 |. Transportability of the misclassification probabilities
The validity of the proposed approach rests of the assumption that a and b or a(t, z) and b(t, z) are the same in the study population as in the external validation study population. However, unlike other approaches to account for outcome misclassification,15–18 the proposed approach does not require equivalence of π12(t, z) = P(C = 1 C⋆ = 2, T = t, Z = z) and π21(t, z) = P(C = 2|C⋆ = 1, T = t, Z = z) between the study sample and validation data. Note that these probabilities are the positive and negative predictive values if the outcome is binary and occurs at a single time point. Like the predictive values, which depend on both the misclassification probabilities a and b and the prevalence of disease in the study sample, π12(t, z) and π21(t, z) depend on the cause-specific hazards of both event types in the population in addition to the misclassification probabilities. Specifically,
and
where λ1(t; z) and λ2(t; z) are the cause specific hazard functions for outcomes 1 and 2, respectively.
Consequently, using π12(t, z) and π21(t, z) from an external validation study would require the assumption that both the event frequency for both event types and the misclassification probabilities are equal between the main and the external validation study. This assumption is much stronger than our assumption of transportability of only the misclassification probabilities between the two populations. Like sensitivity and specificity in the diagnostic testing context, the classification probabilities a(t, z) and b(t, z) are intrinsic properties of the event classification method, and therefore more likely to be transportable than the predictive values π12(t, z) and π21(t, z).
3 |. SIMULATION STUDIES
3.1 |. Simulation set up
We assess the finite sample performance of the proposed approach to account for outcome misclassification in cumulative incidence functions using a Monte Carlo simulation experiment. In each of 2000 simulated data sets, we let i = 1, …, 2000 index simulated individuals. We examined the performance of the proposed approach under scenarios in which the amount of misclassification, the size of the external validation set, and the proportion experiencing the competing event varied. We compared the proposed estimators under settings with no misclassification, nondifferential misclassification, and misclassification that was differential with respect to a predictor of the outcome.
In each simulated data set, half of simulated participants were randomly assigned covariate Zi = 1 and half were assigned to Zi = 0. Among simulated participants with Zi = 1, 30% of the simulated subjects were randomly assigned to cause Ci = 1 and 70% were assigned to cause Ci = 2, while among those with Zi = 0, 60% were assigned to cause Ci = 1 and 40% were assigned to cause Ci = 2, so that the overall P(Ci = 1) = 45%. Among those assigned to cause Ci = 1, the time to failure Ti followed a Weibull distribution with the shape parameter α set to 1.25 and the scale parameter λ1 determined by Zi such that
where λ1z = 2+log(10)−log(10)z. The time to failure for those assigned to causeCi = 2 followed an exponential distribution with λ2 determined by Zi such that
where λ2z = 4 + log(5) − log(5)z. The time to censoring Ui was a uniform random number between 0 and 5. The observed failure time was the minimum of the time to failure from either cause and the time to censoring, Xi = min(Ti, Ui). If Xi = Ui, then Ci = 0. The observed cause of failure, , was determined by values a(t, z) and b(t, z). Note that, for simplicity, we allowed the misclassification of cause Ci to vary as a function of z, but not of t, such that a(t, z) = a(z) and b(t, z) = b(z). For simulated subjects with Ci = 1, was drawn from a Bernoulli distribution with probability 1(z = 1){1 − a(1)} + 1(z = 0){1 − a(0)}. For subjects with Ci = 2, ; otherwise, . If Ci = 0, then .
3.1.1 |. Simulations under nondifferential misclassification
First, we compared the performance of the proposed estimator under a range of settings with nondifferential misclassification. In these scenarios, we set a(1) = a(0) = a and b(1) = b(0) = b. We compared estimators under four sets of values for a and b, ie, (0, 0), (0.1, 0.1), (0.3, 0.1), and (0.1, 0.3).
For each simulated data set, we simulated a corresponding external validation data set with the number of events set to n0, where n0 took on a value between 25 and 1000 for the various simulated scenarios. In the first set of simulations, the external validation data set had the same data generating mechanism as the main data set. Specifically, in the first set of simulations, the proportion assigned to Ci = 1 was 0.45 in both the main study data and the validation data, or pmain = pval = 0.45. To explore whether the proposed approach was sensitive to differences in the incidence of the competing event between the main study and the validation data, the second set of simulations assigned pval = 0.75 while maintaining pmain = 0.45. In all simulation experiments, values of a and b were the same in the main study and in the validation data. In the simulated external validation data, we assumed that both Ci and were observed.
The true cumulative incidence at time t of event type 1 was F1(t) = ∑zP(Z = z) × pmain,z × (1 − exp{(−t∕λ1z)1.25}) in all scenarios. For each of the simulated scenarios, we compared the bias, standard deviation of the bias, mean squared error, and 95% CI coverage between the following two estimators:
The standard Aalen-Johansen estimator based on the misclassified cause of failure ;
The proposed estimator , which accounted for the misclassification in the cumulative incidence function using Equation (1), with information on a and b from the external validation data.
We report bias, empirical standard error, average estimated standard error, root mean squared error, and 95% CI coverage probability for each approach under each scenario. For ease of interpretation, we have rescaled all values to refer to 100 times the cumulative incidence, which can be thought of as the percentage of study participants experiencing a particular outcome type. Bias was defined as 100 times the average of the difference between the estimated cumulative incidence and the true value, the empirical standard error was the standard deviation of the estimated cumulative incidence (multiplied by 100) over all simulated data sets, and the average estimated standard error was the average standard error of the cumulative incidence (multiplied by 100) estimated using the bootstrap approaches across all simulated data sets. Root mean squared error was the square root of the sum of the bias squared and the empirical standard error squared, and the 95% CI coverage was the proportion of 95% CIs constructed in the simulated data sets that contained the true value.
Confidence intervals for were the α∕2 and 1 − α∕2 percentiles of across 1000 bootstrap samples of the main study data. We compared coverage under two strategies, ie, the first approach constructed 95% CIs as the 2.5th and 97.5th percentiles of across 1000 bootstrap samples in which the main study data only was resampled (the single bootstrap approach). In the second approach, 95% CIs around were constructed as the 2.5th and 97.5th of in 1000 bootstrap samples in which both the main study data were resampled and and were redrawn from their asymptotic distributions in each bootstrap. Results were evaluated at times t = (1, 2, 3).
3.1.2 |. Simulations under differential misclassification
We next compared the performance of the standard estimator , the proposed estimator , which assumed nondifferential misclassification, and the second proposed estimator , which allowed for differential misclassification, under settings with no misclassification, nondifferential misclassification, and misclassification that was differential with respect to Z.
The second set of simulations used the same data generating mechanism as the simulations in Section 3.1.1 above, with the exception of how was generated. In this set of simulations, we examined scenarios with three types of misclassification.
No misclassification: a(1) = a(0) = b(1) = b(0) = 0;
Nondifferential misclassification: a(1) = a(0) = 0.1; b(1) = b(0) = 0.3;
Differential misclassification: a(1) = 0; a(0) = 0.3; b(1) = 0; b(0) = 0.2.
In each scenario, the main study data was composed of 2000 individuals and the validation data included 300 simulated events. Results were evaluated at times t = 1.5 (true F1(1.5) = 15%) and t = 3 (true F1(3) = 26%). In all scenarios in this set of simulations, and were resampled from their asymptotic distribution in each bootstrap sample.
3.1.3 |. Simulations with error in and
In some settings, outcomes measured in external validation data may not share the misclassification properties of outcomes in the main study data. In a final set of simulation experiments, we explored the setting with nondifferential misclassification in which the true value of a was 0.1 and the true value of b was 0.3. As in the aforementioned scenarios, the true prevalence of outcome type C = 1 in the main study data was 0.45. In each of 2000 simulated worlds, we simulated a series of 25 external validation data sets of size n0 = 500 in which and took on values corresponding to a±0.1 and b±0.1, respectively. In each scenario, we estimated the cumulative incidence of event type C = 1 using the standard approach (ie, ) and using the proposed estimator in conjunction with and estimated from each external validation data set. We defined bias as 100 times the average difference between the true cumulative incidence at t = 3 and the estimated cumulative incidence over the 2000 iterations of the experiment.
3.2 |. Simulation results
Table 1 compares the performance of the standard estimator and the proposed in settings with nondifferential misclassification and a validation data set composed of n0 = 300 events. In scenarios with no misclassification (a = b = 0), both estimators had little bias and appropriate 95% CI coverage. Bias in the standard estimator increased as the amount of misclassification in C⋆ increased. Bias was smaller for than for in all scenarios examined where a ≠ 0 or b ≠ 0. The reduction in bias was slightly offset by an increase in the standard deviation of the bias for , but root mean squared error was smaller for than for for many scenarios examined. Notably, had appropriate CI coverage in all scenarios when standard errors were estimated by bootstrapping both the main study data and the validation data. In contrast, 95% CI coverage was too low for (likely due to bias in the point estimates), and for when only the main study was resampled as part of the bootstrap (due to the failure to account for uncertainty in and ).
TABLE 1.
Biasa, empirical standard error (ESE)b, average estimated standard error (ASE)c, root mean squared error (RMSE)d, and 95% CI coverage of cumulative incidence estimates in a standard analysis and an analysis accounting for misclassification from 2000 simulated cohorts of 2000 individuals evaluated at three time points t = (1, 2, 3)
| pvale | a | b | Bias | ESE | ASE | RMSE | Coverage | Bias | ESE | ASE | RMSE | Coveragef | Coverageg |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| t = 1 | |||||||||||||
| 0.45 | 0 | 0 | −0.02 | 0.76 | 0.75 | 0.76 | 0.94 | −0.02 | 0.76 | 0.75 | 0.76 | 0.94 | 0.94 |
| 0.45 | 0.1 | 0.1 | 0.08 | 0.76 | 0.75 | 0.77 | 0.94 | −0.04 | 1.01 | 1.02 | 1.01 | 0.90 | 0.95 |
| 0.45 | 0.1 | 0.3 | 2.29 | 0.84 | 0.82 | 2.45 | 0.20 | −0.06 | 1.36 | 1.37 | 1.36 | 0.88 | 0.94 |
| 0.45 | 0.3 | 0.1 | −1.86 | 0.66 | 0.68 | 1.98 | 0.22 | −0.08 | 1.34 | 1.41 | 1.34 | 0.89 | 0.95 |
| 0.75 | 0 | 0 | −0.02 | 0.76 | 0.75 | 0.76 | 0.94 | −0.02 | 0.76 | 0.75 | 0.76 | 0.94 | 0.94 |
| 0.75 | 0.1 | 0.1 | 0.08 | 0.76 | 0.75 | 0.77 | 0.94 | −0.08 | 0.99 | 1.01 | 0.99 | 0.92 | 0.94 |
| 0.75 | 0.1 | 0.3 | 2.29 | 0.84 | 0.82 | 2.45 | 0.20 | −0.08 | 1.40 | 1.40 | 1.40 | 0.86 | 0.95 |
| 0.75 | 0.3 | 0.1 | −1.86 | 0.66 | 0.68 | 1.98 | 0.22 | −0.09 | 1.24 | 1.33 | 1.24 | 0.92 | 0.96 |
| t = 2 | |||||||||||||
| 0.45 | 0 | 0 | −0.01 | 0.98 | 1.00 | 0.98 | 0.94 | −0.01 | 0.98 | 1.00 | 0.98 | 0.94 | 0.94 |
| 0.45 | 0.1 | 0.1 | 0.06 | 0.95 | 1.00 | 0.95 | 0.96 | 0.06 | 1.51 | 1.54 | 1.51 | 0.86 | 0.95 |
| 0.45 | 0.1 | 0.3 | 3.99 | 1.02 | 1.07 | 4.12 | 0.02 | −0.03 | 2.16 | 2.15 | 2.16 | 0.83 | 0.94 |
| 0.45 | 0.3 | 0.1 | −3.75 | 0.83 | 0.91 | 3.84 | 0.02 | 0.15 | 2.14 | 2.29 | 2.15 | 0.82 | 0.95 |
| 0.75 | 0 | 0 | −0.01 | 0.98 | 1.00 | 0.98 | 0.94 | −0.01 | 0.98 | 1.00 | 0.98 | 0.94 | 0.94 |
| 0.75 | 0.1 | 0.1 | 0.06 | 0.95 | 1.00 | 0.95 | 0.96 | 0.00 | 1.43 | 1.51 | 1.43 | 0.89 | 0.94 |
| 0.75 | 0.1 | 0.3 | 3.99 | 1.02 | 1.07 | 4.12 | 0.02 | −0.05 | 2.16 | 2.23 | 2.16 | 0.79 | 0.94 |
| 0.75 | 0.3 | 0.1 | −3.75 | 0.83 | 0.91 | 3.84 | 0.02 | 0.12 | 1.92 | 2.12 | 1.92 | 0.86 | 0.96 |
| t = 3 | |||||||||||||
| 0.45 | 0 | 0 | 0.02 | 1.15 | 1.11 | 1.15 | 0.93 | 0.02 | 1.15 | 1.11 | 1.15 | 0.93 | 0.93 |
| 0.45 | 0.1 | 0.1 | 0.04 | 1.12 | 1.11 | 1.12 | 0.95 | 0.02 | 1.81 | 1.87 | 1.81 | 0.83 | 0.95 |
| 0.45 | 0.1 | 0.3 | 5.36 | 1.16 | 1.18 | 5.48 | 0.00 | −0.10 | 2.59 | 2.69 | 2.59 | 0.79 | 0.95 |
| 0.45 | 0.3 | 0.1 | −5.25 | 0.98 | 1.03 | 5.34 | 0.00 | 0.17 | 2.71 | 2.92 | 2.71 | 0.75 | 0.97 |
| 0.75 | 0 | 0 | 0.02 | 1.15 | 1.11 | 1.15 | 0.93 | 0.02 | 1.15 | 1.11 | 1.15 | 0.93 | 0.93 |
| 0.75 | 0.1 | 0.1 | 0.04 | 1.12 | 1.11 | 1.12 | 0.95 | 0.02 | 1.79 | 1.86 | 1.79 | 0.85 | 0.96 |
| 0.75 | 0.1 | 0.3 | 5.36 | 1.16 | 1.18 | 5.48 | 0.00 | −0.13 | 2.75 | 2.87 | 2.75 | 0.76 | 0.95 |
| 0.75 | 0.3 | 0.1 | −5.25 | 0.98 | 1.03 | 5.34 | 0.00 | 0.12 | 2.49 | 2.68 | 2.49 | 0.80 | 0.95 |
Abbreviations: ASE, average estimated standard error; ESE, empirical standard error; RMSE: root mean squared error.
Bias was 100 times the average difference between the true and estimated values of the cumulative incidence.
ESE was the standard deviation of the estimated cumulative incidence across all simulations.
ASE was the standard error estimated using the proposed bootstrap approach averaged across all simulated data sets.
RMSE was the sum of the bias squared and the ESE squared.
P(C = 1|C > 0) in the validation data. In the main study data, P(C = 1|C > 0) = 0.45.
Coverage based on 1000 bootstrap resamples of main study data only (treating and as fixed).
Coverage based on 1000 bootstrap samples in which both the main study data were resampled and and were redrawn from their asymptotic distributions.
Figure 1 illustrates the relationship between the size of the external validation data set n0 and estimated standard error and 95% CI coverage under each approach. Using the standard estimator , the standard error was small and independent of the size of the external validation data, as expected. In addition, due to the bias in the estimated cumulative incidence, 95% CI coverage was low for all scenarios examined. The standard error was also independent of the size of the validation sample when we applied , but ignored the uncertainty in estimation of a and b by bootstrapping only the main study data to obtain standard errors. When this estimator was applied, 95% CI coverage was much too low at low validation data set sizes, and rose as the size of the validation data grew. Finally, when we applied accounting for the uncertainty in a and b using the approach that both bootstrapped the main study data and resampled a and b for each bootstrap sample, standard error was dependent on the size of the external validation data, and 95% coverage remained near its nominal value for all sizes of the validation data examined. Specifically, for external validation data sets with n0 > 100, bias was low (approximately 0.05%) and 95% CI coverage was near 95%. At smaller validation data set sizes (eg, n0 = 25), bias was slightly elevated (approximately 0.50%), though remained smaller than bias under the standard approach (3.30%). Moreover, at smaller validation study sizes, 95% CI coverage was somewhat elevated (approximately 98%).
FIGURE 1.
(A) Standard error and (B) 95% confidence interval coverage in simulations with nondifferential misclassification in which a = 0.1 and b = 0.3 under various sizes of the external validation data set. Solid black curves represent values using the standard approach that ignores misclassification, dotted curves represent values obtained by plugging in estimated values of and into Equation (1) and ignoring uncertainty in the estimation of and , and the dashed curves represent values obtained using the proposed approach, including bootstrapping of the validation and main study data. The plot of (B) 95% confidence interval coverage also contains a grey reference line at 95%
Table 2 compares the performance of , , and in scenarios with no misclassification, nondifferential misclassification, and differential misclassification. When no misclassification was present, the three methods yielded nearly identical results with little bias and appropriate 95% CI coverage. When misclassification was nondifferential with respect to Z, the standard estimator was biased and had poor coverage, while and produced results will little bias and appropriate coverage. When misclassification was differential with respect to Z, produced results with little bias and appropriate coverage, while produced biased results with very poor coverage. Notably, using the estimator that assumed nondifferential misclassification in the setting where differential misclassification was present produced results with some bias and coverage probability that was too low. However, the reduction in bias seen when implementing in settings with differential misclassification came at the cost of decreased precision.
TABLE 2.
Performance of estimators of the cumulative incidence at t = 1.5 and t = 3 years in 2000 simulated cohorts of 2000 individuals with external validation data of size n0 = 300 under scenarios with no misclassification, nondifferential misclassification, and differential misclassification
| t=1.5 | t=3 | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Approach | Bias | ESE | ASE | RMSE | Coverage | Bias | ESE | ASE | RMSE | Coverage |
| a1 = a2 = b1 = b2 = 0 | ||||||||||
| −0.01 | 0.92 | 0.90 | 0.90 | 0.94 | −0.03 | 1.16 | 1.10 | 1.10 | 0.94 | |
| −0.01 | 0.92 | 0.90 | 0.90 | 0.94 | −0.03 | 1.16 | 1.10 | 1.10 | 0.94 | |
| 0.00 | 0.32 | 0.90 | 0.90 | 0.94 | −0.01 | 1.16 | 1.10 | 1.10 | 0.94 | |
| a1 = a1 = 0.1; b1 = b2 = 0.3 | ||||||||||
| 3.22 | 0.99 | 0.97 | 3.37 | 0.08 | 5.35 | 1.20 | 1.18 | 5.50 | 0.00 | |
| −0.03 | 1.69 | 1.82 | 1.69 | 0.97 | −0.08 | 2.23 | 2.28 | 2.23 | 0.96 | |
| −0.02 | 1.71 | 1.86 | 1.71 | 0.97 | −0.05 | 2.55 | 2.88 | 2.55 | 0.97 | |
| a1 = 0; a2 = 0.3; b1 = 0; b2 = 0.2 | ||||||||||
| −1.66 | 0.87 | 0.85 | 1.87 | 0.50 | −3.45 | 1.10 | 1.05 | 3.62 | 0.11 | |
| −0.12 | 1.21 | 1.23 | 1.21 | 0.95 | −0.71 | 1.69 | 1.76 | 1.84 | 0.92 | |
| 0.10 | 1.47 | 1.55 | 1.47 | 0.96 | 0.19 | 2.32 | 2.52 | 2.33 | 0.97 | |
Abbreviations: ASE, average estimated standard error; ESE, empirical standard error; RMSE, root mean squared error.
Bias was 100 times the average difference between the true and estimated values of the cumulative incidence.
ESE was the standard deviation of the estimated cumulative incidence across all simulations.
ASE was the standard error estimated using the proposed bootstrap approach averaged across all simulated data sets.
RMSE was the sum of the bias squared and the ESE squared.
Table 3 illustrates the sensitivity of the proposed approach to violations in the assumption of transportability of a and b between the main study and the validation data. Each row displays the bias, estimated standard error, average standard error, root mean squared error, and 95% CI coverage of the standard approach (employing ) and the proposed approach (employing ) applied in tandem with external validation data sets in which a0 ranges from 0 to 0.2 and b0 ranges from 0.2 to 0.4. Under most scenarios in which the misclassification probabilities in the external data, a0 and b0, were somewhat close to the true value of a and b in the main study data, bias and root mean squared error are reduced using the proposed approach. Moreover, 95% CI coverage is closer to the nominal value under the proposed approach (compared to the standard approach) under all scenarios examined.
TABLE 3.
Performance of estimators of the cumulative incidence at t = 3 years in 2000 simulated cohorts of 2000 individuals with external validation data of size n0 = 500 when a = 0.1 and b = 0.3 in the main study under scenarios where the misclassification probabilities in the validation sample, a0 and b0, differ from a and b in the main study
| Estimates | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| a0 | b0 | Bias | ESE | ASE | RMSE | Coverage | Bias | ESE | ASE | RMSE | Coverage |
| 0.00 | 0.20 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −0.01 | 1.49 | 1.49 | 1.49 | 0.96 |
| 0.00 | 0.25 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −1.82 | 1.62 | 1.62 | 2.44 | 0.80 |
| 0.00 | 0.30 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −3.87 | 1.78 | 1.78 | 4.26 | 0.40 |
| 0.00 | 0.35 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −6.23 | 1.96 | 1.98 | 6.53 | 0.10 |
| 0.00 | 0.40 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −8.95 | 2.22 | 2.22 | 9.22 | 0.01 |
| 0.05 | 0.20 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 1.74 | 1.67 | 1.72 | 2.41 | 0.84 |
| 0.05 | 0.25 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −0.06 | 1.81 | 1.85 | 1.81 | 0.95 |
| 0.05 | 0.30 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −2.14 | 1.96 | 2.01 | 2.91 | 0.83 |
| 0.05 | 0.35 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −4.55 | 2.16 | 2.21 | 5.03 | 0.45 |
| 0.05 | 0.40 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −7.37 | 2.43 | 2.47 | 7.76 | 0.13 |
| 0.10 | 0.20 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 3.79 | 1.92 | 1.97 | 4.25 | 0.51 |
| 0.10 | 0.25 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 2.00 | 2.07 | 2.10 | 2.88 | 0.85 |
| 0.10 | 0.30 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −0.08 | 2.23 | 2.28 | 2.23 | 0.96 |
| 0.10 | 0.35 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −2.52 | 2.44 | 2.49 | 3.50 | 0.83 |
| 0.10 | 0.40 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −5.43 | 2.73 | 2.77 | 6.08 | 0.49 |
| 0.15 | 0.20 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 6.16 | 2.23 | 2.28 | 6.55 | 0.20 |
| 0.15 | 0.25 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 4.42 | 2.39 | 2.43 | 5.03 | 0.56 |
| 0.15 | 0.30 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 2.37 | 2.56 | 2.63 | 3.49 | 0.87 |
| 0.15 | 0.35 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −0.07 | 2.78 | 2.87 | 2.78 | 0.96 |
| 0.15 | 0.40 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −3.04 | 3.09 | 3.18 | 4.34 | 0.85 |
| 0.20 | 0.20 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 8.88 | 2.60 | 2.67 | 9.25 | 0.05 |
| 0.20 | 0.25 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 7.24 | 2.78 | 2.86 | 7.76 | 0.23 |
| 0.20 | 0.30 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 5.28 | 2.98 | 3.09 | 6.06 | 0.62 |
| 0.20 | 0.35 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | 2.89 | 3.24 | 3.38 | 4.34 | 0.90 |
| 0.20 | 0.40 | 5.35 | 1.20 | 1.18 | 5.48 | 0.00 | −0.08 | 3.60 | 3.75 | 3.60 | 0.97 |
Abbreviations: ASE, average estimated standard error; ESE, empirical standard error; RMSE, root mean squared error.
Bias was 100 times the average difference between the true and estimated values of the cumulative incidence.
ESE was the standard deviation of the estimated cumulative incidence across all simulations.
ASE was the standard error estimated using the proposed bootstrap approach averaged across all simulated data sets.
RMSE was the sum of the bias squared and the ESE squared.
4 |. DATA ANALYSIS
We applied the proposed approach to estimate the 1-year cumulative incidence of death while in care at a clinic supported by the FACES program in western Kenya. Dates of death were collected among people in care in FACES by clinic staff. Patients who did not return for follow-up visits and were not known to have died were classified as lost to clinical care. However, vital status among these patients who did not return for scheduled follow-up visits was uncertain; some of these patients were truly lost to clinical care while others likely died. Here, we apply the proposed approach to estimate the cumulative incidence of mortality in this cohort accounting for misclassification of some deaths as loss to care.
Let the 3886 patients enrolled in FACES between November 2001 and July 2011 be indexed from i = 1 to 3886. A patient was considered to be “lost to care” after 90 days without a clinic visit. The time from enrollment to the first of death or loss to care will be represented by Ti and the time from enrollment to censoring at 1 year or database close on July 27, 2011 will be represented by Ui. Xi = min(Ti, Ui) is the observed time each participant spends in the cohort. is the observed event type. if the participant has an observed date of death prior to loss to clinic or censoring, if the participant is lost to clinic prior to censoring, and if the participant is censored. Ci, the true event type at time Xi, was not observed in this cohort.
However, an Academic Model for the Prevention and Treatment of HIV/AIDS (AMPATH) program in the same geographic area conducted a tracing study during the same time period in which participants who did not appear at the clinic for a scheduled follow-up visit were sampled for an outreach program to determine vital status. While it is conceivable that some patients included in the FACES data set also sought care within the AMPATH program, the distance between clinics served by each program makes such overlap unlikely. The AMPATH conducted outreach visits among 4238 patients who had not returned to the clinic within 2 months of a scheduled visit, and were therefore classified as lost to care. Of these patients originally classified as lost to care, 1143 had died within 2 months of their missed visit. Data from this tracing study of 4238 patients were then used to impute the outcome status for the other 23 861 patients who were lost to care but were not traced conditional on age, sex, and time of loss to care, under the assumption that patients successfully traced were a stratified random subset of all patients lost. Note that all observed deaths were assumed to be true deaths. This implies that b = 0 by assumption in this example. After applying the approach describe above, we estimated that 77% of deaths were classified as loss to follow-up, implying that . Data from the AMPATH program were used as validation data to account for misclassification of loss to care in FACES.
We compare the 1-year cumulative incidence of mortality estimated using the standard estimator and the proposed estimator for nondifferential misclassification . Specifically, the steps to estimate were as follows.
Estimate in the validation data. As mentioned above, estimating first required imputing outcomes for patients lost to care and not traced, and then calculating in the imputed data set. To ease computational burden in this example, we used parametric fractional imputation,19 rather than multiple imputation, to impute outcomes for those not traced in the validation sample. was set to 0 by assumption in this example, and was estimated to be 0.77.
Plug and into Equation (1) to obtain point estimate .
Draw 500 bootstrap samples from the validation data and the main study data, respectively.
Estimate in each bootstrap sample from the validation data. Due to the nature of the validation data, estimating required performing the imputation of outcomes for patients lost but not traced in each bootstrap sample of the validation data to allow uncertainty in the imputation process to propagate through the analysis to the final CI. By assumption, was set to 0 for all k.
Plug and into Equation (1) to estimate .
The α∕2 and 1 − α∕2 percentiles of the distribution of the estimates over the M bootstrap samples at time t year are the limits of the 1 − α CI for the estimated cumulative incidence of mortality.
Of the 3886 patients who enrolled in the FACES program during the study time period, 34% were male and the mean age at enrollment was 34 years. Of these patients, 67 died in the year after enrollment and 874 appeared to become lost to care. Using the standard approach, the estimated 1-year cumulative incidence of mortality was 1.76% (95% CI: 1.33, 2.19) and the estimated 1 year cumulative incidence of loss to care was 23.54% (95% CI: 22.17, 24.91). After applying the proposed approach to account for misclassification of some deaths as loss to care, the estimated 1-year cumulative incidence of mortality was 6.59% (95% CI: 4.92, 8.27) and the estimated 1-year cumulative incidence of loss to care was 18.71% (95% CI: 16.77, 20.65). As expected, standard errors estimated using the proposed approach (including bootstrapping the main and validation data) for cumulative incidence of mortality (standard error ) and loss to care were larger than standard errors resulting from an approach that accounted for the outcome misclassification without rebootstrapping the validation data ( and 0.89, respectively).
Figure 2 displays the estimated cumulative incidence of mortality and loss to care over time using the standard approach and the proposed approach and pointwise 95% confidence limits calculated at each event time. The proposed approach yielded estimates of mortality that were higher than the standard approach at all time points under study and of loss to care that were lower than the standard approach at all time points.
FIGURE 2.
Estimated cumulative incidence of mortality and loss to care and pointwise 95% confidence intervals over 1 year after enrollment in the FACES program in western Kenya between 2001 and 2011 under the standard approach to estimate cumulative incidence (dashed lines) and the proposed approach (solid lines)
5 |. DISCUSSION
Misclassification of event type is common in a variety of settings and can lead to biased estimates and invalid conclusions. Here, we have described a nonparametric approach to account for this misclassification when estimating cumulative incidence functions in the presence of competing events. This work extends previously proposed approaches7 in two ways. First, the approach proposed here allows consistent estimation of cumulative incidence functions in the presence of misclassification that is differential with respect to time and/or a set of covariates; second, we propose bootstrapping both the validation data and the main study data to incorporate uncertainty in the misclassification probabilities into the final 95% CI around the estimated cumulative incidence. The proposed estimators where shown to be uniformly consistent. Finally, in the example presented here, applying the proposed approach to estimate mortality in an HIV treatment program resulted in mortality estimates that were nearly four times the mortality that would be estimated assuming perfect outcome classification.
In simulations, we illustrated that the proposed approach works well in settings with both false negatives (a > 0) and false positives (b > 0). However, in the example, misclassification was unidirectional (ie, no patients lost to follow-up were erroneously classified as dead), meaning that b was set to 0. Other applications of this approach could leverage its ability to account for bidirectional misclassification. For example, this approach would be valuable to estimate the cumulative incidence of cause-specific mortality in the presence of misclassification of cause of death.
We presented two options for allowing uncertainty in the estimation of the misclassification probability estimates into the final CIs. In the first, the investigator would resample from both the main study data and the validation data in each bootstrap iteration. However, when the validation data set is small, this approach could fail due to small or zero cell sizes in a given sample of the validation data. To mitigate this challenge, we proposed a second option that involved resampling the misclassification probabilities from their asymptotic distributions in each bootstrap sample. This approach is appealing because it allows inference in a wide variety of settings (including settings where the validation data themselves are unavailable), but relies on correct specification of a model for the misclassification probabilities.
A key assumption of the proposed methodology is that the misclassification probabilities are the same between the main study and the external validation study populations. This assumption is reasonable in many settings as these probabilities are properties of the event ascertainment methods and typically do not depend on the underlying event frequency of the events for binary outcomes. Therefore, the assumptions required to implement the proposed approach are weaker than the assumptions required for methods that rely on transportability of the predictive values π12(t, z) and π21(t, z), which depend on the event frequency of both the event types, from external studies to account for misclassification.20 However, results may be sensitive to differences in the misclassification probabilities a and b between the validation study and the main study data, as we illustrate in Table 3. In settings where the transportability of a and b is uncertain, sensitivity analysis in which the investigator varies these probabilities around the values estimated in the validation study (as in the work of Fox et al21) will provide insight into the possible bias due to the violation of this assumption.
To allow for differential misclassification, we also assumed correct specification of the logistic regression models for the misclassification probabilities. As an alternative, considering nonparametric regression functions of the covariates under study in the framework of generalized additive models or machine learning approaches22 could relax these assumptions. When the size of the validation data set is small, parametric models may be preferable to more flexible approaches. Small validation sample sizes may limit the number of factors by which misclassification probabilities may be allowed to vary. If logistic regression is used to estimate the misclassification probabilities in the validation data, one could use the rule of thumb that from five to ten events are needed per predictor included in the model.23,24
In settings where the size of the validation data set is very small, estimation of a and b may be difficult, and, as a result, the asymptotic distribution of and may be unreliable or unavailable. In these settings, one may stabilize inference by applying a penalty to each of the parameters in the models used to estimate for and (eg, as described in the works of Cole et al25 and Greenland and Mansournia26) or pursue Bayesian approaches to stabilization.27
Finally, we did not consider the potential for mismeasurement of the event time in addition to, or instead of, misclassification of event type. Such mismeasurement could result in censoring participants who in fact experienced one of the events of interest during the study. This is an important topic that has been addressed in a limited number of settings (eg, see the works of Meier et al28 and Korn et al29) and is an interesting area of future research.
ACKNOWLEDGEMENTS
Research reported in this publication was supported by the National Institute Of Allergy And Infectious Diseases (NIAID) under award numbers K01AI125087 and R21AI145662. Research was also supported by NIAID, Eunice Kennedy Shriver National Institute Of Child Health and Human Development (NICHD), National Institute On Drug Abuse (NIDA), National Cancer Institute (NCI), and the National Institute of Mental Health (NIMH), in accordance with the regulatory requirements of the National Institutes of Health under award number U01AI069911 East Africa IeDEA Consortium. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. This research has also been supported by the National Institutes of Health under grant R01-AI102710 “Statistical Designs and Methods for Double-Sampling for HIV/AIDS” and, by the President’s Emergency Plan for AIDS Relief (PEPFAR) through USAID under the terms of cooperative agreement number AID-623-A-12-0001, it is made possible through joint support of the United States Agency for International Development (USAID). The contents of this journal article are the sole responsibility of AMPATH and do not necessarily reflect the views of USAID or the United States Government.
APPENDIX
PROOF OF CONSISTENCY
In this appendix, we provide the proofs of uniform consistency for both estimators. Define ‖f(t)‖∞ = supt∈[0,τ]|f(t)|. We first consider the case of nondifferential misclassification, ie, estimator . It follows that
where the second term in the right side of the aforementioned inequality converges to 0 in probability.7 Denote
and
Then, the first term becomes
where is equal to wj, j = 1, 2, with a and b replaced by their consistent estimates and . By the consistency of and and Slutsky’s theorem, it follows that , and thus, . Consequently,
and thus, the proof of uniform consistency of over [0, τ] is complete.
For the case of differential misclassification, we first need to prove the consistency of . As before
| (A1) |
where the second term in the right side of the aforementioned inequality converges to 0 in probability.14 Next, it can be easily shown that
due to the Lipschitz continuity of the exponential function. Note that Kj < ∞, j = 1, 2, by the usual regularity condition of boundedness of both the logistic regression parameters and the covariate vector W. Therefore, the first term in the right side of inequality (A1) is
where , j = 1, 2, is the nonparametric Nelson-Aalen estimator of the cumulative cause-specific hazard for the jth cause of failure at the maximum follow-up time τ, based on the misclassified counting processes , i = 1, …, n. Now, the usual regularity conditions that , which means that the expected number of subjects at risk at the maximum follow-up time is positive and the boundedness of the cumulative cause-specific hazard at the maximum follow-up time, along with the consistency of and in probability lead to , and thus,
| (A2) |
The same arguments can also be used to show that . Now, consider the competing risks model studied in this manuscript as the special case of a nonhomogeneous continuous time Markov process with one initial state and two absorbing (ie, terminal) states 1 and 2. The transition probability matrix of this process is
The Aalen-Johansen estimator of this matrix can be expressed as a product integral of the form30
where
with , j = 1, 2, being the Nelson-Aalen estimator of the cumulative cause-specific hazard when there is no misclassification in the cause of failure. Moreover, it can be shown that the product-integral estimator is equivalent to30
This means that our estimator under differential misclassification can be expressed as
where
By the uniform consistency (A2) of the components of and a continuity result from the Duhamel equation,30 it follows that
uniformly over [0, τ], and therefore,
Footnotes
DATA AVAILABILITY STATEMENT
Data are available after a concept sheet review process from IeDEA East Africa (https://www.iedea-ea.org/).
REFERENCES
- 1.Cole SR, Hudgens MG, Brookhart MA, Westreich D. Risk. Am J Epidemiol. 2015;181(4):246–250. https://doi.org/1093/aje/kwv001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Benichou J, Gail MH. Estimates of absolute cause-specific risk in cohort studies. Biometrics. 1990;46(3):813–826. [PubMed] [Google Scholar]
- 3.Lu K, Tsiatis AA. Multiple imputation methods for estimating regression coefficients in the competing risks model with missing cause of failure. Biometrics. 2001;57(4):1191–1197. [DOI] [PubMed] [Google Scholar]
- 4.Bakoyannis G, Siannis F, Touloumi G. Modelling competing risks data with missing cause of failure. Statist Med. 2010;29(30):3172–3185. [DOI] [PubMed] [Google Scholar]
- 5.Lee M, Dignam JJ, Han J. Multiple imputation methods for nonparametric inference on cumulative incidence with missing cause of failure. Statist Med. 2014;33(26):4605–4626. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Bakoyannis G, Zhang Y, Yiannoutsos CT. Nonparametric inference for Markov processes with missing absorbing state. Statistica Sinica. 2019;29(4):2083–2104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Bakoyannis G, Yiannoutsos CT. Impact of and correction for outcome misclassification in cumulative incidence estimation. PlOS One. 2015;10(9):e0137454 https://doi.org/1371/journal.pone.0137454 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Greenland S Bayesian perspectives for epidemiologic research: III. Bias analysis via missing-data methods. Int J Epidemiol. 2009;38(6):1662–1673. https://doi.org/1093/ije/dyp278 [DOI] [PubMed] [Google Scholar]
- 9.Chatterjee S, Chatterjee S. Estimation of misclassification probabilities by bootstrap methods. Commun Stat - Simul Comput. 1983;12(6):645–656. 10.1080/03610918308812350 [DOI] [Google Scholar]
- 10.Kalbfleisch JD, Prentice RL. The Statistical Analysis of Failure Time Data. Vol. 360 Hoboken, NJ: John Wiley & Sons; 2011. [Google Scholar]
- 11.Putter H, Fiocco M, Geskus RB. Tutorial in biostatistics: competing risks and multi-state models. Stat Medicine. 2007;26(11):2389–2430. [DOI] [PubMed] [Google Scholar]
- 12.Bakoyannis G, Touloumi G. Practical methods for competing risks data: a review. Stat Methods Med Res. 2012;21(3):257–272. [DOI] [PubMed] [Google Scholar]
- 13.Aalen OO, Johansen S. An empirical transition matrix for non-homogeneous Markov chains based on censored observations. Scand J Stat. 1978;5(3):141–150. [Google Scholar]
- 14.Ha J, Tsodikov A. Isotonic estimation of survival under a misattribution of cause of death. Lifetime Data Anal. 2012;18(1):58–79. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Edwards JK, Cole SR, Troester MA, Richardson DB. Accounting for misclassified outcomes in binary regression models using multiple imputation with internal validation data. Am J Epidemiol. 2013;177(9). https://doi.org/1093/aje/kws340 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Egger M, Spycher BD, Sidle J, et al. Correcting mortality for loss to follow-up: a nomogram applied to antiretroviral treatment programmes in sub-Saharan Africa. PLOS Medicine. 2011;8(1):e1000390 https://doi.org/1371/journal.pmed.1000390 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Yiannoutsos CT, Johnson LF, Boulle A, et al. Estimated mortality of adult HIV-infected patients starting treatment with combination antiretroviral therapy. Sex Transm Infect. 2012;88(Suppl 2):i33–i43. https://doi.org/1136/sextrans-2012-050658 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Anderegg N, Johnson LF, Zaniewski E, et al. All-cause mortality in HIV-positive adults starting combination antiretroviral therapy: correcting for loss to follow-up. AIDS. 2017;31:S31–S40 English (US). https://doi.org/1097/QAD.0000000000001321 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Kim JK. Parametric fractional imputation for missing data analysis. Biometrika. 2011;98(1):119–132. https://doi.org/1093/biomet/asq073 [Google Scholar]
- 20.Altman DG, Bland JM. Statistics notes: diagnostic tests 2: predictive values. BMJ. 1994;309(6947):102 https://doi.org/1136/bmj.309.6947.102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Fox MP, Lash TL, Greenland S. A method to automate probabilistic sensitivity analyses of misclassified binary variables. Int J Epidemiol. 2005;34(6):1370–1376. https://doi.org/1093/ije/dyi184 [DOI] [PubMed] [Google Scholar]
- 22.Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. 2nd ed. New York, NY: Springer; 2009. Springer Series in Statistics. [Google Scholar]
- 23.Peduzzi P, Concato J, Kemper E, Holford TR, Feinstein AR. A simulation study of the number of events per variable in logistic regression analysis. J Clin Epidemiol. 1996;49(12):1373–1379. https://doi.org/1016/S0895-4356(96)00236-3 [DOI] [PubMed] [Google Scholar]
- 24.Vittinghoff E, McCulloch CE. Relaxing the rule of ten events per variable in logistic and Cox regression. Am J Epidemiol. 2007;165(6):710–718. https://doi.org/1093/aje/kwk052 [DOI] [PubMed] [Google Scholar]
- 25.Cole SR, Chu H, Greenland S. Maximum likelihood, profile likelihood, and penalized likelihood: a primer. Am J Epidemiol. 2014;179(2):252–260. https://doi.org/1093/aje/kwt245 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Greenland S, Mansournia MA. Penalization, bias reduction, and default priors in logistic and related categorical and survival regressions. Statist Med. 2015;34(23):3133–3143. https://doi.org/1002/sim.6537 [DOI] [PubMed] [Google Scholar]
- 27.Hamra GB, MacLehose RF, Cole SR. Sensitivity analyses for sparse-data problems-using weakly informative Bayesian priors. Epidemiology 2013;24(2):233–239. https://doi.org/1097/EDE.0b013e318280db1d [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Meier AS, Richardson BA, Hughes JP. Discrete proportional hazards models for mismeasured outcomes. Biometrics. 2003;59(4):947–954. https://doi.org/1111/j.0006-341X.2003.00109.x [DOI] [PubMed] [Google Scholar]
- 29.Korn EL, Dodd LE, Freidlin B. Measurement error in the timing of events: effect on survival analyses in randomized clinical trials. Clinical Trials. 2010;7(6):626–633. https://doi.org/1177/1740774510382801 [DOI] [PubMed] [Google Scholar]
- 30.Andersen PK, Borgan O, Gill RD, Keiding N. Statistical Models Based on Counting Processes . New York, NY: Springer Science & Business Media; 2012. [Google Scholar]


