Skip to main content
Biometrika logoLink to Biometrika
. 2011 Apr 28;98(2):325–340. doi: 10.1093/biomet/asq083

Nonparametric inference for competing risks current status data with continuous, discrete or grouped observation times

M H Maathuis 1, M G Hudgens 2
PMCID: PMC3372275  PMID: 22822257

Abstract

New methods and theory have recently been developed to nonparametrically estimate cumulative incidence functions for competing risks survival data subject to current status censoring. In particular, the limiting distribution of the nonparametric maximum likelihood estimator and a simplified naive estimator have been established under certain smoothness conditions. In this paper, we establish the large-sample behaviour of these estimators in two additional models, namely when the observation time distribution has discrete support and when the observation times are grouped. These asymptotic results are applied to the construction of confidence intervals in the three different models. The methods are illustrated on two datasets regarding the cumulative incidence of different types of menopause from a cross-sectional sample of women in the United States and subtype-specific HIV infection from a sero-prevalence study in injecting drug users in Thailand.

Keywords: Competing risk, Confidence interval, Current status data, Interval censoring, Nonparametric maximum likelihood estimator, Survival analysis

1. Introduction

Current status data with competing risks arise in cross-sectional studies that assess the current status of individuals in the sample with respect to an event that can be caused by several mechanisms. An example is Cycle I of the Health Examination Survey in the United States (MacMahon & Worcestor, 1966). This study recorded the age and menopausal status of the female participants, where menopausal status could be pre-menopausal, post-menopausal due to an operation or post-menopausal due to natural causes. Based on these data, the cumulative incidence of natural and operative menopause can be estimated as a function of age. A second example is the Bangkok Metropolitan Administration injecting drug users cohort study (Kitayaporn et al., 1998; Vanichseni et al., 2001). This study recorded the age and HIV status of injecting drug users, where HIV status could be HIV negative, HIV positive with subtype B, HIV positive with subtype E or HIV positive with some other subtype. Based on these data, the subtype-specific cumulative incidence of HIV can be estimated as a function of age.

New methods and theory have recently been developed to nonparametrically estimate cumulative incidence functions based on current status data with competing risks. Hudgens et al. (2001) and Jewell et al. (2003) derived and studied the nonparametric maximum likelihood approach and also introduced several other estimators, including the so-called naive estimator of Jewell et al. (2003). Groeneboom et al. (2008b,c) and M.H. Maathuis, in an unpublished 2006 University of Washington PhD thesis, available at http://stat.ethz.ch/~maathuis/papers, derived the large-sample behaviour of the maximum likelihood estimator and the naive estimator in a smooth model that imposes certain smoothness conditions on the cumulative incidence functions and the observation time distribution. In this model, the local rate of convergence of the maximum likelihood estimator is n1/3 (Groeneboom et al., 2008b, Theorem 4.17), slower than the usual n1/2 rate. Moreover, its limiting distribution is nonstandard and involves a self-induced system of slopes of convex minorants of Brownian motion processes plus parabolic drifts (Groeneboom et al., 2008c, Theorems 1.7 and 1.8). The naive estimator has the same local rate of convergence as the maximum likelihood estimator, but its limiting distribution is simpler, since it does not involve a self-induced system (Groeneboom et al., 2008c, Theorem 1.6).

In practice, recorded observation times are often discrete, making the smooth model unsuitable. We therefore study the large sample behaviour of the maximum likelihood estimator and the naive estimator in two additional models: a discrete model in which the observation time distribution has discrete support, and a grouped model in which the observation times are assumed to be rounded in the recording process, yielding grouped observation times.

We show that the large sample behaviour of the estimators in the discrete model is fundamentally different from that in the smooth model: the maximum likelihood estimator and the naive estimator converge locally at rate n1/2, and their limiting distributions are identical and normal. These results are related to the work of Yu et al. (1998), who studied the asymptotic behaviour of the maximum likelihood estimator for current status data with discrete observation times in the absence of competing risks.

The grouped model is related to the work of Woodroofe & Zhang (1999) and Zhang et al. (2001), who considered the maximum likelihood estimator for a nonincreasing density when the observations are grouped. We are not aware, however, of any work on the maximum likelihood estimator for interval censored data with grouped observation times, even though such grouping frequently occurs in practice. For example, in the menopause data the ages of the women were grouped in the intervals 25–30, 30–35, 35–36, 36–37, …, 58–59 and recorded as the midpoints of these intervals. The menopausal status, on the other hand, was determined at the exact but unrecorded time of interview, yielding a mismatch between the recorded status and the recorded observation time. For example, if a 30.7-year old pre-menopausal woman is interviewed, she is recorded as pre-menopausal with rounded age 32.5. When ignoring the rounding, as done in previous analyses of these data, this is taken to mean that she was interviewed at age 32.5 and that she was pre-menopausal at that age. A correct interpretation of the data is, however, that she was pre-menopausal at some unknown age in the interval (30, 35]. In particular, the data do not reveal her menopausal status at age 32.5; in actuality, she might have been post-menopausal at that age, for example due to an operation.

The grouped model accounts for such grouping of observation times. We show that the likelihood in this model can be written in the same form as in the discrete model, but in terms of different parameters, representing weighted averages of the cumulative incidence functions over the grouping intervals, where the weights are determined by the observation time distribution. This similarity with the discrete model implies that the maximum likelihood estimator and the naive estimator in the grouped model can be computed with existing software, and that their limiting distributions can be derived as in the discrete model. However, since the likelihood is written in terms of different parameters, the estimates under the grouped model must be interpreted differently. The ideas incorporated in the grouped model can be easily extended to other forms of interval censored data.

The asymptotic results in the three models are applied to the construction of confidence intervals, a problem that has received little attention until now. In the discrete and grouped models, confidence intervals can be constructed by standard methods, for example using the bootstrap or the limiting distributions derived in this paper. In the smooth model, the nonstandard limiting behaviour of the estimators makes the construction of confidence intervals less straightforward. In this case, we advocate using likelihood ratio confidence intervals (Banerjee & Wellner, 2001) based on the naive estimator.

2. Models

2.1. Exact observation times

Consider the usual competing risks setting where an event can be caused by K competing risks, with K ∈ {1, 2, …} fixed. The random variables of interest are (X, Y), where X ∈ ℝ is the time of the event of interest, and Y ∈ {1, …, K} is the corresponding cause. The goal is to estimate the cumulative incidence functions F0 = (F01, …, F0K), where F0k(t) = pr(Xt, Y = k) for k = 1, …, K. The cumulative incidence functions are nonnegative, monotone nondecreasing, and satisfy k=1K F0k(t)=pr(Xt)1.

The difficulty in estimating the cumulative incidence functions is that we cannot observe (X, Y) directly. Rather, we observe the current status of a subject at a single random observation time C ∈ ℝ. Thus, at time C we observe whether or not the event of interest has occurred, and if and only if the event has occurred, we also observe the cause Y. We assume that C is independent of (X, Y). Let G denote the distribution of C, and let (C, Δ) denote the observed data, where Δ = (Δ1, …, ΔK + 1) is an indicator vector for the status of the subject at time C:

Δk=1{XC,Y=k}(k=1,,K),ΔK+1=1{X>C}. (1)

To make this concrete, consider the HIV data discussed in § 1, where X is the age at HIV infection, C is the age at screening and there are K = 3 competing risks representing the HIV subtypes: Y = 1 for subtype B, Y = 2 for subtype E and Y = 3 for other subtypes.

We consider the maximum likelihood estimator for F0 based on n independent and identically distributed observations of (C, Δ), denoted by (Ci, Δi) (i = 1, …, n), where Δi=(Δ1i,,ΔK+1i). For any K-tuple (x1, …, xK) let x+=k=1Kxk and, unless otherwise defined, let xK + 1 = 1 − x+. Moreover, define the set 𝒡K = {F = (F1, …, FK) : F1, …, FK are cumulative incidence functions and F+(t) ⩽ 1 for all t ∈ ℝ}. A maximum likelihood estimator for F0 is defined as any n = (n1, …, nK) ∈ 𝒡K satisfying ln(n) = maxF∈𝒡K ln(F), where ln(F) is the loglikelihood

ln(F)=1ni=1nk=1K+1Δkilog{Fk(Ci)}, (2)

with the convention 0 log 0 = 0; see also Jewell et al. (2003), equation (1).

We also consider the naive estimator n = (n1, …, nK) of Jewell et al. (2003), whose kth component is defined as any nk𝒡1 satisfying lnk(nk) = max Fk∈𝒡1 lnk(Fk), where

lnk(Fk)=1ni=1n[Δkilog{Fk(Ci)}+(1Δki)log{1Fk(Ci)}] (3)

is the marginal loglikelihood for the reduced current status data (Ci,Δki) (i = 1, …, n), and 𝒡1 is obtained from 𝒡K by taking K = 1. Since nk only uses the kth entry of the Δ-vector, the naive estimator splits the estimation problem into K well-known univariate current status problems. Therefore, its computation and asymptotic theory follow straightforwardly from known results on current status data. But this simplification comes at a cost. For example, n+ need not be bounded by unity, and the naive estimator has been empirically shown to be less efficient than the maximum likelihood estimator in the smooth model (Groeneboom et al., 2008c).

The R-package MLEcens provides an efficient and stable method to compute the maximum likelihood estimator. This algorithm first uses the Height Map Algorithm of Maathuis (2005) to compute the areas to which the maximum likelihood estimator can possibly assign probability mass, called maximal intersections. Next, it computes the amounts of mass that must be assigned to the maximal intersections. This involves solving a high-dimensional convex optimization problem, which is done using the support reduction algorithm of Groeneboom et al. (2008a). Jewell & Kalbfleisch (2004) describe an alternative algorithm for the computation of the maximum likelihood estimator, based on the pool adjacent violators algorithm of Ayer et al. (1955).

The maximum likelihood estimator and the naive estimator are not defined uniquely at all times. Gentleman & Vandal (2002) defined two types of nonuniqueness for estimators based on censored data: mixture nonuniqueness and representational nonuniqueness. Mixture nonuniqueness occurs when the probability masses assigned to the maximal intersections are nonunique. Representational nonuniqueness refers to the fact that the estimator is indifferent to the distribution of mass within the maximal intersections. The maximum likelihood estimator for current status data with competing risks is always mixture unique, see Theorem 2.20 in the thesis by M.H. Maathuis, and mixture uniqueness of the naive estimator follows as a special case of this. One can account for representational nonuniqueness of the estimators by providing a lower bound that assigns all mass to the right endpoints of the maximal intersections, and an upper bound that assigns all mass to the left endpoints of the maximal intersections.

2.2. Exact observation times with discrete support

Section 2.1 does not impose any assumptions on the observation time distribution G, and hence is valid for both continuous and discrete observation times. However, the formulas can be simplified when G is discrete. In this case, let G({s}) denote the point mass of G at s, and let 𝒮 = {s ∈ ℝ : G({s}) > 0} denote the support of G, where 𝒮 is countable but possibly infinite. Defining

Nk(s)=1ni=1nΔki1{Ci=s}(k=1,,K+1),s𝒮,

and N(s)=k=1K+1Nk(s), the loglikelihood (2) reduces to

ln(F)=s𝒮k=1K+1Nk(s)log{Fk(s)}, (4)

and the marginal loglikelihood (3) for the naive estimator becomes

lnk(Fk)=s𝒮[Nk(s),log{Fk(s)}+{N(s)Nk(s)}log{1Fk(s)}].

The spaces 𝒡K and 𝒡1 can also be simplified, as the nonnegativity, monotonicity and boundedness constraints only need to hold at points s𝒮.

2.3. Grouped observation times

In many applications, only rounded versions of the observation times are recorded, yielding grouped observation times. We introduce a new model for this type of data, called the grouped model. For any interval I on the real line, define G(I) = ∫c∈IdG(c). Let be a countable but possibly infinite set of mutually exclusive intervals such that G(I) > 0 for all I. For each I, let m(I) denote a unique point in the interval, for example its midpoint, and let 𝒨 = {m(I) ∈ ℝ : I}. For each m𝒨, let I (m) denote the corresponding interval in .

The observation scheme in the grouped model is as follows. As before, the current status of a subject is assessed at a single random time C ∈ ℝ, where C is independent of (X, Y). The difference is that we no longer observe C. Instead, all observation times falling into interval I are grouped and rounded to m(I). Thus, the observed data are (D, Δ), where D = ∑I∈ℐm(I)1{C ∈ I} is the rounded version of C, and Δ is the indicator vector corresponding to the status of the subject at the exact time C, as defined in (1). We study the maximum likelihood estimator and the naive estimator based on n independent and identically distributed observations of (D, Δ), which we denote by (Di, Δi) (i = 1, …, n).

To derive the likelihood in the grouped model, we compute pr(D = d, Δ = δ) for d𝒨 and δ ∈ {e1, …, eK+1}, where ek is the unit vector in ℝK+1 with a 1 at the kth entry. Conditioning on the exact observation time C yields

pr(D=d,Δ=δ)=pr(D=d,Δ=δ|C=c)dG(c)=cI(d)pr(Δ=δ|C=c)dG(c)=k=1K+1{cI(d)F0k(c)dG(c)}δk=G{I(d)}k=1K+1[H0k{I(d)}]δk, (5)

where

H0k{I(d)}=[G{I(d)}]1cI(d)F0k(c)dG(c)(k=1,,K),

and H0,K+1{I (d)} = 1 − H0+{I (d)} are weighted averages of F01, …, F0,K + 1 over I (d) with weights determined by G. It is convenient to work with these weighted averages, as they must obey the same constraints as the cumulative incidence functions. More precisely, considering H0k (k = 1, …, K), as functions that maps m to H0k{I (m)}, the constraints on F01, …, F0K imply that H01, …, H0K must be nonnegative and nondecreasing and satisfy H0+{I (m)} ⩽ 1 for all m𝒨. Let K denote the space of such allowable K-tuples (H1, …, HK).

The term G{I (d)} in the right-hand side of (5) can be dropped from the likelihood, as it does not depend on F. Hence, a maximum likelihood estimator or H0 = (H01, …, H0K) is defined as any ĤnK satisfying lngroup(H^n)=maxHKlngroup(H), where

lngroup(H)=1ni=1nk=1K+1Δkilog[Hk{I(Di)}]. (6)

Expression (6) has the same form as (2), but with Fk(Ci) replaced by the weighted average Hk{I (Di)}. As in the discrete model, (6) can be simplified further:

lngroup(H)=Ik=1K+1Mk(I)log{Hk(I)}, (7)

where

Mk(I)=1ni=1nΔki1{Di=m(I)}(k=1,,K+1),I.

Since the loglikelihood (7) has the same form as (4), and also the constraints on the maximization problems for the discrete and grouped models are equivalent, the maximum likelihood estimator in the grouped model can be computed with existing software. Moreover, its asymptotic theory follows straightforwardly from the theory for the discrete model. The important difference between the two models is, however, that the resulting estimates must be interpreted differently. In the discrete model, one estimates the cumulative incidence functions at points s𝒮. In the grouped model, the cumulative incidence functions are unidentifiable in general, and one estimates the weighted averages of the cumulative incidence functions over intervals I.

The naive estimator n in the grouped model can be derived analogously. Defining M(I)=k=1K+1Mk(I), I, the marginal loglikelihood for the kth component is

lnkgroup(Hk)=I[Mk(I)log{Hk(I)}+{M(I)Mk(I)}log{1Hk(I)}],

and nk1 is defined by lnkgroup(H˜nk)=maxHk1lnkgroup(Hk).

Remark 1. In general, F0k(m) ≠ H0k{I (m)}, but equality can occur in special situations. For example, F0k(m) = H0k{I (m)} if

  1. F0k is constant on I (m), or

  2. both F0k and G are linear on I (m) and m is the midpoint of I (m), or

  3. the only mass of G on I (m) consists of a point mass at m.

Condition (iii) shows that the grouped model generalizes the discrete model.

3. Local asymptotics of the estimators

3.1. Strong consistency in the discrete and grouped models

The maximum likelihood estimator and the naive estimator are Hellinger consistent when the observation times are recorded exactly for any observation time distribution G, see Theorem 4.6 in the thesis by M. H. Maathuis. Using the equivalence between Hellinger distance and total variation distance, this implies consistency in total variation, see Corollary 4.7 in the thesis by M. H. Maathuis, which in turn implies strong pointwise consistency at all points s𝒮 in the discrete model, as stated in Theorem 1.

Theorem 1. In the discrete model, F̂nk(s) → F0k(s) and F̃nk(s) → F0k(s) almost surely as n → ∞ for all s𝒮.

Since the form of the loglikelihood and the constraints on the allowable functions are identical in the discrete and grouped models, the proofs for the discrete model carry over directly to the grouped model. This leads to Theorem 2, which we give without proof.

Theorem 2. In the grouped model, Ĥnk(I) → H0k(I) and H̃nk(I) → H0k(I) almost surely as n → ∞ for all I.

3.2. Limiting distributions in the discrete model

Denote the infimum and supremum of 𝒮 by sinf and ssup, and define

s=sup{x𝒮:x<s}for s𝒮with ssinf,s+=inf{x𝒮:x<s}for s𝒮with sssup.

Define s𝒮 to be a regular point if F0k(s) = 0 for all k = 1, …, K or the following two conditions hold.

  • Condition 1. If ssinf then s𝒮 and for each k = 1, …, K either F0k(s) < F0k(s) or F0k(s) = 0.

  • Condition 2. If sssup then s+𝒮 and for each k = 1, …, K either F0k(s) < F0k(s+) or F0k(s) = 0.

If 𝒮 is a finite set and s𝒮 \ {sinf, ssup}, then s and s+ are simply the points directly to the left and right of s, and Conditions 1 and 2 are equivalent to requiring that for each k = 1, …, K either F0k(s) < F0k(s) < F0k(s+) or F0k(s) = 0. As a second example, suppose that 𝒮 is the set of rational numbers. Then for any point s𝒮 we have s ∉ {sinf, ssup} and s = s = s+. Hence, Conditions 1 and 2 are only satisfied if F0k(s) = 0 for all k = 1, …, K. Yu et al. (1998) introduced regular points in the current status model without competing risks. Our definition generalizes theirs by allowing for competing risks. Moreover, we allow the parameters to be on the boundary of the parameter space. For example, s𝒮 can be a regular point when F0k(s) = 0 for some or all of the F0ks, and s = ssup can be a regular point when k=1KF0k(s)=1 or when F0k(s) = limt→∞ F0k(t) for some of the F0ks.

We now introduce the following simple estimator for F0k(s):

Fnk(s)=Nk(s)/N(s)(k=1,,K),s𝒮,

where we set 0/0 = 0. This estimator is very simple, in the sense that nk does not obey monotonicity constraints and uses only the kth component of the Δ-vector. Lemma 1 below states that n is the maximum likelihood estimator for F0 if the monotonicity constraints on the cumulative incidence functions are discarded. Next, Lemma 2 establishes that for any regular point s𝒮, n(s) = n(s) = n(s) with probability tending to one as n → ∞. Hence, at such points the limiting distributions of n(s) and n(s) equal the limiting distribution of n(s). This yields asymptotic normality of n(s) and n(s) at regular points, as stated in Theorem 3. All proofs are deferred to the Appendix.

Lemma 1. Let K* = {F= (F1,…,FK) : Fk(s) ⩾ 0 for k = 1, …, K and F+(s) ⩽ 1 for all s𝒮}. Then ln(n) ⩾ ln(F) for all FK*, and ln(n) > ln(F) for all FK* such that F(s) ≠ n(s) for some s𝒮 with N(s) > 0.

Lemma 2. For any regular point s𝒮 in the discrete model, pr{n(s) = n(s) = n(s)} → 1, as n → ∞.

Theorem 3. For any regular point s𝒮 in the discrete model,

n1/2{F^n(s)F0(s)}=n1/2(F^n1(s)F01(s)F^nK(s)F0K(s))

is asymptotically normal with zero mean and covariance matrix V (s), where V (s) is a K × K matrix with entries

{V(s)}k,=[F0k(s)1{k=}F0k(s)F0(s)]/G({s}),k,,{1,,K}.

For any finite collection of regular points s1, …, sp in 𝒮, the stacked vector n1/2{n(s1) − F0(s1), …, n(sp) − F0(sp)} is asymptotically normal with zero mean and block diagonal covariance matrix with blocks V (s1), …, V (sp). Consistent estimators for the elements of V (s), s𝒮, are

{V^n(s)}k,=[F^nk(s)1{k=}F^nk(s)F^n(s)]/N(s),k,{1,,K}.

The same results hold for the naive estimator, that is, when F̂n is replaced by F̃n.

Remark 2. If F0k(s) > 0 for all k = 1, …, K and k=1KF0k(s)=1, then the matrix V (s) is positive-semidefinite with rank K − 1. If F0k(s) = 0 or F0k(s) = 1, then the kth row and the kth column of V (s) are zero vectors, and the corresponding limiting distributions of nk(s) and nk(s) should be interpreted as degenerate distributions consisting of a point mass at zero. More details can be found in the proof of Theorem 3.

3.3. Limiting distributions in the grouped model

Denote the infimum and supremum of 𝒨 by minf and msup, and define

{m(I)}=sup{x:x<m(I)}for Iwith m(I)minf,{m(I)}+=inf{x:x>m(I)}for Iwith m(I)msup.

If {m(I)}𝒨 let I = I [{m(I)}], and if {m(I)}+𝒨 let I+ = I [{m(I)}+]. We say that I is a regular interval if H0k(I) = 0 for all k = 1, …, K or the following two conditions hold: (i) if m(I) ≠ minf then {m(I)}𝒨 and for each k = 1, …, K either H0k(I) < H0k(I) or H0k(I) = 0, and (ii) if m(I) ≠ msup then {m(I)}+𝒨 and for each k = 1, …, K either H0k(I) < H0k(I+) or H0k(I) = 0.

Analogously to n in the discrete model, we define a simple estimator in the grouped model:

Hnk(I)=Mk(I)/M(I)(k=1,,K),.

The proofs and results for the discrete model can now be translated directly to the grouped model, by replacing regular points s𝒮 by regular intervals I, n(s) by Ĥn(I), n(s) by n(I), n(s) by n(I), F0(s) by H0(I) and Nk(s) by Mk(I) for k = 1, …, K + 1. We therefore only give the main result in Theorem 4, without proof.

Theorem 4. For any regular interval Iℐ in the grouped model,

n1/2{H^n(I)H0(I)}=n1/2(H^n1(I)H01(I)H^nK(I)H0K(I))

is asymptotically normal with zero mean and covariance matrix U(I), where U(I) is a K × K matrix with entries

{U(I)}k,=[H0k(I)1{k=}H0k(I)H0(I)]/G(I),k,{1,,K}.

Moreover, for any finite collection of regular intervals I1, …, Ip, the stacked vector n1/2{Ĥn(I1) − H0(I1), …, Ĥn(Ip) − H0(Ip)} is asymptotically normal with mean vector zero and block diagonal covariance matrix with blocks U(I1), …, U(Ip). Consistent estimators for the elements of U(I), Iℐ, are

{U^n(I)}k,=[H^nk(I)1{k=}H^nk(I)H^n(I)]/M(I),k,{1,,K}.

The same results hold for the naive estimator, that is, when Ĥn is replaced by H̃n.

As in Theorem 3, a degenerate limiting distribution should be interpreted as point mass at zero.

3.4. Theoretical motivation for the grouped model

The asymptotic results provide a theoretical motivation for the grouped model, since a contradiction arises with respect to rates of convergence when the grouping of observation times is ignored. To see this, consider the menopause data and the HIV data, and suppose that the grouping of observation times is ignored, meaning that the recorded observation times are interpreted as exact observation times. This assumption was made in previous analyses of the menopause data, see Krailo & Pike (1983); Jewell et al. (2003); Jewell & Kalbfleisch (2004). Under this assumption, the discrete model is most appropriate for the menopause data, since there are numerous ties in the recorded observation times, see § 5.2. On the other hand, the smooth model seems most appropriate for the HIV data, since this dataset contains very few ties in the recorded observation times, see § 5.3. This would imply that the local rate of convergence of the maximum likelihood estimator and the naive estimator at the recorded observation times is n1/2 for the menopause data, while it is n1/3 for the HIV data.

In reality, however, the observation times were continuous in both datasets, and they were rounded in the recording process. In the menopause data, this rounding was substantial, into one-year or five-year intervals, while in the HIV data it was minimal, into one-day intervals. Since rounding implies discarding information, it seems impossible that more rounding, as in the menopause data, leads to a faster local rate of convergence at the recorded observation times. This apparent contradiction can be resolved by modelling the grouping of the observation times. For the grouped model, rounding or grouping of the observation times indeed yields a faster rate of convergence, but not for the cumulative incidence functions at the recorded observation times, but for weighted averages of the cumulative incidence functions over the grid cells. These weighted averages are smooth functionals of the cumulative incidence functions and thus can be estimated at rate n1/2, see Jewell et al. (2003) and Chapter 7 in the thesis by M. H. Maathuis.

4. Construction of pointwise confidence intervals

4.1. Confidence intervals in the discrete and grouped models

In the discrete and grouped models, the large-sample behaviour of the maximum likelihood estimator and the naive estimator at regular points or intervals is standard, and hence confidence intervals can be constructed by any standard method, for example using the asymptotic normal distribution or the bootstrap. For instance, let s𝒮 be a regular point in the discrete model. Then an asymptotic (1 − α) 100% confidence interval for F0k(s) is

F^nk(s)±n1/2z1α/2[{V^n(s)}k,k]1/2,

where z1−α/2 is the (1 − α/2)-quantile of the standard normal distribution. Similarly, considering a regular interval I in the grouped model, an asymptotic (1 − α) 100% confidence interval for H0k(I) is

H^nk(I)±n1/2z1α/2[{U^n(I)}k,k]1/2. (8)

4.2. Confidence intervals in the smooth model

In the smooth model, the large-sample behaviour of the maximum likelihood estimator and the naive estimator is nonstandard, making the construction of confidence intervals less straightforward. In principle, one can construct confidence intervals using the limiting distribution of the maximum likelihood estimator, but this approach entails several difficulties. First, the limiting distribution involves parameters from the underlying distributions that need to be estimated. Moreover, Theorems 1.7 and 1.8 of Groeneboom et al. (2008c) suggest that these parameters cannot be separated from the limiting distribution, in the sense that it seems impossible to write the limiting distribution as cZ, where c is a constant depending on the underlying distribution and Z is a universal limit. Hence, one would need to simulate the limiting distribution on a case by case basis. Conducting such simulations is nontrivial, see Groeneboom & Wellner (2001).

One might also consider the nonparametric bootstrap to construct confidence intervals based on the maximum likelihood estimator or the naive estimator. However, it is likely that the bootstrap is inconsistent in this setting, given recent results of Kosorok (2008) and Sen et al. (2010) on inconsistency of the bootstrap for the closely related Grenander estimator.

Subsampling (Politis & Romano, 1994), a variant of the bootstrap, can be applied to construct asymptotically valid confidence intervals for the cumulative incidence functions based on the maximum likelihood estimator or the naive estimator. A drawback of subsampling is that it requires a tuning parameter, the subsample size, which is difficult to choose in practice.

Finally, one can consider likelihood ratio confidence intervals based on the naive estimator. Although the naive estimator has been shown empirically to be less efficient than the maximum likelihood estimator (Groeneboom et al., 2008c, Fig. 3), it has the advantage that its large sample behaviour is simpler. For a fixed failure cause, the limiting distribution of the naive estimator is identical to the limiting distribution of the maximum likelihood estimator for current status data without competing risks (Groeneboom et al., 2008c, Theorem 1.6). Hence, the likelihood ratio theory of Banerjee & Wellner (2001) applies, and confidence intervals can be constructed by inverting likelihood ratio tests (Banerjee & Wellner, 2005). These confidence intervals have the appealing property that they do not require estimation of parameters from the underlying distribution, nor any tuning parameters. Simulation studies by Banerjee & Wellner (2005) showed that for current status data without competing risks, likelihood ratio based confidence intervals are typically preferable over confidence intervals based on the limiting distribution or subsampling. In the smooth model, we therefore recommend using likelihood ratio confidence intervals based on the naive estimator.

Fig. 3.

Fig. 3.

HIV data. The maximum likelihood estimator n (dashed) and the naive estimator n (solid) for the cumulative incidence of HIV subtypes (a) B and (b) E as a function of age, using the smooth model. The solid vertical lines represent 95% pointwise confidence intervals at times 16, …, 34, based on the likelihood ratio method for the naive estimator.

5. Examples

5.1. Simulation

It is not clear how well the asymptotic distributions of § 3.2 and § 3.3 approximate the finite sample behaviour of the estimators, especially for grids that are dense relative to n. We therefore conducted a simulation study, using the following discrete model: pr(Y = 1) = 0.6, pr(Y = 2) = 0.4, X | Y = 1 ∼ Ga(5, 3), and X | Y = 2 ∼ Ga(9, 2) where Ga(a, b) denotes a Gamma distribution with shape parameter a and scale parameter b. The distribution of C was taken to be uniform on one of the following grids: (i) {10, 20, 30}, called gap 10, (ii) {6, 8, …, 34}, called gap 2, (iii) {5.5, 6.0, …, 35.0}, called gap 0.5 and (iv) {5.1, 5.2, …, 35.0}, called gap 0.1. For each of the four resulting models, 1000 datasets of sample size n = 1000 were simulated. For each dataset we computed symmetric 95% asymptotic confidence intervals for the cumulative incidence functions at the points t0 = (10, 20, 30), based on the normal distribution and the bootstrap, using both the maximum likelihood estimator and the naive estimator.

The results for F01 are shown in Fig. 1. The results for F02 are similar, and are therefore omitted. Confidence intervals based on the maximum likelihood estimator behave very similarly to confidence intervals based on the naive estimator, while there is a large difference between normal and bootstrap based confidence intervals for the denser grids. The increase in width of the normal based confidence intervals for the denser grids is caused by the decrease of nG({t0}), which can be viewed as the expected effective sample size for the simple estimator n at t0. As a result, the variance of the asymptotic normal distribution increases by a factor of 5 or 6 between each pair of successive grids. The empirical variance of the estimators, on the other hand, increases somewhat for the denser grids, but not by much, due to the stabilizing effect of the monotonization that takes place in the maximum likelihood estimator and the naive estimator. As a result, the normal based confidence intervals give substantial overcoverage. This breakdown of the normal limit is already apparent for the larger time-points in the relatively coarse grid, gap 2, which has an average of 67 observation times per grid point. The bootstrap variance was found to be a better approximation of the empirical variance of the estimators, suggesting the use of bootstrap intervals over asymptotic normal intervals in practice. However, the undercoverage of the bootstrap intervals at t0 = 10 becomes more substantial as the grids become denser. This points to inconsistency of the bootstrap for very dense grids, which is in line with the theory discussed in § 4.2.

Fig. 1.

Fig. 1.

Simulation: Coverage (a)–(d) and average width (e)–(h) of the four 95% confidence intervals for F01(t0) as a function of t0, corresponding to grids gap 10 {(a) and (e)}, gap 2 {(b) and (f)}, gap 0.5 {(c) and (g)}, and gap 0.1 {(d) and (h)}. The confidence intervals were based on the normal distribution (○) and the bootstrap (Δ), using the maximum likelihood estimator (solid line) and the naive estimator (dashed line). The bootstrap confidence intervals are based on 750 bootstrap samples.

5.2. Menopause data

We consider data on 2423 women in the age range 25–59 years from Cycle I of the Health Examination Survey of the National Center for Health Statistics (MacMahon & Worcestor, 1966). Among other things, these women were asked to report: their current age; whether they were pre- or postmenopausal; and if they were postmenopausal, the age and cause of menopause, where the cause could be natural or operative. Since MacMahon & Worcestor (1966) found marked terminal digit clustering in the reported ages of menopause, Krailo & Pike (1983) excluded these from the analysis. The remaining information can be viewed as current status data with competing risks. Nonparametric estimates of the cumulative incidences of the two types of menopause were computed by Jewell et al. (2003) and Jewell & Kalbfleisch (2004) under the assumption that the recorded ages of the women at the time of the interview were exact. However, this was not the case. Instead, the ages were grouped into the intervals 25–30, 30–35, 35–36, 36–37 …, 58–59 and recorded as the midpoints of these intervals, yielding 26 age groups with a minimum of 45 and an average of 93 observations per age group. This is comparable to gap 2 in our simulation study, see § 5.1.

We add to the previous analyses of these data in two ways: we use the grouped model, which is clearly appropriate for these data, and we provide confidence intervals. Figure 2 shows the maximum likelihood estimator and the naive estimator for the weighted averages of the cumulative incidence functions, together with 95% normal and bootstrap confidence intervals based on the maximum likelihood estimator. As in our simulation study, the confidence intervals based on the normal distribution are wider than those based on the bootstrap.

Fig. 2.

Fig. 2.

Menopause data. The maximum likelihood estimator Ĥn (○) and the naive estimator n (×) for the weighted averages of the cumulative incidence of (a) operative and (b) natural menopause over the age groups. The estimators are plotted at the midpoints of the age groups which are indicated by the dotted vertical lines. The two solid vertical line segments in each age group are 95% asymptotic confidence intervals based on the maximum likelihood estimator: the left line segment is based on the normal approximation (8) and the right line segment is a symmetric bootstrap confidence interval based on 1000 bootstrap samples.

5.3. HIV data

The Bangkok Metropolitan Administration injecting drug users cohort study (Kitayaporn et al., 1998; Vanichseni et al., 2001) was established in 1995 to better understand HIV transmission and to assess the feasibility of conducting a phase III HIV vaccine efficacy trial in an injecting drug users population in Bangkok. We consider data on 1366 injecting drug users in this study who were screened from May to December 1996 and who were under 35 years of age. Among this group, 393 were HIV positive, with 114 infected with subtype B, 238 infected with subtype E, 5 infected by another or mixed subtype and 36 infected with missing subtype. The subjects with other, mixed or missing subtypes were grouped in a remainder category. All ages were recorded in days, leading to a small number of ties: among the 1366 subjects, there were 1212 distinct ages, and the mean number of observations per distinct age was 1.13. In light of this, we analyse these data using the smooth model. Figure 3 shows the maximum likelihood estimator and the naive estimator for the subtype-specific cumulative incidence of HIV, together with 95% likelihood ratio confidence intervals based on the naive estimator.

6. Observation time distribution or grouping dependent on n

There are interesting connections between our work and a 2011 unpublished technical report by Tang, Banerjee and Kosorok, available at http://www.stat.lsa.umich.edu/~moulib/information.html, who studied current status data without competing risks when the observation time distribution depends on the sample size n. More precisely, let X be a random event time with distribution F0 and let C(n) be a random observation time with distribution G(n), where G(n) is a discrete distribution on an equidistant grid with spacings nγ for some γ ∈ (0, 1). Without loss of generality, assume this grid is on [0, 1]. Consider the nonparametric maximum likelihood estimator n for F0 based on n independent and identically distributed observations of (C(n), Λ(n)), where Λ(n) = 1{XC(n)}. Let t0 ∈ (0, 1) be a time-point of interest, and let tn be the largest support point of G(n) smaller than t0. Assuming F0 satisfies certain smoothness conditions in a neighbourhood of t0, Tang, Banerjee and Kosorok found that the limiting distribution of the maximum likelihood estimator depends crucially on γ. For γ < 1/3 the limiting distribution of n(1−γ)/2{n(tn) − F0(tn)} is normal with zero mean and variance F0(t0){1 − F0(t0)}. Hence, for such sparse grids, the maximum likelihood estimator behaves as in the discrete model, up to a different rate of convergence. For γ > 1/3, on the other hand, the limiting distribution of n1/3{n(t0) − F0(t0)} is determined by the slope of the convex minorant of a Brownian motion process plus parabolic drift, showing that the maximum likelihood estimator behaves as in the smooth model. The case γ = 1/3 forms the boundary between these two scenarios and yields a new limiting distribution.

Combining our work with that of Tang, Banerjee and Kosorok yields two extensions. First, consider a grouped model for current status data without competing risks, where the grouping intervals depend on n. More precisely, let X be an event time with distribution F0, let C be an observation time with distribution G, and let Λ = 1{XC}. Assume the support of G is [0, 1], and let n be the set of intervals formed by the grid cells of an equidistant grid on [0, 1] with spacings nγ for some γ ∈ (0, 1). Assume that the observation time C is rounded to the midpoint of the interval in which it falls, and denote this rounded observation time by D(n). One can now consider the nonparametric maximum likelihood estimator for F0 based on n independent and identically distributed copies of (D(n), Λ). Since the likelihood in this grouped model can be written in exactly the same form as the likelihood in the discrete model, and since also the constraints on the two optimization problems are equivalent, the work of Tang, Banerjee and Kosorok should carry over to this model, with the only difference that everything should be written in terms of weighted averages of F0 over the grid cells. Second, consider the discrete model for current status data with competing risks, where the support of G depends on n. Then the results of Tang, Banerjee and Kosorok should carry over to the naive estimator nk, since this estimator can be viewed as a maximum likelihood estimator based on reduced current status data without competing risks. The same holds for the naive estimator nk in the grouped model when the grouping intervals depend on n.

Acknowledgments

We thank Kachit Choopanya, Dwip Kitayaporn, Timothy D. Mastro, Philip A. Mock and Suphak Vanichseni for allowing us to use the data from the Bangkok Metropolitan Administration Injecting Drug Users cohort study. We thank Moulinath Banerjee for discussing the connections between this paper and his work with Runlong Tang and Michael Kosorok. We thank the associate editor and two anonymous referees for helpful suggestions and comments. M.G. Hudgens was supported by a grant from the National Institutes of Health, U.S.A.

Appendix

Proof of Lemma 1. Due to the absence of monotonicity constraints on K*, the maximizer of ln(F) over K* can be determined separately for each s𝒮. Thus, fix s𝒮, and define ln(F,s)=k=1K+1Nk(s)log{Fk(s)}. Moreover, define 𝒦 = {k ∈ {1, …, K + 1} : Nk (s) > 0} and 𝒦C = {1, …, K +1} \ 𝒦. First, suppose 𝒦 = ∅. Then ln(F, s) = 0 for any choice of Fk (s) (k = 1, …, K), and hence n(s) is a maximizer of ln(F, s). Next, suppose 𝒦 ≠ ∅, or equivalently, N(s) > 0. Then any maximizer of ln(F, s) subject to the constraint F+(s) ⩽ 1 must set Fk (s) = 0 for k𝒦C. Hence, for k𝒦C the maximizer is unique and equals nk(s). If |𝒦| = 1, then ln(F, s) contains only one nonzero term, and it is clear that the corresponding Fk(s) should be set to 1, which equals nk(s). If |𝒦| > 1, we define k* = max 𝒦. Then Nk* (s) = N(s) − ∑k𝒦\{k*}Nk (s) and any maximizer of ln(F, s) over K* must satisfy Fk* (s) = 1 − ∑k𝒦\{k*}Fk (s). Hence, we can write ln(F, s) = ∑k𝒦\{k*}Nk (s) log{Fk (s)} + {N(s) − ∑k𝒦\{k*}Nk (s)} log{1 − ∑k𝒦\{k*}Fk (s)}. This function is strictly concave in Fk (s) for k𝒦 \ {k*}. The unique maximizer can be determined by solving ∂ln(F, s)/∂Fk (s) = 0 for k𝒦 \ {k*}, which yields nk(s), k𝒦.

Proof of Lemma 2. Let s𝒮 be a regular point in the discrete model. We first consider the maximum likelihood estimator for the so-called basic case, where s ∉ {sinf, ssup}. Let 𝒦+ = {k ∈ {1, …, K} : F0k (s) > 0}. For k ∈ {1, …, K} \ 𝒦+, we have Nk (s) = 0. Hence, the corresponding Fk (s)s do not contribute to the likelihood and we directly obtain that the corresponding estimators satisfy nk(s) = nk(s) = 0. So we are done if 𝒦+ = ∅. Otherwise, we are left to show pr[∩k𝒦+ {nk (s) = nk (s)}] → 1 as n → ∞. Define the events

An(s)=k𝒦+{F^nk(s)<F^nk(s)<F^nk(s+)},Bn(s)=k𝒦+{K+1}{Nk(s)>0}.

The assumptions on s imply F0k (s) < F0k (s) < F0k (s+) for k𝒦+. By combining this with the consistency of n in Theorem 1, it follows that pr{An(s)}→ 1 as n → ∞. Moreover, the law of large numbers, G({s}) > 0, F0k (s) > 0 for k𝒦+ and F0+(s) < 1 imply pr{Bn(s)}→ 1 as n → ∞. Hence, pr{An(s) ∩ Bn(s)}→ 1 as n → ∞, and the proof for the basic case can be completed by showing that the event {An(s) ∩ Bn(s)} implies ∩k𝒦+{nk (s) = F̆nk (s)}. We do this using contraposition. Thus, suppose {An(s) ∩ Bn(s)} holds. This implies k* = K + 1 in the proof of Lemma 1, and it follows that nk (s), k𝒦+, is the unique solution of ∂ln(F)/∂Fk (s) = 0, k𝒦+. Now assume there is a j𝒦+ such that nj(s) ≠ nj(s). Then there must be a 𝒦+ such that ∂ln(F)/∂F(s)|n(s) ≠ 0. Let σ ∈ {−1, +1} be the sign of ∂ln(F)/∂F(s)|n(s), and define F^nnew(s)=F^n(s)+γσek¯, where ek is the unit vector in ℝK with a 1 at the kth entry. Then for γ > 0 sufficiently small, replacing n(s) by F^nnew(s) increases the loglikelihood. Moreover, this replacement does not violate the constraints of 𝒡K, as for γ > 0 sufficiently small we have F^nk¯(s)<F^nk¯new(s)<F^nk¯(s+) and F^n+new(s)<F^n+(s+)1. This shows that n cannot be the maximum likelihood estimator, which is a contradiction.

If ssinf and s = ssup, we distinguish two cases. If F0+(s) < 1, the proof of the basic case goes through with the only change that An(s) = ∩k𝒦+ {nk (s) < nk (s)} ∩ {n+(s) < 1}. If F0+(s) = 1, then NK + 1(s) = 0 and 1 − F+(s) does not contribute to the loglikelihood. Hence, the maximum likelihood estimator must satisfy n+(s) = 1 and this equals n+(s) if N(s) > 0. If K = 1, this implies pr{n(s) = n(s)} → 1 as n → ∞, so that we are done. If K > 1, we use the proof for the basic case with the following changes. We define An(s) = ∩k𝒦+{nk (s) < F̂nk (s)} and Bn(s) = ∩k𝒦+{Nk (s) > 0}. As before, we have pr{An(s) ∩ Bn(s)}→ 1 as n → ∞. We will therefore show that {An(s) ∩ Bn(s)} implies ∩k𝒦+ {nk (s) = nk(s)}, using contraposition. Thus, assume {An(s) ∩ Bn(s)} holds. This implies k* = max 𝒦+ in the proof of Lemma 1, meaning that nk, k𝒦+, are found by solving ∂ln(F)/∂Fk(s) = 0 for k𝒦+ \ {k*} and setting nk*(s) = 1 − ∑k𝒦+\{k*} nk(s). Assume nk(s) ≠ nk(s) for some k𝒦+. Then there must be a 𝒦+ \ {k*} such ∂ln(F)/∂F(s)|n(s) ≠ 0. Define σ as the sign of ∂ln(F)/∂F(s)|n(s), and define F^nnew(s)=F^n(s)+γσek¯γσek*. Then for γ > 0 sufficiently small, replacing n(s) by F^nnew(s) increases the loglikelihood. Moreover, this replacement does not violate the constraints of 𝒡K, as for γ > 0 sufficiently small we have F^nk¯(s)<F^nknew(s), F^nk*(s)<F^nk*new(s) and F^n+new(s)=F^n+(s)=1. Hence, n cannot be the maximum likelihood estimator, again a contradiction.

The proof for the maximum likelihood estimator is completed by considering two remaining special cases. If s = sinf and sssup, then the proof for the basic case goes through with the only change that An(s) = ∩k∈𝒮+{0 < F̂nk (s) < nk (s+)}. If s = sinf = ssup, then |𝒮| = 1 and monotonicity constraints do not play any role in the maximum likelihood estimator, so that n = n follows immediately.

The proof for the naive estimator follows directly from the proof for the maximum likelihood estimator by taking K = 1. To see this, let k ∈ {1, …, K} and recall that the naive estimator is the maximum likelihood estimator for the reduced current status data (Δki,Ci) (i = 1, …, n). Hence, the proof for the maximum likelihood estimator implies pr{F˜nk(s)=Fnkred(s)}1 as n → ∞, where Fnkred is the simple estimator based on the reduced data. The proof is completed by observing that Fnkred=Fnk.

Proof of Theorem 3. Because of Lemma 2, it is sufficient to derive the limiting distribution of n. Let k ∈ {1, …, K} and s𝒮. Since pr{N(s) > 0} → 1 as n → ∞, we can assume N(s) > 0. We first consider the case 0 < F0k (s) < 1. Then

n1/2{Fnk(s)F0k(s)}=N(s)1n1/2{Nk(s)F0k(s)N(s)}=N(s)1n1/2i=1n{ΔkiF0k(s)}1{Ci=s},

and the result follows from N(s) → p G({s}), the multivariate central limit theorem, and Slutsky’s lemma, e.g., Van der Vaart (1998, Lemma 2.8(iii)).

If F0k (s) = 0, then Nk (s) = 0 and hence nk (s) = 0 = F0k (s) always. Similarly, if F0k (s) = 1, we have Nk (s) = N(s) and hence nk (s) = 1 = F0k (s) whenever N(s) > 0. These results are in agreement with the theorem, since in these cases {V (s)}k,k = 0, leading to a degenerate limiting distribution that should be interpreted as a point mass at zero. It can be easily verified that the off-diagonal elements {V (s)}k, j = 0 for j ∈ {1, …, K}, jk, are also correct in these cases.

References

  1. Ayer M, Brunk HD, Ewing GM, Reid WT, Silverman E. An empirical distribution function for sampling with incomplete information. Ann Math Statist. 1955;26:641–47. [Google Scholar]
  2. Banerjee M, Wellner JA. Likelihood ratio tests for monotone functions. Ann Statist. 2001;29:1699–731. [Google Scholar]
  3. Banerjee M, Wellner JA. Confidence intervals for current status data. Scand J Statist. 2005;32:405–24. [Google Scholar]
  4. Gentleman R, Vandal AC. Nonparametric estimation of the bivariate CDF for arbitrarily censored data. Can J Statist. 2002;30:557–71. [Google Scholar]
  5. Groeneboom P, Jongbloed G, Wellner JA. The support reduction algorithm for computing non-parametric function estimates in mixture models. Scand J Statist. 2008a;35:385–99. doi: 10.1111/j.1467-9469.2007.00588.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Groeneboom P, Maathuis MH, Wellner JA. Current status data with competing risks: consistency and rates of convergence of the MLE. Ann Statist. 2008b;36:1031–63. doi: 10.1214/009053607000000983. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Groeneboom P, Maathuis MH, Wellner JA. Current status data with competing risks: limiting distribution of the MLE. Ann Statist. 2008c;36:1064–89. doi: 10.1214/009053607000000983. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Groeneboom P, Wellner JA. Computing Chernoff’s distribution. J Comp Graph Statist. 2001;10:388–400. [Google Scholar]
  9. Hudgens MG, Satten GA, Longini IM. Nonparametric maximum likelihood estimation for competing risks survival data subject to interval censoring and truncation. Biometrics. 2001;57:74–80. doi: 10.1111/j.0006-341x.2001.00074.x. [DOI] [PubMed] [Google Scholar]
  10. Jewell NP, Kalbfleisch JD. Maximum likelihood estimation of ordered multinomial parameters. Biostatistics. 2004;5:291–306. doi: 10.1093/biostatistics/5.2.291. [DOI] [PubMed] [Google Scholar]
  11. Jewell NP, Van der Laan MJ, Henneman T. Nonparametric estimation from current status data with competing risks. Biometrika. 2003;90:183–97. [Google Scholar]
  12. Kitayaporn D, Vanichseni S, Mastro TD, Raktham S, Vaniyapongs T, Des Jarlais DC, Wasi C, Young NL, Sujarita S, Heyward WL, Esparza J. Infection with HIV 1 subtypes B and E in injecting drug users screened for enrollment into a prospective cohort in Bangkok, Thailand. J Acq Immun Defic Synd Hum Retrovir. 1998;19:289–95. doi: 10.1097/00042560-199811010-00012. [DOI] [PubMed] [Google Scholar]
  13. Kosorok MR. Bootstrapping the Grenander estimator. In: Balakrishnan N, Peña EA, Silvapulle MJ, editors. Beyond Parametrics in Interdisciplinary Research: Festschrift in Honor of Professor Pranab K. Sen. Beachwood, OH: Institute of Mathematics and Statistics; 2008. pp. 282–292. Institute of Mathematics and Statistics Collection 1. IMS Collections. [Google Scholar]
  14. Krailo MD, Pike MC. Estimation of the distribution of age at natural menopause from prevalence data. Am J Epidemiol. 1983;117:356–61. doi: 10.1093/oxfordjournals.aje.a113548. [DOI] [PubMed] [Google Scholar]
  15. Maathuis MH. Reduction algorithm for the MLE for the distribution function of bivariate interval censored data. J Comp Graph Statist. 2005;14:352–62. [Google Scholar]
  16. MacMahon B, Worcestor J. Age at menopause, United States 1960–1962. National Center for Health Statist Vital Health Statist. 1966;11:161–73. [PubMed] [Google Scholar]
  17. Politis DN, Romano JP. Large sample confidence regions based on subsamples under minimal assumptions. Ann Statist. 1994;22:2031–50. [Google Scholar]
  18. Sen B, Banerjee M, Woodroofe M. Inconsistency of the bootstrap: the Grenander estimator. Ann Statist. 2010;38:1953–77. [Google Scholar]
  19. Van der Vaart AW. Asymptotic Statistics. Cambridge: Cambridge University Press; 1998. Cambridge Series in Statistical and Probabilistic Mathematics. [Google Scholar]
  20. Vanichseni S, Kitayaporn D, Mastro TD, Mock PA, Raktham S, Des Jarlais DC, Sujarita S, Srisuwanvilai L, Young NL, Wasi C, Subbarao S, Heyward WL, Esparza J, Choopanya K. Continued high HIV-1 incidence in a vaccine trial preparatory cohort of injection drug users in Bangkok, Thailand. AIDS. 2001;15:397–405. doi: 10.1097/00002030-200102160-00013. [DOI] [PubMed] [Google Scholar]
  21. Woodroofe M, Zhang R. Isotonic estimation for grouped data. Statist Prob Lett. 1999;45:41–7. [Google Scholar]
  22. Yu Q, Schick A, Li L, Wong GYC. Asymptotic properties of the GMLE in the case 1 interval-censorship model with discrete inspection times. Can J Statist. 1998;26:619–27. [Google Scholar]
  23. Zhang R, Kim J, Woodroofe M. Asymptotic analysis of isotonic estimation for grouped data. J Statist Plan Infer. 2001;98:107–17. [Google Scholar]

Articles from Biometrika are provided here courtesy of Oxford University Press

RESOURCES