Skip to main content
BMC Medical Research Methodology logoLink to BMC Medical Research Methodology
. 2023 Apr 4;23:82. doi: 10.1186/s12874-023-01903-x

A pairwise pseudo-likelihood approach for regression analysis of left-truncated failure time data with various types of censoring

Li Shao 1, Hongxi Li 1,, Shuwei Li 1, Jianguo Sun 2
PMCID: PMC10071649  PMID: 37016341

Abstract

Background

Failure time data frequently occur in many medical studies and often accompany with various types of censoring. In some applications, left truncation may occur and can induce biased sampling, which makes the practical data analysis become more complicated. The existing analysis methods for left-truncated data have some limitations in that they either focus only on a special type of censored data or fail to flexibly utilize the distribution information of the truncation times for inference. Therefore, it is essential to develop a reliable and efficient method for the analysis of left-truncated failure time data with various types of censoring.

Method

This paper concerns regression analysis of left-truncated failure time data with the proportional hazards model under various types of censoring mechanisms, including right censoring, interval censoring and a mixture of them. The proposed pairwise pseudo-likelihood estimation method is essentially built on a combination of the conditional likelihood and the pairwise likelihood that eliminates the nuisance truncation distribution function or avoids its estimation. To implement the presented method, a flexible EM algorithm is developed by utilizing the idea of self-consistent estimating equation. A main feature of the algorithm is that it involves closed-form estimators of the large-dimensional nuisance parameters and is thus computationally stable and reliable. In addition, an R package LTsurv is developed.

Results

The numerical results obtained from extensive simulation studies suggest that the proposed pairwise pseudo-likelihood method performs reasonably well in practical situations and is obviously more efficient than the conditional likelihood approach as expected. The analysis results of the MHCPS data with the proposed pairwise pseudo-likelihood method indicate that males have significantly higher risk of losing active life than females. In contrast, the conditional likelihood method recognizes this effect as non-significant, which is because the conditional likelihood method often loses some estimation efficiency compared with the proposed method.

Conclusions

The proposed method provides a general and helpful tool to conduct the Cox’s regression analysis of left-truncated failure time data under various types of censoring.

Supplementary Information

The online version contains supplementary material available at 10.1186/s12874-023-01903-x.

Keywords: Cox model, EM algorithm, Interval censoring, Left truncation, Partly interval-censored data

Introduction

Failure time data are frequently encountered in various scientific areas, including clinical trials, epidemiology surveys, and biomedical studies. A key feature of such data is the presence of censoring, which usually poses great computational challenges for their analysis [1, 2]. The type of censoring that has been investigated most is apparently right censoring [36]. Other types of censored data that often occur in practice include interval-censored and partly interval-censored data [713]. In particular, Gao et al. [10] recently proposed an efficient semiparametric estimation approach for the analysis of partly interval-censored data under the accelerated failure time model. Zhou et al. [13] also studied the analysis of partly interval-censored failure time but via the transformation models.

For failure time data, in addition to censoring, left truncation also often arises due to the use of cross-sectional sampling strategy and can substantially complicate the data analysis. For example, in the Canadian Study of Health and Aging Study, the failure time of interest is defined as the duration from the onset of dementia to death [14]. Since only dementia patients who had not experienced the death at the enrollment are included in the study, the patient’s death time is expected to suffer from left truncation, where the truncation time is the gap time between the onset of dementia and the enrollment. Therefore, the sampled patients are no longer representative of the whole population under study, and it is well-known that ignoring the left truncation in the data analysis often leads to biased parameter estimation.

Due to the ubiquity of left truncation in failure time studies, extensive efforts have been devoted to the method developments for the analysis of the left-truncated failure time data under various types of censoring scheme [1525]. For instance, Wang et al. [16] considered the left-truncated and right-censored data, and developed a conditional estimation approach under the proportional hazards (PH) model, while Pan and Chappell [17] investigated the analysis of left-truncated and interval-censored data and suggested a marginal likelihood approach and a monotone maximum likelihood approach for the PH model. Gao and Chan [24] discussed the same model and data structure as Pan and Chappell [17], but further assumed that the truncation times follow the uniform distribution, which is usually referred to as the stationary or length-biased assumption in the literature. However, it is worth noting that this approach may produce biased parameter estimation when the length-biased assumption is violated in practical applications. For the left-truncated and partly interval-censored data, Wu et al. [25] provided a conditional likelihood approach for the PH model in the presence of a cured subgroup.

In addition to the work described above, Huang and Qin [14] also studied left-truncated and right-censored data and proposed an estimation procedure for the additive hazards model by combining a pairwise pseudo-score function and the conditional estimating function. This approach is appealing since it utilizes the marginal likelihood of the truncation times and can thus improve the estimation efficiency. In addition, the employed pairwise pseudo-likelihood can eliminate nuisance parameters from the marginal likelihood of the truncation times, leading to an estimating equation function with tractable form, and can yield more efficient estimation compared with the conditional estimating equation approach. Inspired by the work of Huang and Qin [14], Wu et al. [26] proposed a pairwise likelihood augmented estimator for the PH model with the left-truncated and right-censored data. Furthermore, Wang et al. [27] considered the analysis of left-truncated and interval-censored data with the PH model, and developed a sieve maximum likelihood estimation procedure by accommodating the pairwise likelihood function of the truncation times.

In the following, we will consider regression analysis of left-truncated failure time data under the PH model and various types of censoring mechanism, including the interval censoring, right censoring and a mixture of them. Specifically, motivated by Huang and Qin [14] and Wu et al. [26], we propose a nonparametric maximum likelihood estimation (NPMLE) approach by combining the conditional likelihood of the failure times with the pairwise likelihood obtained from the marginal likelihood of the truncation times, rendering an efficient estimation for the PH model. A flexible EM algorithm that can accommodate various types of censored data will be developed to implement the NPMLE. Through the desirable data augmentation, the objective function in the M-step of the algorithm has a tractable form, and one can estimate the regression coefficients and the nuisance parameters related to the cumulative baseline hazard function separately. In particular, by utilizing the spirit of self-consistent estimation equation, we obtain the explicit estimators of the possibly large-dimensional nuisance parameters, which can greatly relieve the computational burden in the optimization procedure. The numerical results obtained from extensive simulation studies demonstrate that the proposed method is computationally stable and reliable and can improve the estimation efficiency of the conditional likelihood approach. In other words, the proposed method provides a general and helpful tool to conduct the Cox’s regression analysis of left-truncated failure time data under various types of censoring.

The remainder of this paper is organized as follows. In Section Notation, model, and likelihood, we will first introduce some notation, data structure and the model, and then present the observed data likelihood function. Section Estimation procedure presents the developed EM algorithm to implement the NPMLE. In Section Simulation studies, extensive simulation studies are conducted to evaluate the empirical performance of the proposed method, followed by an application to a set of real data in Section An application. Section Discussion and concluding remarks gives some discussion and concluding remarks.

Notation, model, and likelihood

Consider a failure time study involving left truncation, and for a subject from the target population, let T denote the underlying failure time, that is, the time to the onset of the failure event. Let A be the underlying truncation time (i.e. the time to the study enrolment), which is assumed to be independent of T, and Z be the p-dimensional vector of covariates. For a subject enrolled in the study (i.e. satisfying TA), denoted by T, A and Z the failure time, the truncation time and the vector of covariates, respectively. Then (T,A,Z) has the same joint distribution as (T,A,Z) conditional on TA.

Let f and S denote the density and survival functions of T, respectively. Let h be the density function of A. Then the joint density function of (TA) at (ta) is

f(t)h(a)0S(u)h(u)du=f(t)S(a)×S(a)h(a)0S(u)h(u)du,(0at),

where f(t)/S(a) is the conditional density of T given A, S(a)h(a)/0S(u)h(u)du is the marginal density of A. To describe the effect of Z on the failure time T, we assume that T follows the PH model with the conditional cumulative hazard function of T given Z taking the form

Λ(tZ)=Λ(t)exp(Zβ). 1

In the above, Λ(t) is an unspecified baseline cumulative hazard function and β denotes a p-dimensional vector of regression coefficients.

As mentioned above, censoring always exists in failure time studies. Define Δ=1 if T can be observed exactly and 0 otherwise. If Δ=0, let (LR] be the smallest interval that brackets T with LA. Clearly, T is left-censored if L=A, T is right-censored if R=, and T is interval-censored if R<. In the sequel, notations with the subscript i represent the corresponding sample analogues. Therefore, we have partly interval-censored data if the obtained data consist of n independent observations denoted by (Ai,Ti,Δi,Zi) if Δi=1 and (Ai,Li,Ri,Δi,Zi) if Δi=0 for i=1,,n. Notably, the data above reduce to interval-censored data if Δi=0 for i=1,,n, and right-censored data if Ri= for i=1,,n.

Let S(tZi)=exp{-Λ(t)exp(Ziβ)} and λ(t)=dΛ(t)/dt. Assume that (Li,Ri) is conditionally independent of (A,T) given AT and Z, and that A is independent of Z, the observed data likelihood function takes the form

Ln(β,Λ,h)=LnC(β,Λ)×LnM(β,Λ,h), 2

where

LnC(β,Λ)=i=1n{λ(t)exp(Ziβ)S(TiZi)}Δi{S(LiZi)-S(RiZi)}1-ΔiS(AiZi)=i=1nλ(t)exp(Ziβ)exp{-(Λ(Ti)-Λ(Ai))exp(Ziβ)}Δi×exp{-(Λ(Li)-Λ(Ai))exp(Ziβ)}-exp{-(Λ(Ri)-Λ(Ai))exp(Ziβ)}1-Δi,

and

LnM(β,Λ,h)=i=1nS(AiZi)h(Ai)0S(uZi)h(u)du.

In the above, LnC(β,Λ) is the conditional likelihood of {ΔiTi,(1-Δi)Li,(1-Δi)Ri,Δi} given (Ai,Zi), and LnM(β,Λ,h) is the marginal likelihood of Ai given Zi. Note that the observed data likelihood Ln(β,Λ,h) has an intractable form due to the complex data structure and the involvement of the nuisance functions Λ and h. For the estimation, it is apparent that performing direct maximization of Ln(β,Λ,h) with respect to all parameters is quite challenging and unstable even after approximating Λ and h with some smooth functions with finite-dimensional parameters. To address this issue, in the next section, we will develop a flexible EM algorithm by introducing some Poisson latent variables in the data augmentation procedure, which can greatly simplify the form of LnC(β,Λ). In addition, by following Liang and Qin [28] and others, we will employ the pairwise likelihood approach to eliminate the nuisance function h from the marginal likelihood LnM(β,Λ,h). The above two manipulations make the estimation procedure appealing and easily implemented.

Estimation procedure

To estimate β and Λ, we adopt the NPMLE approach and develop an EM algorithm for its implementation. For this, we will first discuss the data augmentation and then present the pairwise likelihood method as well as the E-step and M-step of the algorithm.

Data augmentation

First note that the likelihood function above depends on Λ(t) only through its values at the finite observation times, exactly-observed failure times and truncation times. Let t1<tKn< denote the ordered sequence of these unique time points, and assume that Λ(t) is a step function at tk with the non-negative jump size λk for k=1,,Kn. Then the conditional likelihood LnC(β,Λ) can be re-expressed as

L1nC(θ)=i=1nk=1KnλkI(Ti=tk)exp(Ziβ)exp-AitkTiλkexp(Ziβ)Δi×exp-AitkLiλkexp(Ziβ)-I(Ri<)exp-AitkRiλkexp(Ziβ)1-Δi,

where θ=(β,λ1,,λKn).

To simplify L1nC(θ), for the ith subject, we introduce a set of new independent latent variables {Wik;k=1,2,,Kn} relating to t1,t2,,tKn respectively, where Wik is a Poisson random variable with mean λkexp(Ziβ). Then L1nC(θ) can be equivalently expressed as

L2nC(θ)=i=1nPAitk<TiWik=0PWik|tk=Ti=1Δi×PAitkLiWik=0PLi<tkRiWik>0I(Ri<)1-Δi,

where Wik|tk=Ti denotes the variable in {Wik;k=1,2,,Kn} that satisfies tk=Ti.

Define Ri=(1-Δi)(LiI(Ri=)+RiI(Ri<))+ΔiTi, and let p{Wikλkexp(Ziβ)} be the probability mass function of Wik with mean λkexp(Ziβ). By treating the latent variables Wik’s as observable, the augmented likelihood function is given by

LC(θ)=i=1nk=1Knp{Wikλkexp(Ziβ)}I(AitkRi)=i=1nk=1Kn{λkexp(Ziβ)}WikWik!exp{-λkexp(Ziβ)}I(AitkRi),

which subjects to the constraints that Aitk<TiWik=0 and Wik|Ti=tk=1 if Δi=1, AitkLiWik=0 and Li<tkRiWik>0 if Δi=0 and Ri<; and AitkLiWik=0 if Δi=0 and Ri=.

Pairwise likelihood

Since the density function h in the marginal likelihood LnM(β,Λ,h) is a nuisance function, we follow the work of Liang and Qin [28] and apply the pairwise likelihood method to LnM(β,Λ,h) to eliminate h. Note that, for ij, by conditioning on (Zi,Zj) and having observed (Ai,Aj) but without knowing the order of Ai and Aj, the pairwise pseudo-likelihood of the observed (Ai,Aj) is given by

S(AiZi)h(Ai)0S(aZi)h(a)da×S(AjZj)h(Aj)0S(aZj)h(a)daS(AiZi)h(Ai)0S(aZi)h(a)da×S(AjZj)h(Aj)0S(aZj)h(a)da+S(AiZj)h(Ai)0S(aZj)h(a)da×S(AjZi)h(Aj)0S(aZi)h(a)da=11+Rij(θ),

where

Rij(θ)=S(AiZj)S(AjZi)S(AiZi)S(AjZj)=expk=1KnI(tkAi)-I(tkAj)λkexp(Ziβ)-exp(Zjβ).

Therefore, the pairwise likelihood LnP(θ) of all pairs is given by

LP(θ)=ij{1+Rij(θ)}-1.

Notably, through the above manipulation, LP(θ) depends on the parameters in the survival model, β and λ1,,λKn, but not on the density function h of truncation time A.

EM algorithm

Combing the augmented likelihood LC(θ) with the pairwise likelihood LP(θ), and taking into account the different magnitudes of LC(θ) and LP(θ), we can derive the composite complete-data log-likelihood as follows

l(θ)=1ni=1nk=1KnI(AitkRi)Wiklog{λkexp(Ziβ)}-λkexp(Ziβ)-log(Wik!)-1n(n-1)ijlog{1+Rij(θ)}.

In the E-step of the algorithm, we take the conditional expectations with respect to the latent variables Wik’s in l(θ), and for notational simplicity, we will ignore the conditional arguments including the observed data and the estimate of θ at the lth iteration denoted by θ(l) in all conditional expectations. This step yields the following objective function

lE(θ)=1ni=1nk=1KnI(AitkRi)[E(Wik)log{λkexp(Ziβ)}-λkexp(Ziβ)]-1n(n-1)ijlog{1+Rij(θ)}.

We now present the expressions of E(Wik)’s in lE(θ). Specifically, in the case of Δi=1 (exactly-observed Ti), we have E(Wik)=0 if Aitk<Ti, and E(Wik)=1 if Ti=tk. In the case of Δi=0 and AiTiLi (left censoring), we have

E(Wik)=λk(l)exp(Ziβ(l))1-exp{-AitkLiλk(l)exp(Ziβ(l))},ifAitkLi.

In the case of Δi=0 and Ri< (interval censoring), we have E(Wik)=0 if AitkLi, and

E(Wik)=λk(l)exp(Ziβ(l))1-exp{-Li<tkRiλk(l)exp(Ziβ(l))},ifLi<tkRi.

In the case of Δi=0 and Ri= (right censoring), we have E(Wik)=0 if AitkLi.

Differentiating lE(θ) with respect to β and λk’s yields the following composite score functions

Uβ(θ)=1ni=1nk=1KnI(AitkRi)Zi{E(Wik)-λkexp(Ziβ)}-1n(n-1)ijk=1KnλkQij(1)(tk)1+Rij-1(θ),

and

Uλk(θ)=1ni=1nI(AitkRi)E(Wik)λk-exp(Ziβ)-1n(n-1)ijQij(0)(tk)1+Rij-1(θ),

where Qij(m)(t;β)={Zimexp(Ziβ)-Zjmexp(Zjβ)}{I(tAi)-I(tAj)} for m=0 or 1, Z0=1 and Z1=Z.

Specifically, at the (l+1)th iteration, based on estimating equation Uλk(θ)=0, one can derive a self-consistent solution to update each λk :

λk(l+1)=1ni=1nI(AitkRi)E(Wik)1ni=1nI(AitkRi)exp(Ziβ(l))+1n(n-1)ijQij(0)(tk;β(l))1+1/Rij(θ(l)). 3

By combining the discussion above, the proposed EM algorithm can be summarized as follows:

Step 0:

Choose initial values for β(0) and λk(0) for k=1,,Kn, and set l=0.

Step 1:

At the (l+1)th iteration, calculate each E(Wik) based on the observed data and the parameter estimates at the lth iteration.

Step 2:

Update each λk with the closed-form expression (3).

Step 3:

Update β by solving the estimation equation Uβ(θ)=0 with the one-step Newton-Raphson method, and increase l by 1.

Step 4:

Repeat Steps 1 - 3 until the convergence is achieved.

The resulting estimators of β and Λ(t) are denoted as β^ and Λ^(t)=tktλ^k, respectively, where λ^k is the estimate of λ for k=1,,Kn. For the standard error estimation of β^ and Λ^(t), we propose to simply employ the nonparametric bootstrap approach ([29], for example), and the numerical results below suggest that it seems to work well in finite samples. The numerical results also indicate that the performance of the proposed algorithm is quite robust to the choices of the initial values of β and λk’s. In the practical implementation of the proposed algorithm, one can simply set the initial value of each regression parameter to 0 and the initial value of each λk to 1/Kn. The algorithm is declared to achieve convergence if the sum of the absolute differences between two successive estimates of all parameters is less than a small positive constant, say 0.001. We implement the proposed algorithm under the Rcpp environment, which guarantees that the computation is efficient and tractable.

Simulation studies

Simulation studies were conducted to assess the empirical performance of the proposed estimation procedure. In the study, the failure time T was generated from model (1) with Z=(Z1,Z2), Z1Bernoulli(0.5), Z2Uniform(-0.5,0.5), β=(β1,β2)=(1,1), and Λ(t)=t2, which corresponds to the Weibull distribution with the scale parameter 1 and the shape parameter 2. The truncation time A was generated from either Uniform(0, τ) or exponential distribution with rate θ, where τ or θ was chosen to yield about 50% average truncation rate. Note that when the truncation time follows the uniform distribution or satisfies the stationary assumption, we have the length-biased data, a special type of the left-truncated data as discussed above. Under the left truncation mechanism, the observed failure time T was equal to T if T>A. We firstly considered the situation with left-truncated and partly interval-censored data. To construct censoring, for each subject, we mimicked the periodical follow-up study and generated a sequence of examination times with the first observation time being A and the gap times of two successive observation times being 0.05+Uniform(0,0.5). Then we used the above simulated failure time T instead of the interval-censored observation if interval length is less than 0.2 to construct the uncensored or exactly observed T. The length of study was set to be 1.5, beyond which no further examinations were conducted.

For comparison, we considered the following three competing methods: the proposed pairwise pseudo-likelihood method (Proposed method), the NPMLE method without adjusting for the left truncation (Ignoring truncation) and the conditional likelihood method (CL method). Specifically, in the supplementary materials, we developed an EM algorithm with Poisson latent variables to implement the conditional likelihood method, and the “Ignoring truncation” method can be implemented with the EM algorithm by setting each Ai=0. We set n=100, 300 or 500, and used 1000 replicates. Under the above configurations, the proportions of exactly-observed failure times ranged from 4% to 26%; left censoring rates ranged from 16% to 37%; right censoring rates ranged from 7% to 33% and interval censoring rates ranged from 24% to 58%.

Table 1 presents the simulation results for the estimated regression parameters and the cumulative hazards function at t=0.4, 0.8 or 1.2 with partly interval-censored data. They include the estimated bias (Bias) given by the average of the 1000 estimates minus the true value, the sample standard error (SSE) of the 1000 estimates, the average of the 1000 standard error estimates (SEE), and the 95% empirical coverage probability (CP) yielded by the normal approximation. Specifically, the standard errors of the proposed pairwise pseudo-likelihood estimators were calculated via the nonparametric bootstrapping with 100 bootstrap samples. For CL and “Ignoring truncation” methods, we followed Zeng et al. [30] and proposed to adopt the profile likelihood approach to perform the variance estimation. This approach is simple and easy to implement, but can only provide the variance estimation for the estimated regression parameter, finite-dimensional parameter of interest. Thus, the SEEs of the cumulative hazards function estimates of the CL and “Ignoring truncation” methods were not available in Table 1. Given that Λ(t) is always positive, we used the log-transformation and constructed its confidence band with the delta method as Mao and Lin [31] among others. For any t, the confidence interval of Λ(t) is given by [Λ^(t)exp{-z0.975σ^(t)/Λ^(t)}, Λ^(t)exp{z0.975σ^(t)/Λ^(t)}], where σ^(t) is the standard error estimate of Λ^(t), and z0.975 is the upper 97.5th percentile of the standard normal distribution.

Table 1.

Simulation results with partly interval-censored data, including the estimated bias (Bias), the sample standard error (SSE) of the estimates, the average of the standard error estimates (SEE), and the 95% empirical coverage probability (CP)

Proposed method CL method Ignoring truncation
n Par True Bias SSE SEE CP Bias SSE SEE CP Bias SSE SEE CP
A follows the uniform distribution
100 β1 1 0.041 0.246 0.254 96.6 0.061 0.290 0.278 92.7 0.192 0.273 0.311 94.1
β2 1 0.045 0.403 0.408 95.6 0.055 0.492 0.467 93.7 0.180 0.474 0.445 90.8
Λ(0.4) 0.16 0.009 0.094 0.087 95.6 0.002 0.092 -0.084 0.043
Λ(0.8) 0.64 -0.037 0.165 0.161 93.9 -0.048 0.164 -0.245 0.103
Λ(1.2) 1.44 -0.038 0.240 0.248 93.5 -0.067 0.240 -0.445 0.209
300 β1 1 0.008 0.134 0.129 93.5 0.013 0.156 0.151 93.1 0.120 0.148 0.306 95.9
β2 1 0.012 0.212 0.212 94.6 0.025 0.248 0.253 95.2 0.146 0.244 0.305 92.4
Λ(0.4) 0.16 0.020 0.067 0.064 94.2 0.020 0.067 -0.081 0.026
Λ(0.8) 0.64 0.020 0.104 0.105 95.3 0.020 0.107 -0.240 0.061
Λ(1.2) 1.44 -0.028 0.179 0.189 96.5 -0.024 0.182 -0.436 0.116
500 β1 1 0.014 0.101 0.099 95.4 0.018 0.115 0.117 96.0 0.154 0.101 0.197 79.6
β2 1 0.014 0.161 0.163 94.6 0.020 0.191 0.193 95.0 0.146 0.193 0.219 85.7
Λ(0.4) 0.16 0.012 0.048 0.048 96.7 0.012 0.048 -0.081 0.018
Λ(0.8) 0.64 0.010 0.076 0.075 95.0 0.009 0.077 -0.244 0.044
Λ(1.2) 1.44 -0.012 0.133 0.131 94.6 -0.012 0.135 -0.443 0.094
A follows the exponential distribution
100 β1 1 0.045 0.242 0.251 94.9 0.062 0.272 0.266 93.1 0.146 0.266 0.294 95.3
β2 1 0.047 0.396 0.405 95.3 0.071 0.451 0.453 95.3 0.149 0.435 0.435 90.5
Λ(0.4) 0.16 0.009 0.083 0.080 95.5 0.009 0.085 -0.068 0.046
Λ(0.8) 0.64 -0.036 0.159 0.160 93.7 -0.038 0.159 -0.178 0.114
Λ(1.2) 1.44 -0.042 0.234 0.240 92.7 -0.042 0.235 -0.280 0.251
300 β1 1 0.011 0.131 0.133 95.9 0.016 0.147 0.148 94.9 0.084 0.137 0.250 97.4
β2 1 -0.001 0.210 0.217 95.9 0.007 0.228 0.246 96.7 0.082 0.229 0.294 95.6
Λ(0.4) 0.16 0.017 0.053 0.053 96.5 0.017 0.054 -0.065 0.025
Λ(0.8) 0.64 0.014 0.097 0.093 93.8 0.013 0.100 -0.173 0.066
Λ(1.2) 1.44 -0.016 0.184 0.182 94.5 -0.015 0.184 -0.300 0.140
500 β1 1 0.012 0.100 0.101 95.2 0.016 0.115 0.113 94.4 0.069 0.116 0.263 92.3
β2 1 0.010 0.165 0.167 94.7 0.008 0.187 0.188 94.7 0.097 0.173 0.247 92.3
Λ(0.4) 0.16 0.014 0.044 0.045 95.3 0.015 0.044 -0.067 0.020
Λ(0.8) 0.64 0.013 0.075 0.073 94.6 0.013 0.077 -0.174 0.061
Λ(1.2) 1.44 -0.003 0.136 0.138 97.5 -0.003 0.138 -0.292 0.118

Note: “Proposed method” denotes the proposed pairwise pseudo-likelihood method, “CL method” denotes the conditional likelihood method, and “Ignoring truncation” denotes the NPMLE approach that ignores the existence of left truncation

One can see from Table 1 that the estimators of the proposed pairwise pseudo-likelihood method are virtually unbiased, the corresponding sample standard error estimates are close to the average standard error estimates, and the empirical coverage probabilities are all around the nominal value 95%, implying that the normal approximation of the asymptotic distribution of the proposed estimator seems reasonable. In addition, one can clearly find that the proposed method is more efficient than the conditional likelihood method, and this efficiency gain can be anticipated since the proposed method utilizes the information of the marginal distribution of the truncation time. Since the generated data are subject to biased sampling, as seen from Table 1, the “Ignoring truncation” method is expected to yield much larger estimation biases than the proposed and the conditional likelihood methods.

In the second study, we considered the left-truncated and interval-censored data. For this, we generated the truncation time A in the same way as before, and set the first examination time being A. The gap time of two successive observation times was set to be 0.05+Uniform(0,0.5), and the other model specifications were kept the same as above. Then we had the left-truncated and interval-censored data by contrasting the generated T with the observation times. Under the aforementioned simulation setups, the left censoring rates were from 20% to 56%; the right censoring rates ranged from 7% to 32%; interval censoring rates ranged from 27% to 67%. The simulation results summarized in Table 2 again indicate that the proposed method performs reasonably well and has some advantages over the conditional likelihood and the “Ignoring truncation” methods.

Table 2.

Simulation results with interval-censored data, including the estimated bias (Bias), the sample standard error (SSE) of the estimates, the average of the standard error estimates (SEE), and the 95% empirical coverage probability (CP)

Proposed method CL method Ignoring truncation
n Par True Bias SSE SEE CP Bias SSE SEE CP Bias SSE SEE CP
A follows the uniform distribution
100 β1 1 0.057 0.253 0.258 96.1 0.100 0.302 0.262 89.5 0.212 0.290 0.323 93.3
β2 1 0.079 0.408 0.411 95.2 0.114 0.510 0.427 87.4 0.214 0.471 0.451 88.2
Λ(0.4) 0.16 0.013 0.103 0.104 94.0 0.009 0.103 -0.076 0.052
Λ(0.8) 0.64 -0.053 0.186 0.184 93.9 -0.039 0.185 -0.236 0.113
Λ(1.2) 1.44 -0.033 0.306 0.301 92.1 -0.065 0.308 -0.431 0.230
300 β1 1 0.025 0.134 0.132 93.9 0.042 0.155 0.150 93.1 0.154 0.149 0.209 90.3
β2 1 0.016 0.212 0.214 95.3 0.032 0.249 0.244 93.2 0.169 0.250 0.240 83.4
Λ(0.4) 0.16 0.023 0.075 0.073 96.5 0.022 0.076 -0.068 0.034
Λ(0.8) 0.64 -0.006 0.131 0.132 94.2 0.002 0.127 -0.224 0.069
Λ(1.2) 1.44 -0.021 0.229 0.222 95.2 -0.017 0.225 -0.427 0.139
500 β1 1 0.005 0.100 0.100 96.1 0.017 0.121 0.115 94.3 0.132 0.107 0.189 86.6
β2 1 0.022 0.165 0.163 93.6 0.036 0.202 0.189 91.8 0.148 0.188 0.179 75.4
Λ(0.4) 0.16 0.019 0.059 0.055 93.9 0.018 0.063 -0.066 0.028
Λ(0.8) 0.64 -0.014 0.098 0.101 95.1 -0.008 0.097 -0.222 0.057
Λ(1.2) 1.44 -0.022 0.186 0.182 94.6 -0.019 0.189 -0.429 0.107
A follows the exponential distribution
100 β1 1 0.084 0.250 0.266 96.8 0.115 0.292 0.259 88.9 0.164 0.273 0.309 94.2
β2 1 0.084 0.411 0.428 96.3 0.125 0.484 0.424 88.9 0.162 0.449 0.437 90.9
Λ(0.4) 0.16 0.008 0.096 0.101 96.2 0.007 0.097 -0.062 0.056
Λ(0.8) 0.64 -0.046 0.178 0.174 93.6 -0.05 0.179 -0.169 0.134
Λ(1.2) 1.44 -0.043 0.289 0.303 93.5 -0.037 0.290 -0.267 0.272
300 β1 1 0.030 0.136 0.135 94.3 0.046 0.152 0.145 92.5 0.126 0.145 0.179 91.4
β2 1 0.019 0.223 0.221 94.9 0.034 0.254 0.239 93.0 0.136 0.248 0.218 82.1
Λ(0.4) 0.16 0.021 0.066 0.068 95.9 0.023 0.067 -0.056 0.035
Λ(0.8) 0.64 -0.012 0.117 0.116 94.4 -0.005 0.119 -0.169 0.079
Λ(1.2) 1.44 -0.027 0.206 0.211 95.1 -0.235 0.205 -0.288 0.161
500 β1 1 0.014 0.105 0.102 94.5 0.024 0.116 0.111 93.5 0.100 0.110 0.187 90.4
β2 1 0.017 0.168 0.168 95.0 0.027 0.191 0.185 94.2 0.103 0.190 0.185 82.7
Λ(0.4) 0.16 0.019 0.054 0.052 97.7 0.017 0.054 -0.055 0.027
Λ(0.8) 0.64 -0.006 0.096 0.099 96.5 -0.003 0.094 -0.169 0.063
Λ(1.2) 1.44 -0.002 0.184 0.184 95.1 -0.200 0.186 -0.288 0.131

Note: “Proposed method” denotes the proposed pairwise pseudo-likelihood method, “CL method” denotes the conditional likelihood method, and “Ignoring truncation” denotes the NPMLE approach that ignores the existence of left truncation

Note that Wu et al. [26] considered the left-truncated and right-censored data and proposed an iterative estimation procedure to implement the pairwise pseudo-likelihood method. It is clear that the proposed method can deal with such data too. Therefore, one may be interested in comparing the performance of the proposed method with that of Wu et al. [26]. To investigate this, we generated the failure time T from model (1) with Z=(Z1,Z2), Z1Bernoulli(0.5), Z2Uniform(-1,1), β1=β2=1, and Λ(t)=t2. The truncation time A was generated in the same way as before. The right censoring time C was generated independently from Uniform(0,Cmax), where Cmax were chosen to yield about 30% right censoring rate. The results given in Table 3 imply that the two methods can both perform well and give similar performance.

Table 3.

Simulation results for the comparison of the proposed method with Wu et al. (2018)’s method under right censored data, including the estimated bias (Bias), the sample standard error (SSE) of the estimates, the average of the standard error estimates (SEE), and the 95% empirical coverage probability (CP)

Proposed method Wu et al. (2018)’s method
n Par True Bias SSE SEE CP Bias SSE SEE CP
A follows the uniform distribution
100 β1 1 0.025 0.244 0.247 95.0 0.026 0.244 0.227 92.8
β2 1 0.027 0.391 0.398 94.9 0.027 0.391 0.368 93.6
300 β1 1 0.011 0.129 0.133 96.0 0.012 0.129 0.130 95.2
β2 1 0.005 0.246 0.216 94.8 0.005 0.216 0.211 95.1
500 β1 1 0.005 0.100 0.102 95.2 0.005 0.100 0.100 95.1
β2 1 0.002 0.166 0.165 95.1 0.003 0.166 0.162 94.9
A follows the exponential distribution
100 β1 1 0.024 0.248 0.257 95.8 0.024 0.248 0.237 94.3
β2 1 0.015 0.398 0.416 95.7 0.015 0.398 0.383 93.6
300 β1 1 0.003 0.134 0.138 95.6 0.003 0.134 0.135 95.3
β2 1 0.008 0.218 0.223 95.4 0.008 0.218 0.219 95.2
500 β1 1 0.010 0.107 0.106 94.7 0.010 0.107 0.105 95.2
β2 1 0.011 0.172 0.171 94.8 0.011 0.172 0.169 94.6

An application

We apply the proposed method to a set of real data arising from the Massachusetts Health Care Panel Study (MHCPS) discussed in Pan and Chappell [17], Gao and Chan [24] and others. In 1975, the MHCPS enrolled elderly people who had not lost the active life in Massachusetts to evaluate the effect of gender (male or female) on the time to loss of active life. To determine when individuals in the study lost the active life, three subsequent follow-ups were taken at the 1.25, 6, and 10 years after the study enrolment. Therefore, age of the loss of active life, the defined failure time of interest T, cannot be recorded exactly and suffered from interval censoring. In the MHCPS, since subjects who had lost the active life before the study were not enrolled, the age of the loss of active life was subject to left truncation with the truncation time A being the age at enrolment [17]. Therefore, we had left-truncated and interval-censored data. After deleting a small amount of unrealistic records of the raw data, 1025 subjects with the age ranging from 65 to 97.3 were considered in the current analysis. In particular, the right censoring rate is 45.8%.

Define Z=1 if the individual is male and 0 otherwise. For the analysis of the MHCPS data, as in the simulation studies, we considered three competing methods: the proposed pairwise pseudo-likelihood method (Proposed method), the conditional likelihood approach (CL method), and the NPMLE method that ignores the existence of left truncation (Ignoring truncation). Table 4 presents the obtained results including the estimated covariate effect (Est), the standard error estimate (Std) and the associated p-value for testing the covariate effect being zero. In the proposed pairwise pseudo-likelihood method, as in the simulation study, we employed the nonparametric bootstrapping with 100 bootstrap samples to calculate the standard error of the estimated regression parameter.

Table 4.

Analysis results of the MHCPS data, including the estimated covariate effect (Est), the standard error estimate (Std) and the p-value

Method Est Std p-value
Proposed method 0.122 0.060 0.041
CL method 0.133 0.082 0.103
Ignoring truncation 0.156 0.095 0.100

Note: “Proposed method” denotes the proposed pairwise pseudo-likelihood method, “CL method” denotes the conditional likelihood method, and “Ignoring truncation” denotes the NPMLE approach that ignores the existence of left truncation

One can see from Table 4 that the estimated coefficient and the standard error estimate of the proposed method are given by 0.122 and 0.060, respectively, meaning that males have significantly higher risk of losing active life than females. This conclusion is in accordance with that given in Gao and Chan [24] where the length-biased assumption was made for the truncation time. One can also find from Table 4 that the CL method recognized the covariate effect as non-significant, which is different from the conclusion obtained by the proposed method. This phenomenon may arise partly due to the fact the CL method often loses some estimation efficiency compared with the proposed method. Moreover, the results given in Table 4 suggested that the NPMLE method that ignores the existence of left truncation tended to overestimate the covariate effect, and this effect was also recognized as non-significant.

Discussion and concluding remarks

In the preceding sections, we proposed a general or unified pairwise pseudo-likelihood approach for the analysis of left-truncated failure time data under the PH model. The proposed method is quite general and flexible since it applies to various types of censored data, including the partly interval-censored, interval-censored, and right-censored data. We devised an EM algorithm to calculate the nonparametric maximum likelihood estimators, which was shown to be computationally stable and reliable in finite samples. Numerical results indicated that, by utilizing the pairwise order information of the truncation times, the proposed method can indeed yield more efficient estimators compared with the conventional conditional likelihood estimation approach. An application to the MHCPS data demonstrated the practical utility of the proposed method.

Notably, in the proposed algorithm, the derivation of the self-consistent solution (3) for λk is the desirable feature, which avoids the use of high-dimensional optimization procedure. In addition, the estimation equation Uβ(θ)=0 for β has tractable form and can be readily solved with some routine optimization procedure, such as the Newton-Raphson method. The two desirable features both make the proposed algorithm computationally stable and reliable. There may also exist some shortcomings of the proposed method. One is that the self-consistent solution (3) may not ensure that the estimate of λk is always non-negative. However, it has been our experience that, given a reasonable initial value, the negative estimate of λk is unlikely to occur in the simulations. As an alternative, by following Zhou et al. [32] and others, one can attempt to reparameterize each λk as exp(λk), where λk is the unconstrained parameter to be estimated. Another is that we adopted the nonparametric bootstrap method to calculate the variance of parameter estimate, which involves repeated data sampling. This procedure will become computationally intensive if the sample size is extremely large. Future efforts will be devoted to develop a simple variance estimation procedure.

There may also exist several potential research directions for future research. One is that in the proposed method, we made a non-informative or independent censoring assumption [33, 34]. In other words, the failure times of interest were assumed to be conditionally independent of the observation times given the covariates. However, it is apparent that this assumption may not hold in some applications, and thus the generalizing of the proposed method to the situation of informative censoring deserves further investigation. In some applications, one may also encounter bivariate or multivariate failure time data [35], and it would be helpful to generalize the proposed method to deal with such data. Also the extensions of the proposed method to other regression models such as the transformation or additive hazards models can be useful.

Supplementary Information

Additional file 1. (148.9KB, pdf)

Acknowledgements

We would like to thank the editor office for the efforts on handing this submission. We also wish to thank the editor, the associate editor, and reviewers for the helpful comments and suggestions that greatly improved this article.

Abbreviations

PH

Proportional hazards

NPMLE

Nonparametric maximum likelihood estimation

EM

Expectation Maximization Algorithm

CL

Conditional likelihood

SSE

Sample standard error

SEE

Standard error estimate

CP

Coverage probability

MHCPS

Massachusetts Health Care Panel Study

Authors’ contributions

SL proposed the idea. LS wrote the R code and created the R package. HL conducted the simulation and real data analysis. SL, LS and HL wrote the original version of the manuscript together and JS polished the manuscript. All authors read and approved the final manuscript.

Funding

Shuwei Li’s research was partially supported by Science and Technology Program of Guangzhou of China (Grant No. 202102010512), the National Nature Science Foundation of China (Grant No. 11901128), and Nature Science Foundation of Guangdong Province of China (Grant Nos. 2021A1515010044 and 2022A1515011901). Li Shao’s work was supported by Guangdong Basic and Applied Basic Research Foundation (Grant No. 2021A1515110926).

Availability of data and materials

The MHCPS data set used in this study can be downloaded at https://onlinelibrary.wiley.com/action/downloadSupplement?doi=10.1111%2Fj.0006-341X.2002.00064.x &file=BIOM_64_sm_010423.txt. The proposed algorithm can be implemented in the R package LTsurv, which is publicly available at https://github.com/lishuwstat/Left-truncation-Cox-Pairwise-likelihood.

Declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Li Shao, Email: sesshaoli@gzhu.edu.cn.

Hongxi Li, Email: lihongxi@e.gzhu.edu.cn.

Shuwei Li, Email: seslishuw@gzhu.edu.cn.

Jianguo Sun, Email: sunj@missouri.edu.

References

  • 1.Kalbfleisch JD, Prentice RL. The statistical analysis of failure time data. New York: Wiley; 2002. [Google Scholar]
  • 2.Sun J. The statistical analysis of interval-censored failure time data. New York: Springer; 2006. [Google Scholar]
  • 3.Cox DR. Regression models and life-tables (with Discussion) J R Stat Soc Ser B. 1972;34(2):187–220. [Google Scholar]
  • 4.Lin DY, Ying Z. Semiparametric analysis of the additive risk model. Biometrika. 1994;81(1):61–71. doi: 10.1093/biomet/81.1.61. [DOI] [Google Scholar]
  • 5.Zeng D, Lin DY. Efficient estimation of semiparametric transformation models for counting processes. Biometrika. 2006;93(3):627–640. doi: 10.1093/biomet/93.3.627. [DOI] [Google Scholar]
  • 6.Chiou SH, Kang S, Yan J. Rank-based estimating equations with general weight for accelerated failure time models: an induced smoothing approach. Stat Med. 2015;34:1495–1510. doi: 10.1002/sim.6415. [DOI] [PubMed] [Google Scholar]
  • 7.Huang J. Efficient Estimation for the Cox Model with Interval Censoring. Ann Stat. 1996;24(2):540–568. doi: 10.1214/aos/1032894452. [DOI] [Google Scholar]
  • 8.Huang J. Asymptotic properties of nonparametric estimation based on partly interval-censored data. Stat Sin. 1999;9:501–519. [Google Scholar]
  • 9.Kim JS. Maximum likelihood estimation for the proportional hazards model with partly interval-censored data. J R Stat Soc Ser B. 2003;65(2):489–502. doi: 10.1111/1467-9868.00398. [DOI] [Google Scholar]
  • 10.Gao F, Zeng D, Lin DY. Semiparametric estimation of the accelerated failure time model with partly interval-censored data. Biometrics. 2017;73(4):1161–1168. doi: 10.1111/biom.12700. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Li J, Ma J. Maximum penalized likelihood estimation of additive hazards models with partly interval censoring. Comput Stat Data Anal. 2019;137:170–180. doi: 10.1016/j.csda.2019.02.010. [DOI] [Google Scholar]
  • 12.Pan C, Cai B, Wang L. A Bayesian approach for analyzing partly interval-censored data under the proportional hazards model. Stat Methods Med Res. 2020;29(11):3192–3204. doi: 10.1177/0962280220921552. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Zhou Q, Sun Y, Gilbert PB. Semiparametric regression analysis of partly interval-censored failure time data with application to an AIDS clinical trial. Stat Med. 2021;40(20):4376–4394. doi: 10.1002/sim.9035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Huang CY, Qin J. Semiparametric estimation for the additive hazards model with left-truncated and right-censored data. Biometrika. 2013;100(4):877–888. doi: 10.1093/biomet/ast039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Wang MC. Nonparametric estimation from cross-sectional survival data. J Am Stat Assoc. 1991;86(413):130–143. doi: 10.1080/01621459.1991.10475011. [DOI] [Google Scholar]
  • 16.Wang MC, Brookmeyer R, Jewell NP. Statistical models for prevalent cohort data. Biometrics. 1993;49:1–11. doi: 10.2307/2532597. [DOI] [PubMed] [Google Scholar]
  • 17.Pan W, Chappell R. Estimation in the Cox proportional hazards model with left-truncated and interval-censored data. Biometrics. 2002;58(1):64–70. doi: 10.1111/j.0006-341X.2002.00064.x. [DOI] [PubMed] [Google Scholar]
  • 18.Shen Y, Ning J, Qin J. Analyzing length-biased data with semiparametric transformation and accelerated failure time models. J Am Stat Assoc. 2009;104(487):1192–1202. doi: 10.1198/jasa.2009.tm08614. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Qin J, Ning J, Liu H, Shen Y. Maximum likelihood estimations and EM algorithms with length-biased data. J Am Stat Assoc. 2011;106(496):1434–1449. doi: 10.1198/jasa.2011.tm10156. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Shen PS. Proportional hazards regression with interval-censored and left-truncated data. J Stat Comput Simul. 2014;84(2):264–272. doi: 10.1080/00949655.2012.705844. [DOI] [Google Scholar]
  • 21.Shen PS. Conditional MLE for the proportional hazards model with left-truncated and interval-censored data. Stat Probab Lett. 2015;100:164–171. doi: 10.1016/j.spl.2015.02.015. [DOI] [Google Scholar]
  • 22.Wang P, Tong X, Zhao S, Sun J. Efficient estimation for the additive hazards model in the presence of left-truncation and interval censoring. Stat Interface. 2015;8(3):391–402. doi: 10.4310/SII.2015.v8.n3.a12. [DOI] [Google Scholar]
  • 23.Shen Y, Ning J, Qin J. Nonparametric and semiparametric regression estimation for length-biased survival data. Lifetime Data Anal. 2017;23(1):3–24. doi: 10.1007/s10985-016-9367-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Gao F, Chan KCG. Semiparametric regression analysis of length-biased interval-censored data. Biometrics. 2019;75(1):121–132. doi: 10.1111/biom.12970. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Wu Y, Chambers CD, Xu R. Semiparametric sieve maximum likelihood estimation under cure model with partly interval censored and left truncated data for application to spontaneous abortion. Lifetime Data Anal. 2019;25(3):507–528. doi: 10.1007/s10985-018-9445-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Wu F, Kim S, Qin J, Saran R, Li Y. A pairwise likelihood augmented Cox estimator for left-truncated data. Biometrics. 2018;74(1):100–108. doi: 10.1111/biom.12746. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Wang P, Li D, Sun J. A pairwise pseudo-likelihood approach for left-truncated and interval-censored data under the Cox model. Biometrics. 2021;77(4):1303–1314. doi: 10.1111/biom.13394. [DOI] [PubMed] [Google Scholar]
  • 28.Liang KY, Qin J. Regression analysis under non-standard situations: a pairwise pseudolikelihood approach. J R Stat Soc Ser B. 2000;62(4):773–786. doi: 10.1111/1467-9868.00263. [DOI] [Google Scholar]
  • 29.Efron B. Censored data and the bootstrap. J Am Stat Assoc. 1981;76:316–319. doi: 10.1080/01621459.1981.10477650. [DOI] [Google Scholar]
  • 30.Zeng D, Mao L, Lin D. Maximum likelihood estimation for semiparametric transformation models with interval-censored data. Biometrika. 2016;103(2):253–271. doi: 10.1093/biomet/asw013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Mao L, Lin DY. Efficient estimation of semiparametric transformation models for the cumulative incidence of competing risks. J R Stat Soc Ser B. 2017;79:573–587. doi: 10.1111/rssb.12177. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Zhou Q, Hu T, Sun J. A Sieve Semiparametric Maximum Likelihood Approach for Regression Analysis of Bivariate Interval-Censored Failure Time Data. J Am Stat Assoc. 2017;112:664–672. doi: 10.1080/01621459.2016.1158113. [DOI] [Google Scholar]
  • 33.Ma L, Hu T, Sun J. Sieve maximum likelihood regression analysis of dependent current status data. Biometrika. 2015;102:731–738. doi: 10.1093/biomet/asv020. [DOI] [Google Scholar]
  • 34.Li S, Hu T, Wang P, Sun J. Regression analysis of current status data in the presence of dependent censoring with applications to tumorigenicity experiments. Comput Stat Data Anal. 2017;110:75–86. doi: 10.1016/j.csda.2016.12.011. [DOI] [Google Scholar]
  • 35.Piao J, Ning J, Shen Y. Semiparametric model for bivariate survival data subject to biased sampling. J R Stat Soc Ser B. 2019;81:409–429. doi: 10.1111/rssb.12308. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Additional file 1. (148.9KB, pdf)

Data Availability Statement

The MHCPS data set used in this study can be downloaded at https://onlinelibrary.wiley.com/action/downloadSupplement?doi=10.1111%2Fj.0006-341X.2002.00064.x &file=BIOM_64_sm_010423.txt. The proposed algorithm can be implemented in the R package LTsurv, which is publicly available at https://github.com/lishuwstat/Left-truncation-Cox-Pairwise-likelihood.


Articles from BMC Medical Research Methodology are provided here courtesy of BMC

RESOURCES