Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2017 Jun 12;36(21):3422–3436. doi: 10.1002/sim.7368

Information criteria for Firth's penalized partial likelihood approach in Cox regression models

Kengo Nagashima 1,, Yasunori Sato 1
PMCID: PMC6084330  PMID: 28608396

Abstract

In the estimation of Cox regression models, maximum partial likelihood estimates might be infinite in a monotone likelihood setting, where partial likelihood converges to a finite value and parameter estimates converge to infinite values. To address monotone likelihood, previous studies have applied Firth's bias correction method to Cox regression models. However, while the model selection criteria for Firth's penalized partial likelihood approach have not yet been studied, a heuristic AIC‐type information criterion can be used in a statistical package. Application of the heuristic information criterion to data obtained from a prospective observational study of patients with multiple brain metastases indicated that the heuristic information criterion selects models with many parameters and ignores the adequacy of the model. Moreover, we showed that the heuristic information criterion tends to select models with many regression parameters as the sample size increases. Thereby, in the present study, we propose an alternative AIC‐type information criterion based on the risk function. A Bayesian information criterion type was also evaluated. Further, the presented simulation results confirm that the proposed criteria performed well in a monotone likelihood setting. The proposed AIC‐type criterion was applied to prospective observational study data. © 2017 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd

Keywords: Akaike's information criterion, model selection, monotone likelihood, penalized partial likelihood, survival analysis

1. Introduction

The Cox regression model 1 is one of the most useful and widely used tools in survival analysis. In Cox regression model estimations, maximum partial likelihood estimates may be infinite in monotone likelihood situations 2. In such cases, partial likelihood converges to a finite value, and the parameter estimates and standard errors converge to infinite values; hence, these results are not interpretable. Such problems arise, for example, in the presence of unbalanced covariates, large parameter effects, and/or heavy censoring.

To address this monotone likelihood problem, Heinze and Schemper proposed Firth's penalized partial likelihood approach 3. They directly applied to the Cox regression model Firth's bias correction method 4, which aims to remove asymptotic bias from maximum likelihood estimates in exponential families with canonical link functions. Firth's penalized partial likelihood approach reduces asymptotic bias and addresses the monotone likelihood problem 3, 5. Firth's bias correction method was also applied to logistic regression models to address the separation problem 5, 6, 7, which is similar to the monotone likelihood problem. This approach reduces asymptotic bias and also overcomes the separation problem.

In this study, we discuss the model selection criteria for Firth's penalized partial likelihood approach based on Akaike's information criterion (AIC) 8 and Bayesian information criterion (BIC) 9. Although model selection is an important issue in data analysis, the model selection criteria for Firth's penalized partial likelihood approach have never been studied. To our best knowledge, only the SAS PHREG procedure can be used to obtain an AIC‐type heuristic information criterion, AIC=−2log(maximum penalized partial likelihood)+2p, where p is the number of regression parameters, and other major statistical software (e.g., Stata and R) can only output the log penalized partial likelihood. However, AIC is not theoretically justified, and especially, we find that AIC tends to select a model that has a large number of regression parameters as n, where n is the sample size; that is, AIC does not have the important property of avoiding over‐fitting. This result indicates that AIC is not a suitable model selection criterion. Similarily, the SAS PHREG procedure implements a BIC‐type heuristic information criterion, BIC=−2log(maximum penalized partial likelihood)+logd, where d is the number of events. BIC is not also theoretically justified. Therefore, we consider alternative model selection criteria in this setting.

The remainder of this paper is organized as follows. In Section 2, we introduce motivating data and issues of AIC. Section 3 briefly reviews Firth's bias correction method and penalized partial likelihood approach, discusses the fundamental problems of AIC and BIC, and proposes appropriate information criteria. Section 4 presents the simulation results to demonstrate the performance of the criteria and to check the property of AIC holds. Section 5 applies the proposed method to real data, and Section 6 concludes the paper with a brief discussion.

2. Motivating example

Yamamoto et al. 10 collected time‐to‐event (e.g., death, local recurrence, and leptomeningeal dissemination) data for 1194 cancer patients with multiple brain metastases. Secondary end points of this study include time to leptomeningeal dissemination (the data on 928 patients were censored, while the MRI results of 121 patients (10%), that is, those who suffered an early death or who deteriorated markedly soon after stereotactic radiosurgery, were not available; 145 patients had an event). We analyzed the following covariates: age (<65, ⩾65), sex (female, male), kps(Karnofsky performance status; ⩾80, ⩽70), ntumor (number of tumors; 1, 2–4, 5–10), diameter (maximum diameter of largest tumor; <1.6cm, ⩾1.6cm), volume (cumulative tumor volume; <1.9 mL, ⩾1.9 mL), ptumor (primary tumor category; lung, breast, gastrointestinal, kidney, other), status(extracerebral disease status; not controlled, controlled), and neuro (neurological symptoms; no, yes). To analyze the competing risk end point, leptomeningeal dissemination, we used cause‐specific proportional hazard models, which are identical to usual Cox regression models 11.

The descriptive statistics for the study data are shown in Table 1. The data have heavy censoring, and the number of events differs considerably for the primary tumor categories. In particular, the kidney cancer group has no events. Further, as we illustrate below, monotone likelihood was observed in these data because of the primary tumor categories, while the parameter estimate of the kidney cancer group converges to −.

Table 1.

Number of events (leptomeningeal dissemination) and censored values for the study data (n=1073).

Covariate Group Event Censored % Censored
age <65 69 393 85.1
⩾65 76 535 87.6
sex Female 62 370 85.6
Male 83 558 87.1
kps ⩾80 132 810 86.0
⩽70 13 118 90.1
ntumor 1 49 365 88.2
2–4 61 412 87.1
5–10 35 151 81.2
diameter <1.6cm 76 455 85.7
⩾1.6 cm 69 473 87.3
volume <1.9 mL 75 459 86.0
⩾1.9 mL 70 469 87.0
ptumor Lung 116 705 85.9
Breast 17 95 84.8
Gastrointestinal 9 66 88.0
Kidney 0 32 100.0
Other 3 30 90.9
status Not controlled 103 634 86.0
Controlled 42 294 87.5
neuro No 105 656 86.2
Yes 40 272 87.2
Total 145 928 86.5

Note: ntumor, number of tumors; kps, Karnofsky performance status; diameter, maximum diameter of largest tumor; volume, cumulative tumor volume; ptumor, primary tumor category; status, extracerebral disease status; neuro, neurological symptoms.

Next, we consider the model selection based on AIC, the results of which are shown in Table 2. The full model, which includes all the covariates, was selected by AIC as the best. In the best model, the hazard ratios were estimated by using Firth's penalized partial likelihood approach (Table 3). To illustrate the problem of monotone likelihood, the hazard ratios estimated by using a usual Cox regression model are also shown. For the usual Cox regression model, when monotone likelihood occurred, the hazard ratio of the kidney cancer group was 0.00 (= exp(−)), standard error was 543.30, p‐value of the Wald test was 0.98, and p‐value of the likelihood ratio test was <0.01. Although the number of events for lung cancer was considerably larger than that for kidney cancer (Table 1), a large p‐value was observed in the Wald test. On the contrary, the results derived by using Firth's penalized partial likelihood approach were plausible. The hazard ratio was 0.12, and standard error was 1.43 (Table 3); therefore, usual Cox regression models were unsatisfactory in the presence of monotone likelihood.

Table 2.

The top five models based on AIC for the study data.

Model AIC
age sex kps ntumor diameter volume ptumor status neuro 1731.50
age sex ntumor diameter volume ptumor status neuro 1731.80
age sex kps ntumor diameter ptumor status neuro 1731.85
age sex kps ntumor volume ptumor status neuro 1731.85
age sex ntumor volume ptumor status neuro 1732.16

Note: ntumor, number of tumors; kps, Karnofsky performance status; diameter, maximum diameter of largest tumor; volume, cumulative tumor volume; ptumor, primary tumor category; status, extracerebral disease status; neuro, neurological symptoms; AIC, Akaike's information criterion.

Table 3.

Parameter estimates of the best model based on AIC for the study data.

Usual Cox Firth's penalized partial
Covariate regression model likelihood approach
HR SE 95% CI p (Wald) p (LR) HR SE 95% CI p (LR)
age (⩾65 vs. <65) 1.01 0.17 0.72 1.40 0.98 0.98 1.00 0.17 0.72 1.40 0.98
sex (male vs. female) 1.20 0.19 0.83 1.73 0.34 0.33 1.19 0.19 0.83 1.72 0.34
ntumor
   1 vs. 2–4 0.72 0.20 0.49 1.06 0.09 0.09 0.72 0.20 0.49 1.06 0.09
   5–10 vs. 2–4 1.58 0.22 1.03 2.41 0.04 0.04 1.59 0.22 1.04 2.43 0.04
kps (⩾80 vs. <70) 1.04 0.32 0.55 1.96 0.90 0.90 1.07 0.32 0.57 2.00 0.83
diameter (⩾1.6 vs. <1.6) 0.90 0.33 0.47 1.69 0.73 0.73 0.90 0.33 0.47 1.70 0.74
volume (⩾1.9 vs. <1.9) 1.10 0.32 0.59 2.08 0.76 0.76 1.11 0.32 0.59 2.08 0.76
ptumor
   Breast vs. lung 1.03 0.29 0.58 1.83 0.91 0.91 1.05 0.29 0.60 1.87 0.85
   GI vs. lung 1.55 0.37 0.75 3.21 0.24 0.26 1.61 0.37 0.79 3.31 0.21
   Kidney vs. lung 0.00 543.30 0.00 0.98 <0.01 0.12 1.43 0.01 1.91 0.02
   Others vs. lung 0.77 0.59 0.24 2.45 0.66 0.64 0.89 0.55 0.30 2.63 0.83
status (nc. vs. controlled) 1.02 0.19 0.71 1.47 0.92 0.92 1.03 0.19 0.71 1.48 0.89
neuro (yes vs. no) 1.50 0.23 0.96 2.36 0.07 0.08 1.51 0.23 0.97 2.37 0.07

Note: HR, hazard ratio; kps, Karnofsky performance status; ntumor, number of tumors; diameter, maximum diameter of largest tumor; volume, cumulative tumor volume; ptumor, primary tumor category; status, extracerebral disease status; neuro, neurological symptoms; GI, gastrointestinal; nc., not controlled; LR, likelihood ratio; SE, standard error.

Now, we return to the model selection result based on AIC when using Firth's penalized partial likelihood approach. As shown in Table 2, model selection based on AIC tends to select models that have many parameters, and to support this statement, we discuss the theoretical property of AIC in Section 3.3. In Table 3, the best model under AIC includes variables that have considerably small effects such as age(H R=1.00, p‐value=0.98) and status (H R=1.03, p‐value=0.89), whose p‐values were very large. Although these variables have little association with the time‐to‐event, such variables were included in the best model and subsequent models ranked in the top 5. Indeed, because model selection based on AIC performs badly, we propose an alternative approach herein to address this problem.

3. Infomation criteria for Firth's penalized partial likelihood approach

3.1. Cox regression model

We consider Cox regression models 1 with Andersen and Gill's 12 counting process formulation. A triplet (Ω,F,P) is a probability space, and {Ft,t[0,1]} is an increasing right continuous family of sub σ‐algebras of F that includes failure time and covariate histories to scaled time t and censoring histories to t +. Let N=(N 1,…,N i,…,N n)T for i=1,…,n be an n‐component multivariate counting process, where N i(t) counts the number of failures (0 or 1) for the ith individual in scaled time t∈[0,1]. The sample paths N 1,…,N i,…,N n are step functions, with 0 at time 0 and no two components having simultaneous jumps. Now, suppose that N i(t) has a random intensity process hi(t)=Yi(t)h0(t)exp{β0TZi(t)}, where h 0(t) is a baseline hazard function, Y i(t) is a predictable process taking the value of 1 if the ith individual is at risk at time t and 0 otherwise, β 0=(β 01,…,β 0p)T is a p‐dimensional vector of the true regression parameters, and the p‐dimensional vector Z i=(Z i1,…,Z ip)T is the predictable covariate process for the ith individual. Note that the superscript ‘T’ indicates the transpose of a matrix or a vector. We assume that (N i,Y i,Z i) are independent and identically distributed. In this case, the processes Mi(t)=Ni(t)0thi(x)dx are independent local square integrable martingales on the scaled time interval [0,1]. N i, Y i, and Z i are assumed to be adapted to {Ft,t[0,1]}.

Under these settings, the log‐partial likelihood function is defined as

l(N,Y,Z;β)=i=1n01βTZi(x)lognS(0)(β,x)dNi(x),

the score function is defined as

U(β)=l(N,Y,Z;β)β=i=1n01Zi(x)S(1)(β,x)S(0)(β,x)dNi(x),

and the observed information matrix is defined as

I(β)=2l(N,Y,Z;β)ββT=i=1n01S(2)(β,x)S(0)(β,x)S(1)(β,x)S(0)(β,x)2dNi(x),

where Y=(Y 1,…,Y n)T, Z=(Z 1,…,Z n)T, the scalar function S (0), the vector function S (1), and the matrix function S (2) are defined as S(k)(β,t)=n1i=1nYi(t)Zi(t)kexp{βTZi(t)} for k=0,1,2. Here, for a vector b, b ⊗0=1, b ⊗1=b, and b ⊗2=b b T. By using this notation, the usual Cox regression estimator β^Cox is obtained by solving U(β)=0.

3.2. Firth's bias correction method for Cox regression models

Heinze and Schemper 3 directly applied Firth's bias correction method 4 to Cox regression models to overcome monotone likelihood. They proposed an estimation method based on the penalized log‐partial likelihood, l (N,Y,Z;β)=l(N,Y,Z;β)+0.5log|I(β)|, and the modified score function U (β)=U(β)+a(β), where |I(β)| is the determinant of the observed information matrix, a(β)={a 1(β),…,a p(β)}xx‐xxT are modification terms, and a j(β)=tr[{I(β)}xx‐xx−1{ I(β)/ β j}]/2. The penalized partial likelihood estimator β^ is obtained by solving U (β)=0, which is different from the usual Cox regression estimator, β^Cox. They only assessed the empirical performance of these methods by using simulation studies. These simulation results confirm the satisfactory performance of the penalized likelihood ratio test and profile penalized likelihood confidence interval under monotone likelihood.

The modification term a(β) can be derived by using an asymptotic expansion of E[U (β)]. It will be convenient to employ the notation of Cox and Snell 13 and Firth 4. Let U j(β)= l(N,Y,Z;β)/ β j, U jk(β)= 2 l(N,Y,Z;β)/ β j β k, U jkl(β)= 3 l(N,Y,Z;β)/ β j β k β l, the null cumulants are κ j,k=n −1E[U j(β 0)U k(β 0)], κ jk=n −1E[U jk(β 0)], κ j,kl=n −1E[U j(β 0)U kl(β 0)], κ j,k,l=n −1E[U j(β 0)U k(β 0)U l(β 0)], and κ jkl=n −1E[U jkl(β 0)]. Based on the asymptotic expansion, the bias of the estimator of the mth regression parameter is given by

E[n1/2(β^mβ0m)]=n1κl,j12κk,l(κj,k,l+κj,kl)+αj(β0)+O(n3/2), (1)

where κ k,l denotes the inverse of the Fisher information matrix, Einstein summation convention is applied, and α j(β 0)=E[a j(β 0)]. From Eq. 1, if α j(β 0)=κ k,l(κ j,k,l+κ j,kl)/2, then the first‐order bias term disappears. Moreover, if κ j,k+κ jk=0, κ j,k,l+κ j,kl+κ k,jl+κ l,jk+κ jkl=0, and κ j,kl=0, then

αj(β)=12κk,l(κj,k,l+κj,kl)=12κk,lκj,k,l=12κk,lκjkl.

From the aforementioned result and paying attention to the summation convention, the modification term can be written as

aj(β)=12tr[{I(β)}1{I(β)/βj}].

However, the relationships κ j,k,l+κ j,kl+κ k,jl+κ l,jk+κ jkl=0 and κ j,kl=0 are a nontrivial result in Cox regression models, and thus, they have never been evaluated. Nevertheless, we proved that these relationships are true in Cox regression models under independent and identically distributed (see Appendix A for more details).

3.3. Problem in heuristic information criteria

As noted earlier, although model selection is an important issue in data analysis, the model selection criteria for the penalized partial likelihood approach have never been studied. To our best knowledge, only the SAS PHREG procedure can be used to obtain an AIC‐type heuristic information criterion, AIC=2l(N,Y,Z;β^)+2p=2l(N,Y,Z;β^)+2plog|I(β^)|. Moreover, other major statistical software (e.g., Stata and R) can only output the penalized log‐partial likelihood. However, AIC is not theoretically justified.

Now, we discuss a property of AIC. After some algebra,

AIC=2l(N,Y,Z;β^)+(2logn)plog{np|I(β^)|}.

The last term on the right‐hand side log{np|I(β^)|} converges to a constant because n1I(β^)PΣ(β0) (see Appendix B for more details) and |n1I(β^)|=np|I(β^)|P|Σ(β0)| as n, where Σ(β0)=01v(β0,x)s(0)(β0,x)h0(x)dx, v=(s (2)/s (0))−{s (1)/s (0)}xx‐xx⊗2, s (0)(β,t)=E[S (0)(β,t)], s (1)(β,t)=E[S (1)(β,t)], and s (2)(β,t)=E[S (2)(β,t)]. If n⩾8, then 2−logn is negative. Because AIC includes the term (2−logn)p, this criterion tends to select models with large p as n. Importantly, this result indicates that AIC does not avoid over‐fitting.

Similarly, the SAS PHREG procedure implements a BIC‐type heuristic information criterion, BIC=2l(N,Y,Z;β^)+plogd, where d is the number of events 14. Let c=1−d/n∈(0,1] be the proportion of censoring, BIC=2l(N,Y,Z;β^)+plog(1c)log{np|I(β^)|}. Because log(1−c)<0, BIC has a negative penalty term in proportion to the number of regression parameters.

3.4. Proposed criteria

As an alternative approach to address the issue discussed in Section 3.3, we propose a criterion termed herein AIC for Firth's penalized partial likelihood approach (AICF). AIC is a model selection criterion used to measure the goodness of fit of a model by using the risk function based on Kullback–Leibler (KL) information between the true model and the candidate model, which is a measure of discrepancy from the true model.

Xu et al. 15 provided a theoretical justification for the use of partial likelihood in AIC under usual Cox regression models, which was also extended to proportional hazards mixed models. These authors developed a profile AIC 16 for selecting a model with minimum KL information based on the profile likelihood under Cox regression models. It is well known that partial likelihood can be considered as profile likelihood 17, 18, 19. Suppose that f denotes the true distribution and g β,λ=g(·;β,λ) denotes candidate models, where λΛ is the nuisance parameter and Λ is the parameter space of λ. The KL information can be written as K L( f,g β,λ)=E(N,Y,Z)∼f[logf(N,Y,Z)−logg β,λ(N,Y,Z)]. Here and subsequently, we write EN=E(N,Y,Z)∼f for convenience. Focusing on the regression parameters β alone and ignoring the constant term EN[logf(N,Y,Z)] in KL, the minimum KL information is given at β 0 such that EN[loggβ0(N,Y,Z)]=maxβEN[loggβ(N,Y,Z)], where EN[logg β(N,Y,Z)]= maxλEN[logg β,λ(N,Y,Z)]. If the model is correctly specified (i.e., f=g β0), EN[loggβ0(N,Y,Z)]=Ωloggβ0(N,Y,Z)dP. Under Cox regression models, the log profile likelihood can be written as maxλi=1ngβ(Ni,Yi,Zi;β,λ)=l(N,Y,Z;β). Xu et al. 15 showed that the risk function, ENEÑ[2l(N˜,Y˜,Z˜;β^Cox)] based on the log profile likelihood, and profile AIC, 2l(N˜,Y˜,Z˜;β^Cox)+2p, as an approximately unbiased estimator of the risk function, where (N˜,Y˜,Z˜) is a future observation. Based on Akaike 8, the minimum risk function corresponds to the minimum KL information using a future observation.

Therefore, we consider a partial likelihood‐based risk function, RISK=ENEÑ[2l(N˜,Y˜,Z˜;β^)], and derive AICF as an approximately unbiased estimator of RISK. In the definition of RISK, the estimator β^ was not the usual Cox regression estimator β^Cox, but rather the Firth's penalized partial likelihood estimator. When we simply estimate RISK by 2l(N,Y,Z;β^), we need to correct bias B. Here, B is defined as

B=ENEÑ[2l(N,Y,Z;β^)2l(N˜,Y˜,Z˜;β^)]=b1+b2+b3,

where

b1=EN[EÑ[2l(N˜,Y˜,Z˜;β0)]EÑ[2l(N˜,Y˜,Z˜;β^)]],b2=EN[2l(N,Y,Z;β0)]EÑ[2l(N˜,Y˜,Z˜;β0)],

and

b3=EN[2l(N,Y,Z;β^)2l(N,Y,Z;β0)].

According to this definition, B includes the true parameter vector β 0. Therefore, we need approximate B. A second‐order Taylor expansion of EÑ[l(N˜,Y˜,Z˜;β^)] around β^=β0 gives

EÑ[l(N˜,Y˜,Z˜;β^)]EÑ[l(N˜,Y˜,Z˜;β0)]12{n(β^β0)T}Σ(β0){n(β^β0)}, (2)

a first‐order Taylor expansion of U(β^)=0 around β^=β0 gives

n(β^β0){n1I(β0)}1{n1/2U(β0)}, (3)

and a second‐order Taylor expansion of l(N,Y,Z;β 0) around β0=β^ gives

l(N,Y,Z;β0)l(N,Y,Z;β^)12{n(β^β0)T}{n1I(β^)}{n(β^β0)}+{a(β^)}T(β^β0), (4)

where I (β)=− U (β)/ β T=I(β)− a(β)/ β T. From the fact E[ f(D)]=E[ tr {f(D)}] for a scalar function f and a random vector D, and by substituting Eqs 24 into b 1, we can show that

b1EN[tr{Σ(β0){n1I(β0)}1{n1J(β0)}{n1I(β0)}1}],

where J(β)=i=1n{Li(β)}2, Li(β)=Li(β)a(β)/n, and Li(β0)=01{Zi(x)S(1)(β0,x)/S(0)(β0,x)}dMi(x). Similarly,

b3EN[tr{{n1I(β^)}{n1I(β0)}1{n1J(β0)}{n1I(β0)}12{a(β^)}T(β^β0)}].

Under the true model n1J(β0)PΣ(β0), n1I(β0)PΣ(β0), n1I(β^)PΣ(β0), and a(β^)=Op(1)(see Data S1 and Appendix B). Therefore, by applying the continuous mapping theorem, we obtain

tr{Σ(β0){n1I(β0)}1{n1J(β0)}{n1I(β0)}1}Ptr{Σ(β0){Σ(β0)}1Σ(β0){Σ(β0)}1}=p,

and

tr{{n1I(β^)}{n1I(β0)}1{n1J(β0)}{n1I(β0)}12{a(β^)}T(β^β0)}Ptr{Σ(β0){Σ(β0)}1Σ(β0){Σ(β0)}1}=p,

as n. Moreover, it is obvious that b 2=0. Further details are presented in Data S1. Hence, b 1p, b 3p, and bias B=b 1+b 2+b 3 can be approximated by 2p. From the aforementioned results, we define AICF as

AICF=2l(N,Y,Z;β^)+2p.

The AICF does not include the penalty term of AIC, 0.5log|I(β^)|. Even in the penalized partial likelihood setting, non‐penalized likelihood should be used for risk estimation. Sometimes, penalty terms for parameter estimation have a strong effect on a model selection criterion.

Similarly, based on the results of a previous study 14, we propose BIC for Firth's penalized partial likelihood approach

BICF=2l(N,Y,Z;β^)+plogd.

The detailed derivation is omitted, but is similar to that described previously 14.

Note that SAS and R programs for AICF and BICF are provided in Data S3.

4. Simulation

4.1. Simulation conditions

Simulation studies were conducted to investigate the performance of model selection critera (AICF, BICF, AIC, and BIC) in a monotone likelihood setting and to check that the property of AIC discussed in Section 3.3 holds. We set the simulation conditions by referring to 3 and the generated observations {N i,Y i,Z i} from exponential distributions with hazard functions hi(t)=h0(t)γi(t)=h0(t)exp{β0TZi}, where h 0(t)=1, β 0=(logθ,logθ,logθ,0,0)T, Z i=(Z i1,Z i2,Z i3,Z i4,Z i5)T, and Z ij∼Bernoulli (q). We further set the proportion of covariates as q=0.5 or 0.8, the regression parameters as θ=1.3,2,4, or 16, the proportion of censoring as c=0,50, or 90 (%), and the total sample size as n=100,200, or 1000. We generated data under simple type I censoring; the observations of each individual were censored at a suitable time τ for each simulation. Time τ was determined to achieve an expected 50% and 90% censoring. We find monotone likelihood in situations of high censoring and high parameter values. For each data configuration, we generated R=20,000 simulations. For each simulation, we calculated AICF, AIC, BICF, and BIC for the following 11 models:

  • 1

    Model 1: logγ i(t)=β 1 Z i1

  • 2

    Model 2: logγ i(t)=β 4 Z i4

  • 3

    Model 3: logγ i(t)=β 1 Z i1+β 2 Z i2

  • 4

    Model 4: logγ i(t)=β 1 Z i1+β 4 Z i4

  • 5

    Model 5: logγ i(t)=β 4 Z i4+β 5 Z i5

  • 6

    Model 6: logγ i(t)=β 1 Z i1+β 2 Z i2+β 3 Z i3(the true model)

  • 7

    Model 7: logγ i(t)=β 1 Z i1+β 2 Z i2+β 4 Z i4

  • 8

    Model 8: logγ i(t)=β 1 Z i1+β 4 Z i4+β 5 Z i5

  • 9

    Model 9: logγ i(t)=β 1 Z i1+β 2 Z i2+β 3 Z i3+β 4 Z i4

  • 10

    Model 10: logγ i(t)=β 1 Z i1+β 2 Z i2+β 4 Z i4+β 5 Z i5

  • 11

    Model 11: logγ i(t)=β 1 Z i1+β 2 Z i2+β 3 Z i3+β 4 Z i4+β 5 Z i5(the full model)

Model 6 is the true model, and Model 11 is the full model, that is, the model with maximum p.

Because it is well known that AIC is designed for optimal prediction and BIC is designed to identify the true model, we assessed the predictive performance of AICF and AIC. We evaluated the mean of the difference between the estimated mean risk (MR) and value of the information criterion in each model and its 5 and 95 percentiles, as well as the estimated MR for the selected model based on new data. The MR and its estimator, MR^, are defined as

MR=1Rr=1REÑ2l(N˜,Y˜,Z˜;β^r)

and

MR^=2Rr=1Ri=1n01β^rTZ˜ri(x)logj=1nY˜rj(x)exp{β^rTZ˜rj(x)}dÑri(x),

where ( N˜, Y˜, Z˜) is another dataset of the same size, β^r is the model estimate of each replication, and Z ri is the covariate vector in each model and replication. The absolute value of the mean difference between the MR^ and the value of the information criterion should be small because AIC is an estimator of risk function; thus, the mean difference can be regarded as an empirical bias of an AIC‐type criterion. The estimated MR for the selected model, MR^sel, is defined as

MR^sel=2Rr=1Ri=1n01β^r,selTZ˜ri(x)logj=1nY˜rj(x)exp{β^r,selTZ˜rj(x)}dÑri(x),

where β^r,sel is an estimate of the selected model. The MR^sel should be small and is considered to be a performance indicator for prediction, as it measures the goodness of fit for the selected model to a future data as a mean deviance.

To assess the performance of the criteria, we evaluated model selection probability P m, where m=1,2,…,11. We estimate the model selection probability by P^m={# of the model m selected}/{# of replication (R=20,000)}, the relative frequency of the model obtained by minimizing the information criterion.

4.2. Simulation results

The mean of the difference between the estimated MR and value of information criterion and its 5 and 95 percentiles for Models 6 and 11 for q=0.5 is shown in Table 4. The mean differences for AICF were smaller than those for AIC under all conditions. The mean differences for AIC increased with the number of events. The 5 and 95 percentiles for AICF were approximately symmetric around 0, whereas the percentiles for AIC were not symmetric. For instance, in the case with c=0%, θ=1, n=100, and Model 6 (true model), the mean difference and its percentiles for AICF and AIC were −0.4(−9.9, 5.9) and −9.8(−19.2, −3.5); in the case with c=0%, θ=1, n=100, and Model 11 (full model), the mean difference and its percentiles for AICF and AIC were −1.1(−11.2, 6.0) and −16.7(−26.7, −9.7). Thus, AIC is clearly biased downward, as AIC includes unnecessary negative terms (Section 3.3). Larger bias was observed in models with large p‐values. Therefore, these models cannot estimate the risk function. The estimated MR for the selected model based on another new dataset, MR^sel, for q=0.5 is shown in Table 4. The MR^sel for AICF was smaller than that for AIC, except in a case with c=50%, θ=4, and n=1000, and c=0%, θ=1, and n=200. Thus, the model selection based on AICF showed small deviance for future data. These results revealed that AICF shows better prediction performance.

Table 4.

The mean of the difference between the estimated MR and the value of the information criterion in each model and its 5 and 95 percentiles, and the estimated MR for the selected model (the proportion of covariates: q=0.5; the number of simulations: R=20,000).

c (%) θ n Mean difference (5 percentile, 95 percentile)
MR^sel
Model 6 (true model) Model 11 (full model)
AICF AIC AICF AIC AICF AIC
90 1 100 −0.4 (−43.9, 45.9) −2.7 (−44.2, 42.1) −1.1 (−44.4, 45.5) −4.8 (−44.7, 39.2) 94.38 95.05
90 1 200 0.1 (−71.4, 74.6) −4.4 (−74.8, 68.9) −0.1 (−71.6, 74.9) −7.7 (−77.1, 65.5) 212.87 214.55
90 1 1000 1.6 (−206.1, 218.1) −8.0 (−215.2, 208.0) 1.6 (−206.9, 218.2) −14.4 (−222.1, 201.4) 1373.64 1375.61
90 2 100 −0.4 (−43.8, 45.8) −2.7 (−44.1, 42.0) −1.0 (−44.2, 45.6) −4.7 (−44.6, 39.3) 94.40 95.04
90 2 200 −0.6 (−72.0, 74.4) −5.2 (−75.4, 68.8) −0.9 (−72.4, 74.6) −8.5 (−77.8, 65.2) 212.82 214.48
90 2 1000 1.3 (−205.9, 217.1) −8.3 (−215.1, 207.0) 1.3 (−206.6, 217.4) −14.8 (−221.8, 200.6) 1375.21 1377.21
90 4 100 −0.7 (−44.2, 45.7) −3.0 (−44.5, 41.9) −1.4 (−44.6, 45.4) −5.1 (−45.0, 39.1) 94.73 95.36
90 4 200 0.0 (−71.6, 74.2) −4.6 (−74.9, 68.6) −0.3 (−71.5, 74.3) −7.9 (−77.1, 65.0) 212.88 214.50
90 4 1000 −1.8 (−216.9, 214.8) −11.4 (−226.1, 204.7) −1.9 (−215.8, 213.4) −17.9 (−231.0, 196.7) 1376.07 1378.04
90 16 100 −0.7 (−43.4, 45.9) −3.0 (−43.8, 42.1) −1.4 (−44.0, 45.2) −5.1 (−44.4, 38.9) 94.49 95.15
90 16 200 −0.6 (−71.3, 74.0) −5.2 (−74.7, 68.3) −0.9 (−71.5, 74.0) −8.5 (−77.1, 64.6) 212.88 214.54
90 16 1000 0.1 (−207.5, 217.3) −9.5 (−216.6, 207.3) 0.1 (−207.6, 217.4) −15.9 (−222.8, 200.6) 1374.48 1376.42
50 1 100 0.1 (−64.8, 62.8) −7.4 (−71.8, 54.9) −0.4 (−65.8, 62.3) −12.8 (−77.4, 49.1) 432.37 434.56
50 1 200 −1.1 (−110.0, 104.1) −10.7 (−119.2, 94.2) −1.3 (−109.7, 104.7) −17.2 (−125.1, 88.2) 1000.80 1002.75
50 1 1000 3.0 (−322.5, 325.7) −11.5 (−336.8, 311.1) 2.9 (−322.5, 325.9) −21.2 (−346.3, 301.5) 6602.19 6603.65
50 2 100 −0.4 (−64.5, 63.2) −7.8 (−71.4, 55.2) −0.9 (−65.6, 62.8) −13.3 (−77.1, 49.7) 432.28 434.21
50 2 200 0.9 (−108.1, 107.4) −8.7 (−117.3, 97.4) 0.6 (−108.1, 106.9) −15.3 (−123.5, 90.4) 999.82 1001.35
50 2 1000 −2.9 (−322.7, 319.4) −17.4 (−337.0, 304.8) −3.0 (−322.5, 318.6) −27.1 (−346.4, 294.3) 6601.73 6601.85
50 4 100 −0.9 (−66.0, 61.9) −8.4 (−73.0, 54.0) −1.6 (−66.9, 61.3) −14.0 (−78.4, 48.2) 432.62 434.21
50 4 200 0.1 (−107.2, 104.3) −9.5 (−116.4, 94.4) −0.3 (−107.3, 103.9) −16.3 (−122.6, 87.4) 999.69 1000.66
50 4 1000 −3.0 (−323.1, 312.3) −17.4 (−337.4, 297.7) −3.2 (−323.1, 311.9) −27.3 (−347.0, 287.6) 6595.55 6595.42
50 16 100 0.1 (−64.3, 63.4) −7.3 (−71.2, 55.5) −0.7 (−65.6, 62.4) −13.1 (−77.0, 49.3) 431.82 433.14
50 16 200 −1.1 (−107.8, 104.4) −10.6 (−117.0, 94.5) −1.6 (−108.1, 104.4) −17.6 (−123.4, 87.9) 999.22 999.75
50 16 1000 0.8 (−320.5, 319.0) −13.7 (−334.8, 304.4) 0.5 (−321.0, 318.5) −23.6 (−344.8, 294.1) 6589.13 6589.34
0 1 100 −0.4 (−9.9, 5.9) −9.8 (−19.2,−3.5) −1.1 (−11.2, 6.0) −16.7 (−26.7,−9.7) 728.44 728.71
0 1 200 −0.1 (−12.1, 8.8) −11.7 (−23.7,−2.8) −0.5 (−13.2, 9.0) −19.8 (−32.4,−10.3) 1722.43 1722.27
0 1 1000 0.2 (−23.8, 21.1) −16.2 (−40.3, 4.6) 0.1 (−24.3, 21.3) −27.4 (−51.8,−6.3) 11779.42 11780.44
0 2 100 −0.5 (−17.2, 14.1) −9.6 (−26.2, 4.7) −1.2 (−18.4, 13.8) −16.6 (−33.6,−1.7) 705.35 706.37
0 2 200 −0.2 (−23.6, 20.8) −11.6 (−34.8, 9.3) −0.5 (−23.9, 20.5) −19.5 (−42.9, 1.4) 1675.67 1676.76
0 2 1000 0.4 (−48.9, 48.0) −15.9 (−65.2, 31.7) 0.3 (−49.3, 48.2) −27.0 (−76.6, 20.8) 11548.20 11549.24
0 4 100 −0.5 (−24.1, 21.7) −9.2 (−32.6, 12.8) −1.1 (−24.9, 21.4) −16.0 (−39.5, 6.2) 657.00 658.21
0 4 200 −0.5 (−33.3, 30.8) −11.4 (−44.1, 19.8) −0.9 (−34.0, 30.7) −19.5 (−52.4, 12.0) 1578.08 1579.22
0 4 1000 −0.4 (−72.1, 69.9) −16.2 (−87.8, 54.0) −0.4 (−72.1, 69.9) −27.3 (−98.8, 43.0) 11053.44 11054.48
0 16 100 −0.3 (−26.8, 26.5) −7.9 (−34.0, 18.4) −0.8 (−27.7, 26.2) −14.6 (−41.0, 12.0) 575.12 576.30
0 16 200 −0.1 (−37.4, 37.8) −10.0 (−47.0, 27.6) −0.4 (−37.8, 37.6) −17.9 (−55.1, 19.8) 1411.05 1412.15
0 16 1000 0.4 (−81.5, 84.4) −14.4 (−96.3, 69.4) 0.4 (−82.0, 84.4) −25.5 (−107.8, 58.5) 10203.57 10204.62

MR, mean risk; Mean difference (5 percentile, 95 percentile), mean of the difference between the estimated mean risk, MR^, and the value of the information criterion in each model and its 5 and 95 percentiles; MR^sel, estimated MR for the selected model based on new data; c, proportion of random censoring; θ, regression parameters; n, total sample size. The values that are superior to other are highlighted.

The selection probability of Models 6 (the true model) and 11 (the full model) for AICF and AIC when q=0.5 is shown in Table 5. For larger parameter values, larger sample sizes, and less censoring, the selection probability of the true model is larger for AICF than for AIC. For smaller parameter values, smaller sample sizes, and more censoring, the selection probability of the true model is larger for AIC than for AICF. The selection probability of the full model for AICF is smaller than that for AIC. On the contrary, the selection probability of the full model for AIC increases with the number of events. In particular, for n=1000, the selection probability of the full model for AIC is equal to one because of the term (2−logn)p in AIC, as discussed in Section 3.3.

Table 5.

The selection probability (the proportion of covariates: q=0.5; the number of simulations: R=20,000).

c (%) θ n Model 6 (true model) Model 11 (full model)
AICF AIC BICF BIC AICF AIC BICF BIC
90 1.3 100 0.060 0.106 0.050 0.087 0.008 0.039 0.007 0.016
90 1.3 200 0.058 0.121 0.026 0.094 0.005 0.239 0.001 0.016
90 1.3 1000 0.059 0.000 0.005 0.094 0.005 1.000 0.000 0.019
90 2 100 0.063 0.110 0.052 0.092 0.008 0.041 0.007 0.019
90 2 200 0.060 0.116 0.024 0.093 0.005 0.238 0.001 0.018
90 2 1000 0.062 0.000 0.006 0.096 0.004 1.000 0.000 0.018
90 4 100 0.061 0.106 0.052 0.086 0.008 0.042 0.007 0.018
90 4 200 0.061 0.115 0.026 0.092 0.006 0.239 0.001 0.018
90 4 1000 0.055 0.000 0.005 0.090 0.005 1.000 0.000 0.019
90 16 100 0.062 0.110 0.052 0.091 0.008 0.043 0.006 0.018
90 16 200 0.058 0.119 0.023 0.093 0.006 0.234 0.001 0.017
90 16 1000 0.060 0.000 0.005 0.098 0.005 1.000 0.000 0.020
50 1.3 100 0.062 0.000 0.011 0.095 0.007 1.000 0.000 0.021
50 1.3 200 0.064 0.000 0.006 0.104 0.006 1.000 0.000 0.022
50 1.3 1000 0.091 0.000 0.003 0.135 0.007 1.000 0.000 0.027
50 2 100 0.082 0.000 0.018 0.122 0.010 1.000 0.000 0.027
50 2 200 0.099 0.000 0.015 0.140 0.010 1.000 0.000 0.031
50 2 1000 0.258 0.000 0.028 0.294 0.025 1.000 0.000 0.070
50 4 100 0.109 0.000 0.030 0.148 0.017 1.000 0.000 0.039
50 4 200 0.154 0.000 0.031 0.194 0.020 1.000 0.000 0.052
50 4 1000 0.469 0.000 0.142 0.462 0.053 1.000 0.000 0.116
50 16 100 0.133 0.000 0.044 0.172 0.022 1.000 0.002 0.050
50 16 200 0.203 0.000 0.054 0.242 0.029 1.000 0.001 0.066
50 16 1000 0.596 0.000 0.305 0.544 0.072 1.000 0.000 0.143
0 1.3 100 0.270 0.000 0.087 0.304 0.033 1.000 0.001 0.067
0 1.3 200 0.468 0.000 0.188 0.468 0.054 1.000 0.001 0.108
0 1.3 1000 0.778 0.000 0.919 0.658 0.085 1.000 0.001 0.155
0 2 100 0.745 0.000 0.820 0.662 0.091 1.000 0.007 0.145
0 2 200 0.778 0.000 0.968 0.670 0.086 1.000 0.003 0.150
0 2 1000 0.786 0.000 0.991 0.670 0.081 1.000 0.001 0.149
0 4 100 0.772 0.000 0.959 0.676 0.089 1.000 0.007 0.144
0 4 200 0.773 0.000 0.971 0.664 0.085 1.000 0.004 0.150
0 4 1000 0.786 0.000 0.990 0.670 0.081 1.000 0.001 0.149
0 16 100 0.774 0.000 0.958 0.690 0.088 1.000 0.007 0.137
0 16 200 0.779 0.000 0.974 0.679 0.085 1.000 0.003 0.144
0 16 1000 0.788 0.000 0.991 0.668 0.081 1.000 0.001 0.151

Note: c, proportion of random censoring; θ, regression parameters; n, total sample size. The values that are superior to other are highlighted.

The selection probability of Models 6 (true model) and 11 (full model) for BICF and BIC when q=0.5 is shown in Table 5. For larger parameter values, larger sample sizes, and less censoring, the selection probability of the true model was larger for BICF than for BIC. For smaller parameter values, smaller sample sizes, and more censoring, the selection probability of the true model was larger for BIC than for BICF. The selection probability of the full model for BICF showed the smallest value. Although a BIC‐type criterion was designed to identify the true model as n, the selection probability of the true model of BIC was smaller than that of BICF for a large number of samples. For instance, in the case with c=0%, θ=16, and n=1000, the selection probability of the true model for BICF and BIC were 0.991 and 0.668, while the selection probability of the full model for BICF and BIC were 0.001 and 0.151. This is because of the properties of BIC discussed in Section 3.3.

The results of the other models and conditions are presented in Data S2 (Tables S1–S8 for bias, Table S9 for prediction performance, and Tables S10–S17 for selection probability); these results reveal the same tendencies as discussed previously.

In summary, AICF can estimate the risk function and shows better prediction performance than AIC. By contrast, because AIC selects models that have many parameters and ignores the adequacy of the model as the number of events increase, it is not as efficient and not recommended for model selection. BICF can select the true model as n. In contrast, because the selection probability of the true model for BIC was smaller than that for BICF for a large number of samples, this method is not as efficient in selecting the true model.

5. Application of the prospective observational study data

Next, we apply the proposed method to the study introduced in Section 2. We select a model based on AICF as in that section, with the result shown in Table 6. As AIC tends to select models that have many parameters (Table 2), the model that includes ntumor (number of tumors), ptumor (primary tumor category), and neuro (neurological symptoms) was selected by AICF as the best. The best model under AICF does not include unnecessary covariates such as age and status whose hazard ratios were almost equal to one (Table 2).

Table 6.

The top five models based on AICF for the study data.

Model AICF
ntumor ptumor neuro 1753.84
ntumor neuro 1754.72
sex ntumor ptumor neuro 1754.90
ntumor diameter ptumor neuro 1755.83
ntumor ptumor status neuro 1755.83

Note: ntumor, number of tumors; diameter, maximum diameter of largest tumor; ptumor, primary tumor category; status, extracerebral disease status; neuro, neurological symptoms; AICF, AIC for Firth's penalized partial likelihood approach.

The estimated hazard ratios for the best model are shown in Table 7. The best model under AICF included ntumor(1 vs. 2–4: H R=0.71, p‐value=0.08; 5–10 vs. 2–4: H R=1.57, p‐value=0.04), ptumor (breast vs. lung: H R=0.94, p‐value=0.82; gastrointestinal vs. lung: H R=1.60, p‐value=0.21; kidney vs. lung: H R=0.12, p‐value=0.02; others vs. lung: H R=0.87, p‐value=0.79), and neuro (H R=1.53, p‐value=0.04). The selected covariates under AICF are better than age (Table 3; H R=1.00, p‐value=0.98) and status(Table 3; H R=1.03, p‐value=0.85). Therefore, the best model under AICF seems to be plausible and may be more appropriate than the best model under AIC. Additionally, the best model under AICF is clinically interpretable because the selected variables are prognostic factors for brain metastases 20, 21, 22.

Table 7.

Parameter estimates of the best model based on AICF for the study data.

Covariate HR SE 95% CI p (LR)
ntumor
   1 vs. 2–4 0.71 0.19 0.49 1.05 0.08
   5–10 vs. 2–4 1.57 0.21 1.03 2.38 0.04
ptumor
   Breast vs. lung 0.94 0.26 0.57 1.57 0.82
   GI vs. lung 1.60 0.36 0.79 3.26 0.21
   Renal cell vs. lung 0.12 1.43 0.01 1.99 0.02
   Others vs. lung 0.87 0.55 0.30 2.56 0.79
neuro (yes vs. no) 1.53 0.20 1.04 2.24 0.04

Note: HR, hazard ratio; ntumor, number of tumors; ptumor, primary tumor category; neuro, neurological symptoms; GI, gastrointestinal; LR, likelihood ratio; SE, standard error.

6. Discussion and conclusion

One solution to the monotone likelihood problem, which is an important issue in Cox regression models, is Firth's penalized partial likelihood approach. However, the model selection criteria for this approach are yet to be studied, and heuristic criteria, AIC and BIC, are used in the SAS PHREG procedure. Therefore, in this study, we proposed alternative criteria, AICF and BICF, which work for Firth's penalized partial likelihood approach. Moreover, we discussed the justification for adopting Firth's bias correction method in Cox regression models.

We showed that AICF, an estimator of the risk function based on KL information, does not include the penalty term of AIC, 0.5log|I(β^)|. Even in the penalized partial likelihood setting, non‐penalized likelihood should be used for risk estimation. In addition, AICF is more efficient than AIC in simulations and works well when addressing monotone likelihood. The simulation results revealed systematic bias in AIC, and the model selection based on AICF showed superior predictive performance. In any case, AIC is not recommended for model selection. An application using real data concluded that AICF has better properties than AIC and that the latter leads to incorrect results. Moreover, we showed that BIC has a negative penalty term in proportion to the number of regression parameters. The selection probability of the true model for BIC was smaller than that of BICF for a large number of samples, indicating that it is not as efficient for selecting the true model. In summary, we showed that AICF or BICF is appropriate for model selection in monotone likelihood cases. AICF would be used for prediction, while BICF can be used for selecting the true model.

Because we obtained impressive results with alternative criteria, future studies should aim to examine other model evaluation criteria such as the C‐index 23, 24. Moreover, a similar problem occurs if one uses AIC and BIC based on the penalized log‐likelihood under separation in logistic regression models.

Appendix A. Justification for using Firth's bias correction method in Cox regression models

Firth's bias correction method is based on the relationships κ j,k+κ jk=0, κ j,k,l+κ j,kl+κ k,jl+κ l,jk+κ j,k,l=0, and κ j,kl=0, as discussed in Section 3.2. However, the relationships κ j,k,l+κ j,kl+κ k,jl+κ l,jk+κ j,k,l=0 and κ j,kl=0 have not yet been evaluated in Cox regression models. Therefore, we only prove that κ j,kl=0; it is well known that κ j,k+κ jk=0, and we can easily show that κ j,k,l+κ jkl=0.

Lemma 1 (κj,kl=0.)

If we insert β 0 in the functions U j(β) and U kl(β),

Uj(β0)=i=1n01Hij(x)dMi(x),
Hij(t)=Zij(t)Sj(1)(β0,t)S(0)(β0,t),

and

Ukl(β0)=Ukl(β0)(1)+i=1n01Gkl(x)dMi(x),
Gkl(t)=Skl(2)(β0,t)S(0)(β0,t)Sk(1)(β0,t)Sl(1)(β0,t){S(0)(β0,t)}2,

where Sj(1)(β,t)=n1i=1nYi(t)Zij(t)exp{βTZi(t)}, Skl(2)(β,t)=n1i=1nYi(t)Zik(t)Zil(t)exp{βTZi(t)}, and Ukl(β0)(t)=n0tGkl(x)S(0)(β0,x)h0(x)dx.

We note that H ij and G kl are predictable processes according to the assumption made in Section 3.1. Here,

Uj(β0)Ukl(β0)=Ukl(β0)(1)i=1n01Hij(x)dMi(x)+i=1n01Hij(x)dMi(x)i=1n01Gkl(x)dMi(x).

From theorem 2.4.4 of 25, we have

κj,kl=n1EUkl(β0)(1)i=1n01Hij(x)dMi(x)+h=1ni=1n01Hhj(x)Gkl(x)dMh,Mi(x).

Because M i(t) is the counting process martingale,

Ei=1n01Hij(x)dMi(x)=0,

and from the orthogonality of martingales

Mh,Mi(t)=0tYh(x)exp{β0TZh(x)}h0(x)dx,h=i0,hi

It follows that

κj,kl=0+n1E01i=1nZij(x)Sj(1)(β0,x)S(0)(β0,x)Yi(x)exp{β0TZi(x)}Gkl(x)h0(x)dx=0.

Therefore, the relationships κ j,k+κ jk=0, κ j,k,l+κ j,kl+κ k,jl+κ l,jk+κ jkl=0, and κ j,kl=0 are also true in Cox regression models.

Appendix B. Consistency of β^

In this section, we discuss the consistency of β^. The following list of conditions will be assumed:

(A)

01h0(x)dx<.

(B)
There exists a neighborhood B of β 0 and a scalar function, s (0), a vector function, s (1), and a matrix function, s (2), defined on B×[0,1] such that for k=0,1,2,
supt[0,1],βB||S(k)(β,t)s(k)(β,t)||P0.
(C)
There exists δ>0 such that
n1/2supt[0,1],1in|Zi(t)|Yi(t)1{β0TZi(t)>δ|Zi(t)|}P0,
where 1{} is an indicator function.
(D)

For all βB, t∈[0,1]: s (0)(β,t), s (1)(β,t)= s (0)(β,t)/ β, and s (2)(β,t)= s (1)(β,t)/ β T are continuous functions of βB, uniformly in t∈[0,1], s (0), s (1), and s (2) are bounded on B×[0,1]; s (0) is bounded away from zero on B×[0,1], and the matrix Σ(β 0) is positive definite.

These conditions are identical to those given by Andersen and Gill 12.

Lemma 2 ( β^β0.)

Consider the process

Xn(β,1)=n1{l(N,Y,Z;β)l(N,Y,Z;β0)}=Xn(β,1)+0.5n1{log|I(β)|log|I(β0)|},

where X n(β,1)=n −1{l(N,Y,Z;β)−l(N,Y,Z;β 0)}. To prove the consistency of the usual Cox regression estimator β^Cox, Andersen and Gill 12 showed that X n(β,1) is a concave function and proved that X n(β,1) converges in probability to

A(β,1)=01(ββ0)Ts(1)(β0,x)logs(0)(β,x)s(0)(β0,x)s(0)(β0,x)h0(t)dx,

for each βB. The first derivative of A(β,1) is zero at β=β 0, and the second derivative is minus a positive definite matrix. In other words, A(β,1) is a concave function of β with a unique maximum at β=β 0. If X n(β,1) is a concave function, then β^CoxPβ0 by applying theorem II.1 of 12.

In the same manner, if Xn(β,1) is concave and Xn(β,1) converges to a concave function of β with a unique maximum at β=β 0, the consistency of β^ can be shown by applying theorem II.1 of 12.

In monotone likelihood settings, the partial log‐likelihood function converges to a finite value. Fixing β 1,β 2,…,β j−1,β j+1,…,β p to be β^1,β^2,,β^j1,β^j+1,,β^p, for the parameter β j, a real number c, and a constant d,

c,βjc,l(N,Y,Z;β^1,β^2,,β^j1,βj,β^j+1,,β^p)=d,

or

c,βjc,l(N,Y,Z;β^1,β^2,,β^j1,βj,β^j+1,,β^p)=d,

in monotone likelihood settings. Therefore, X n(β,1) is a concave function. Note that l(N,Y,Z;β 0) and log|I(β 0)| are obviously finite constants. Thus, it is sufficient to show that log|I(β)| is a concave function. Now, I(β) is a positive semidefinite matrix (see also Prentice 26) because l(N,Y,Z;β) is a concave function. Generally, the function log|C|, where |C| is the determinant of a positive semidefinite matrix C, is concave 27. Therefore, Xn(β,1) is a sum of concave functions.

We next show that Xn(β,1) converges in probability to A(β,1). According to the aforementioned result, X n(β,1) converges in probability to A(β,1). Conditions (B) and (D) imply that, for each βB, n1I(β)PΣ(β). Thus,

log|I(β)|log|I(β0)|=log|n·n1I(β)|log|n·n1I(β0)|=log|n1I(β)|log|n1I(β0)|=log{np|I(β)|}log{np|I(β0)|}

converges in probability to some finite quantity. Therefore,

0.5n1log|I(β)|log|I(β0)|P0,

and Xn(β,1) converges in probability to A(β,1), which is a concave function of β with a unique maximum at β=β 0.

These facts establish that β^Pβ0 as n.

Moreover, from this consistency, we can apply theorem 3.2 of 12; therefore, n1I(β^)PΣ(β0) in monotone likelihood settings.

Supporting information

Supplementary Table S1: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.5, Model 1–6)

Supplementary Table S2: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.5, Model 7–11)

Supplementary Table S3: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.5, Model 1–6)

Supplementary Table S4: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.5, Model 7–11)

Supplementary Table S5: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.8, Model 1–6)

Supplementary Table S6: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.8, Model 7–11)

Supplementary Table S7: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.8, Model 1–6)

Supplementary Table S8: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.8, Model 7–11)

Supplementary Table S9: The estimated MR for the selected model MR^

Supplementary Table S10: The selection probability of AICF (q = 0.5)

Supplementary Table S11: The selection probability of AIC* (q = 0.5)

Supplementary Table S12: The selection probability of BICF (q = 0.5)

Supplementary Table S13: The selection probability of BIC* (q = 0.5)

Supplementary Table S14: The selection probability of AICF (q = 0.8)

Supplementary Table S15: The selection probability of AIC* (q = 0.8)

Supplementary Table S16: The selection probability of BICF (q = 0.8)

Supplementary Table S17: The selection probability of BIC* (q = 0.8)

Code 1: A SAS program for calculating AICF and BICF

Code 2: An R program for calculating AICF and BICF

Acknowledgements

The authors would like to thank the associate editor and two reviewers for very insightful and constructive comments that substantially improved the article. The authors would like to thank Professor Masaaki Yamamoto of Katsuta Hospital Mito Gamma House and Dr. Toru Serizawa of Tsukiji Neurological Clinic for providing the dataset of the JLGK0901 study. This work was supported by the Japan Society for the Promotion of Science KAKENHI grant numbers 26870099 and 16K16014 and the MHLW‐funded Core Clinical Research Centers.

Nagashima, K. , and Sato, Y. (2017) Information criteria for Firth's penalized partial likelihood approach in Cox regression models. Statist. Med., 36: 3422–3436. doi: 10.1002/sim.7368.

The copyright line for this article was changed on 20 November 2017 after original online publication.

References

  • 1. Cox DR. Regression models and life tables (with discussion). Journal of the Royal Statistical Society, Series B 1972; 34(2):187‐220. [Google Scholar]
  • 2. Bryson MC, Johnson ME. The incidence of monotone likelihood in the Cox model. Technometrics 1981; 23(4):381‐383. [Google Scholar]
  • 3. Heinze G, Schemper M. A solution to the problem of monotone likelihood in Cox regression. Biometrics 2001; 57(1):114‐119. [DOI] [PubMed] [Google Scholar]
  • 4. Firth D. Bias reduction of maximum likelihood estimates. Biometrika 1993; 80(1):27‐38. [Google Scholar]
  • 5. Heinze G. The application of Firth's procedure to Cox and logistic regression, Technical Report 10/1999, update in January 2001, Section of Clinical Biometrics, Department of Medical Computer Sciences University of Vienna, 2001. [Google Scholar]
  • 6. Heinze G. A comparative investigation of methods for logistic regression with separated or nearly separated data. Statistics in Medicine 2006; 25(24):4216‐4226. [DOI] [PubMed] [Google Scholar]
  • 7. Heinze G, Schemper M. A solution to the problem of separation in logistic regression. Statistics in Medicine 2002; 21(16):2409‐2419. [DOI] [PubMed] [Google Scholar]
  • 8. Akaike H. Information theory and an extension of the maximum likelihood principle In In Second International Symposium on Information Theory, Petrov BN, Csaki F, eds. Akademiai Kiado: Budapest, 1973; 267‐281. [Google Scholar]
  • 9. Schwarz G. Estimating the dimension of a model. The Annals of Statistics 1978; 6(2):461‐464. [Google Scholar]
  • 10. Yamamoto M, Serizawa T, Shuto T, Akabane A, Higuchi Y, Kawagishi J, Yamanaka K, Sato Y, Jokura H, Yomo S, Nagano O, Kenai H, Moriki A, Suzuki S, Kida Y, Iwai Y, Hayashi M, Onishi H, Gondo M, Sato M, Akimitsu T, Kubo K, Kikuchi Y, Shibasaki T, Goto T, Takanashi M, Mori Y, Takakura K, Saeki N, Kunieda E, Aoyama H, Momoshima S, Tsuchiya K. Stereotactic radiosurgery for patients with multiple brain metastases (JLGK0901): a multi‐institutional prospective observational study. The Lancet Oncology 2014; 15(4):387‐395. [DOI] [PubMed] [Google Scholar]
  • 11. Kalbfleisch JD, Prentice RL. The Statistical Analysis of Failure Time Data. 2nd ed John Wiley & Sons: New Jersey, 2002. [Google Scholar]
  • 12. Andersen PK, Gill RD. Cox's regression model for counting processes: a large sample study. The Annals of Statistics 1982; 10(4):1100‐1120. [Google Scholar]
  • 13. Cox DR, Snell EJ. A general definition of residuals. Journal of the Royal Statistical Society, Series B 1968; 30(2):248‐275. [Google Scholar]
  • 14. Volinsky CT, Raftery AE. Bayesian information criterion for censored survival models. Biometrics 2000; 56(1):256‐262. [DOI] [PubMed] [Google Scholar]
  • 15. Xu R, Vaida F, Harrington DP. Using profile likelihood for semiparametric model selection with application to proportional hazards mixed models. Statistica Sinica 2009; 19(2):819‐842. [PMC free article] [PubMed] [Google Scholar]
  • 16. Claeskens G, Carroll RJ. An asymptotic theory for model selection inference in general semiparametric problems. Biometrika 2007; 94(2):249‐265. [Google Scholar]
  • 17. Bailey KR. The asymptotic joint distribution of regression and survival parameter estimates in the Cox regression model. The Annals of Statistics 1983; 11(1):39‐48. [Google Scholar]
  • 18. Johansen S. An extension of Cox's regression model. International Statistical Review 1983; 51(2):165‐174. [Google Scholar]
  • 19. Murphy SA, van der Vaart AW. On profile likelihood. Journal of the American Statistical Association 2000; 95(450):449‐465. [Google Scholar]
  • 20. Joseph J, Adler JR, Cox RS, Hancock SL. Linear accelerator‐based stereotaxic radiosurgery for brain metastases: the influence of number of lesions on survival. Journal of Clinical Oncology 1996; 14(4):1085‐1092. [DOI] [PubMed] [Google Scholar]
  • 21. Balm M, Hammack J. Leptomeningeal carcinomatosis. Presenting features and prognostic factors. Archives of Neurology 1996; 53(7):626‐632. [DOI] [PubMed] [Google Scholar]
  • 22. Sperduto PW, Kased N, Roberge D, Xu Z, Shanley R, Luo X, Sneed PK, Chao ST, Weil RJ, Suh J, Bhatt A, Jensen AW, Brown PD, Shih HA, Kirkpatrick J, Gaspar LE, Fiveash JB, Chiang V, Knisely JP, Sperduto CM, Lin N, Mehta M. Summary report on the graded prognostic assessment: an accurate and facile diagnosis‐specific tool to estimate survival for patients with brain metastases. Journal of Clinical Oncology 2012; 30(4):419‐425. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Harrell FE, Lee KL, Mark DB. Multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Statistics in Medicine 1996; 15(4):361‐387. [DOI] [PubMed] [Google Scholar]
  • 24. Uno H, Cai T, Pencina MJ, D'Agostino RB, Wei LJ. On the C‐statistics for evaluating overall adequacy of risk prediction procedures with censored survival data. Statistics in Medicine 2011; 30(10):1105‐1117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Fleming TR, Harrington DP. Counting Processes and Survival Analysis. John Wiley & Sons: New York, 1991. [Google Scholar]
  • 26. Prentice RL, Self SG. Asymptotic distribution theory for Cox‐type regression models with general relative risk form. The Annals of Statistics 1983; 11(3):804‐813. [Google Scholar]
  • 27. Cover TM, Thomas JA. Determinant inequalities via information theory. SIAM Journal on Matrix Analysis and Applications 1988; 9(3):384‐392. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Table S1: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.5, Model 1–6)

Supplementary Table S2: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.5, Model 7–11)

Supplementary Table S3: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.5, Model 1–6)

Supplementary Table S4: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.5, Model 7–11)

Supplementary Table S5: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.8, Model 1–6)

Supplementary Table S6: The mean of the difference between the estimated MR and AICF, and its 5 and 95 percentiles (q = 0.8, Model 7–11)

Supplementary Table S7: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.8, Model 1–6)

Supplementary Table S8: The mean of the difference between the estimated MR and AIC*, and its 5 and 95 percentiles (q = 0.8, Model 7–11)

Supplementary Table S9: The estimated MR for the selected model MR^

Supplementary Table S10: The selection probability of AICF (q = 0.5)

Supplementary Table S11: The selection probability of AIC* (q = 0.5)

Supplementary Table S12: The selection probability of BICF (q = 0.5)

Supplementary Table S13: The selection probability of BIC* (q = 0.5)

Supplementary Table S14: The selection probability of AICF (q = 0.8)

Supplementary Table S15: The selection probability of AIC* (q = 0.8)

Supplementary Table S16: The selection probability of BICF (q = 0.8)

Supplementary Table S17: The selection probability of BIC* (q = 0.8)

Code 1: A SAS program for calculating AICF and BICF

Code 2: An R program for calculating AICF and BICF


Articles from Statistics in Medicine are provided here courtesy of Wiley

RESOURCES