Skip to main content
Journal of Applied Statistics logoLink to Journal of Applied Statistics
. 2019 Sep 24;47(6):1084–1108. doi: 10.1080/02664763.2019.1669542

A bivariate inverse Weibull distribution and its application in complementary risks model

Shuvashree Mondal 1, Debasis Kundu 1,CONTACT
PMCID: PMC9042135  PMID: 35706915

ABSTRACT

In reliability and survival analysis the inverse Weibull distribution has been used quite extensively as a heavy tailed distribution with a non-monotone hazard function. Recently a bivariate inverse Weibull (BIW) distribution has been introduced in the literature, where the marginals have inverse Weibull distributions and it has a singular component. Due to this reason this model cannot be used when there are no ties in the data. In this paper we have introduced an absolutely continuous bivariate inverse Weibull (ACBIW) distribution omitting the singular component from the BIW distribution. A natural application of this model can be seen in the analysis of dependent complementary risks data. We discuss different properties of this model and also address the inferential issues both from the classical and Bayesian approaches. In the classical approach, the maximum likelihood estimators cannot be obtained explicitly and we propose to use the expectation maximization algorithm based on the missing value principle. In the Bayesian analysis, we use a very flexible prior on the unknown model parameters and obtain the Bayes estimates and the associated credible intervals using importance sampling technique. Simulation experiments are performed to see the effectiveness of the proposed methods and two data sets have been analyzed to see how the proposed methods and the model work in practice.

KEYWORDS: Marshall–Olkin bivariate distribution, block and basu bivariate distribution, maximum likelihood estimation, EM algorithm, Gamma-Dirichlet prior, complementary risk

AMS SUBJECT CLASSIFICATIONS: Primary: 62E15, Secondary: 62H10

1. Introduction

The Marshall–Olkin [16] bivariate exponential distribution (MOBE) is one of the popular bivariate exponential distributions. The MOBE is a singular distribution with exponentially distributed marginals. The MOBE distribution can be used quite effectively when there are ties in the data set, but if there are no ties, it cannot be used. Block and Basu [2] introduced an absolutely continuous bivariate exponential (BBBE) distribution by removing the singular component from the MOBE distribution and it can be used to analyze a data set when there are no ties. Similar to the BBBE distribution Block and Basu bivariate Weibull distribution (BBBW) also has been studied in the literature and its properties and inferential procedures have been developed, see for example Kundu and Gupta [12] and Pradhan and Kundu [19] and the references cited therein.

Although the Weibull distribution is a very flexible lifetime distribution, it cannot have non-monotone hazard function. In this respect, the inverse Weibull (IW) distribution can be used quite effectively if the data come from a distribution which has a non-monotone hazard function. It is well known that if 0<α1, where α is the shape parameter of an IW distribution as defined in (1), then the mean does not exist, and if 1<α2, then the mean exists but the variance does not exist. Hence, it can be used if the data come from a heavy tailed distribution also with a suitable choice of the shape parameter. Recently, Muhammed [17] and Kundu and Gupta [14] independently introduced a bivariate inverse Weibull (BIW) distribution with the marginals having the IW distributions and discussed different inferential issues of the proposed model. Although, it has been observed that the BIW is a very flexible model, it also has a singular component. Hence, the BIW distribution can be used to analyze a data set if there are ties, similar to the MOBE or Marshall Olkin bivariate Weibull distribution. The main aim of this paper is to introduce an absolutely continuous bivariate inverse Weibull (ACBIW) distribution and use it as a dependent complementary risks model.

In a reliability or in a survival analysis, very often experimental units are exposed to more than one causes of failure and one needs to analyze the effect of one cause in presence of the other causes. In this case one observes the failure time along with the cause of failure. In the statistical literature it is popularly known as the competing risk problem. In a competing risk scenario the failure time T can be written as T=min(X1,X2,,Xm) where Xi denotes the latent failure time due to the i-th risk factor for i=1,2,,m. Here Xi's cannot be observed separately, hence, called latent failure times. An extensive amount of work has been done in developing different competing risks models, analyzing their properties and providing various inferential procedures. Interested readers are referred to the book length treatment of Crowder [4]. Recently, a growing interest is on developing different dependent competing risks models, see for example Feizjavadian and Hashemi [7], Shen and Xu [20] and the references cited therein.

The duality of the competing risk model is known as the complementary risk model and it was originally introduced by Basu and Ghosh [1]. In this case the failure time T can be written as T=max(X1,X2,,Xm), where Xi's are same as defined before. Although an extensive amount of work has been done in the area of competing risks not much work been done in the area of the complementary risks, although survival or reliability data in presence of complementary risks are observed in several areas such as public health, actuarial sciences, biomedical studies, demography and in industrial reliability, see for example Louzada et al. [15] and Han [8]. It is observed that the proposed ACBIW can be used quite effectively as a dependent complementary risks model.

In this paper first we define the ACBIW distribution and obtain its marginals. We provide different properties of the proposed distribution. The distribution has four parameters and the maximum likelihood estimators (MLEs) cannot be obtained in closed form. One has to solve a four dimensional optimization problem to compute the MLEs. To avoid that we have applied the expectation maximization (EM) algorithm, and it is observed that at each ‘E’-step the corresponding ‘M’-step can be performed by solving a one dimensional optimization problem. We have further considered the Bayesian inference of the unknown parameters based on a very general Gamma-Dirichlet (GD) prior on the scale parameters and an independent log-concave prior on the shape parameter. The Bayes estimators and the associated credible intervals of the unknown parameters are obtained based on importance sampling technique. Extensive simulations have been performed to see the effectiveness of the proposed method and one bivariate data set has been analyzed to show how the model can be used in practice.

We further describe how this model can be used as a dependent complementary risks model. We modify both the EM algorithm and the Bayesian algorithm to analyze dependent complementary risks data. We have analyzed one real complementary risks data set to show how the methods can be used in practice.

The rest of the paper is arranged as follows. In Section 2 we describe the model. Different properties of the ACBIW distribution are discussed in Section 3. In Section 4 we propose the EM algorithm to compute the MLEs and in Section 5 the Bayesian procedure has been provided. Simulation experiments and the analysis of a bivariate data set are provided in Section 6. The application of the ACBIW distribution in complementary risk model is discussed in Section 7 along with a real data analysis. Finally, we conclude the paper in Section 8.

2. Model description

A random variable X is said to follow an inverse Weibull distribution with shape parameter α and scale parameter λ, denoted by XIW(α,λ) if its cumulative distribution function (CDF) is of the form

FIW(x;α,λ)={eλxαifx>0,0otherwise, (1)

and the probability density function (PDF) is as follows

fIW(x;α,λ)={αλx(α+1)eλxαifx>0,0otherwise.

The BIW can be defined as follows. Supposes U1,U2,U3 are three independent random variables, where Ui IW(α,λi), for i=1, 2, 3. If Y1=max(U1,U3) and Y2=max(U2,U3), then (Y1,Y2) is said to follow a BIW distribution with parameters α,λ1,λ2,λ3. The joint CDF of (Y1,Y2) where (Y1,Y2) BIW(α,λ1,λ2,λ3) can be written as

FY1,Y2(y1,y2)={λ1+λ2λ1+λ2+λ3Fa(y1,y2)+λ3λ1+λ2+λ3Fs(y1,y2)ify1,y2>0,0otherwise.

Here,

Fs(y1,y2)=e(λ1+λ2+λ3)zαandFa(y1,y2)=λ1+λ2+λ3λ1+λ2eλ1y1αλ2y2αλ3zαλ3λ1+λ2e(λ1+λ2+λ3)zα,

where z=min(y1,y2).

Therefore FY1,Y2(,) consists of Fs() and Fa() where Fs() is the singular part and Fa() is the absolute continuous part. The joint PDF of (Y1,Y2) can be written as

fY1,Y2(y1,y2)={λ1+λ2λ1+λ2+λ3fa(y1,y2)+λ3λ1+λ2+λ3fs(z)ify1,y2>0,0otherwise,

where,

fa(y1,y2)=λ1+λ2+λ3λ1+λ2{fIW(y1;α,λ1+λ3)fIW(y2,α,λ2)if0<y1<y2,fIW(y1;α,λ1)fIW(y2,α,λ2+λ3)if0<y2<y1,

and

fs(y)=fIW(y;α,λ1+λ2+λ3).

The ACBIW distribution can be obtained by omitting the singular part of the BIW distribution and it can be defined as follows. (X1,X2) is said to follow a ACBIW(α,λ1,λ2,λ3), if the joint CDF is of the form

FACBIW(x1,x2)={λ1+λ2+λ3λ1+λ2eλ1x1αλ2x2αλ3uαλ3λ1+λ2e(λ1+λ2+λ3)uαifx1,x2>00otherwise, (2)

where, u=min(x1,x2).

Therefore, the joint PDF can be written as

fACBIW(x1,x2)={λ1+λ2+λ3λ1+λ2fIW(x1;α,λ1+λ3)fIW(x2,α,λ2)if0<x1<x2,λ1+λ2+λ3λ1+λ2fIW(x1;α,λ1)fIW(x2,α,λ2+λ3)if0<x2<x1,0otherwise. (3)

Note that ACBIW has been obtained by removing the singular part from a BIW distribution, and then normalizing it to make a proper distribution function. The main advantage of the ACBIW distribution is that it is an absolutely continuous distribution with respect to two dimensional Lebesgue measure, and it can be used quite effectively to analyze bivariate lifetime data when there are no ties.

The following result provides the shape of the joint PDF of ACBIW distribution.

Theorem 2.1

Let (X1,X2) ACBIW(α,λ1,λ2,λ3). Then

  1. If λ1=λ2, then fACBIW(x1,x2) is continuous for 0<x1,x2<. Moreover, fACBIW(x1,x2) is unimodal and the mode is at (xm,xm), where xm={α(2λ1+λ3)/(2(α+1))}1/α

  2. If λ1+λ3<λ2, then fACBIW(x1,x2) is not continuous at x1=x2. Further, fACBIW(x1,x2) is unimodal and the mode is at (x1m,x2m), where x1m={α(λ1+λ3)/(α+1))}1/α and x2m={αλ2/(α+1)}1/α.

  3. If λ2+λ3<λ1, then fACBIW(x1,x2) is not continuous at x1=x2. Further, fACBIW(x1,x2) is unimodal and the mode is at (x1m,x2m), where x1m={αλ1/(α+1)}1/α and x2m={α(λ2+λ3)/(α+1))}1/α and.

Proof.

It can be easily obtained, see for example Kundu and Gupta [14].

From Theorem 2.1, it is observed that the joint PDF of a ACBIW is unimodal and it can take various shapes. We have provided the surface plots of the joint PDFs of a ACBIW distribution for different parameter values in Figures 15. It is observed that it can be heavy tailed also. From the derivation of the ACBIW it is evident that if (Y1,Y2) ∼ BIW(α,λ1,λ2,λ3) and

(X1,X2)=d(Y1,Y2)|Y1Y2,

then (X1,X2) ACBIW(α,λ1,λ2,λ3). Here, =d means equal in distribution. Therefore, the following algorithm can be applied to generate a random sample from the ACBIW distribution.

Figure 2.

Figure 2.

Surface plot of the joint PDF of ACBIW distribution.

Figure 3.

Figure 3.

Surface plot of the joint PDF of ACBIW distribution.

Figure 4.

Figure 4.

Surface plot of the joint PDF of ACBIW distribution.

Figure 1.

Figure 1.

Surface plot of the joint PDF of ACBIW distribution.

Figure 5.

Figure 5.

Surface plot of the joint PDF of ACBIW distribution.

Algorithm 1

  1. Generate U1,U2,U2 independently where Ui IW(α,λi) for i=1, 2, 3.

  2. If U3>U1 and U3>U2 go back to Step 1, otherwise set X1=max(U1,U3) and X2=max(U2,U3).

3. Different properties of the marginals

In this section we provide different properties of the marginals of the ACBIW distribution. The following result gives the marginal distributions of a ACBIW distribution.

Theorem 3.1

If (X1,X2)ACBIW(α,λ1,λ2,λ3), then the marginal PDFs of X1 and X2 are

fX1(x1)={λ1+λ2+λ3λ1+λ2fIW(x1;α,λ1+λ3)λ3λ1+λ2fIW(x1;α,λ1+λ2+λ3)ifx1>0,0otherwise, (4)

and

fX2(x2)={λ1+λ2+λ3λ1+λ2fIW(x2;α,λ2+λ3)λ3λ1+λ2fIW(x2;α,λ1+λ2+λ3)ifx2>0,0otherwise, (5)

respectively.

Proof.

The marginal PDFs can be derived from (3) using simple integration.

Therefore, the marginal distribution functions can be obtained as

FX1(x1)={λ1+λ2+λ3λ1+λ2e(λ1+λ3)x1αλ3λ1+λ2e(λ1+λ2+λ3)x1αifx1>0,0otherwise, (6)
FX2(x2)={λ1+λ2+λ3λ1+λ2e(λ2+λ3)x2αλ3λ1+λ2e(λ1+λ2+λ3)x2αifx2>0,0otherwise, (7)

Note that, the marginal distributions are the generalized mixture of the IW distributions. Therefore, k th moment of the marginals will exist only when α>k. It is observed that the PDFs of the marginals are unimodal or a decreasing function depending on the values of α. The hazard functions of the marginals are also either unimodal or a decreasing function.

The next result provides the conditional PDFs and CDFs of the marginals of the ACBIW distribution and it will be needed in the application section.

Theorem 3.2

If (X1,X2)ACBIW(α,λ1,λ2,λ3), then the conditional PDF of X1|X2=x2 is given by

fX1|X2=x2(x1)={λ2eλ3x2α(λ2+λ3(1eλ1x2α))fIW(x1;α,λ1+λ2)for0<x1<x2,λ2+λ3λ2+λ3(1eλ1x2α)fIW(x1;α,λ1)for0<x2<x1,0otherwise, (8)

and the conditional PDF of X2|X1=x1 is given by

fX2|X1=x1(x2)={λ1+λ3λ1+λ3(1eλ2x1α)fIW(x2;α,λ2)for0<x1<x2,λ1eλ3x1α(λ1+λ3(1eλ2x1α))fIW(x2;α,λ2+λ3)for0<x2<x1,0otherwise. (9)

Proof.

The results can be derived using (3), (4) and (5).

Some of the following properties will be useful for data analysis purposes and they will be used in Section 6.

Theorem 3.3

If (X1,X2)ACBIW(α,λ1,λ2,λ3), then

  1. P(X1<X2)=λ2/(λ1+λ2),

  2. max(X1,X2)IW(α,λ1+λ2+λ3),

  3. X1|X1>X2IW(α,λ1+λ2+λ3),

  4. X2|X2>X1IW(α,λ1+λ2+λ3).

Proof.

See in Appendix.

From the above result it is interesting to observe that, though any of the marginals are not IW variables, maximum of the two marginals follows a IW distribution. The marginal distribution of one variable provided it is greater than the other variable follows a IW distribution.

4. Maximum likelihood estimation

In this section we derive the MLEs of the unknown parameters α,λ1,λ2,λ3 of a ACBIW distribution when we have a random sample of size n. The data are of the form Data={(x11,x21),,(x1n,x2n)}. We introduce two sets I1,I2, where I1={i:x1i<x2i} and I2={i:x1i>x2i}; |I1|=n1,|I2|=n2. Based on the above data the log-likelihood function can be written as

l(θ|Data)=2nlnα+nln(λ1+λ2+λ3)nln(λ1+λ2)+n1ln(λ1+λ3)+n1lnλ2+n2lnλ1+n2ln(λ2+λ3)(α+1)iI1I2(lnx1i+lnx2i)(λ1+λ3)iI1x1iαλ2iI1x2iαλ1iI2x1iα(λ2+λ3)iI2x2iα. (10)

Here θ=(α,λ1,λ2,λ3). The MLEs of the unknown parameters can be obtained by maximizing (10) with respect to the unknown parameters. Clearly, they cannot be obtained in explicit forms and they have to be obtained by solving a four dimensional optimization problem. To avoid that we propose to use the EM algorithm using the missing information principle. The basic idea behind the proposed EM algorithm is the following. Let us go back to the basic formulation of the ACBIW distributions based on U1,U2 and U3. We will show that along with (X1,X2) if we had observed the associated U1,U2 and U3, then the MLEs of the unknown parameters can be obtained by solving a one dimensional optimization problem. Let us assume that along with (X1,X2), we observe (δ1,δ2), where δ1=i if X1=Ui and δ2=j if X2=Uj, for j=1, 2, 3. It is assumed that we have the complete observation as follows: {((x1i,δ1i),(x2i,δ2i)):i=1,,n}. Let us use the following notations.

I312={i:x1i<x2i,δ1i=1,δ2i=2}I132={i:x1i<x2i,δ1i=3,δ2i=2}I321={i:x1i>x2i,δ1i=1,δ2i=2}I231={i:x1i>x2i,δ1i=1,δ2i=3}.

We have I1=I312I132 and I2=I321I231. In the observed data set I1, δ2i=2, is known, but δ1i is unknown. Similarly, in I2, δ1i=1, is known, but δ2i is unknown. The possible arrangement of the Ui's along with the associated probabilities are provided in Table 1. It will be useful in computing the likelihood function of the complete data set.

Table 1. The possible arrangements of U1,U2 and U3.

Different arrangements Probabilities (δ1,δ2) X1 and X2 Set
U3<U1<U2 λ1λ2(λ1+λ3)(λ1+λ2) (1,2) X1<X2 I312
U1<U3<U2 λ2λ3(λ1+λ3)(λ1+λ2) (3,2) X1<X2 I132
U3<U2<U1 λ1λ2(λ2+λ3)(λ1+λ2) (1,2) X1>X2 I321
U2<U3<U1 λ1λ3(λ2+λ3)(λ1+λ2) (1,3) X1>X2 I231

The likelihood contribution of a data point from the complete data, for each set is given below.

  1. From the set I312 the contribution is αλ1x1i(α+1)e(λ1+λ3)x1iααλ2x2i(α+1)eλ2x2iα,

  2. From the set I132 the contribution is αλ3x1i(α+1)e(λ1+λ3)x1iααλ2x2i(α+1)eλ2x2iα,

  3. From the set I321 the contribution is αλ1λ2x1i(α+1)eλ1x1iααx2i(α+1)e(λ2+λ3)x2iα,

  4. From the set I231 the contribution is αλ1λ3x1i(α+1)eλ1x1iααx2i(α+1)e(λ2+λ3)x2iα.

Based on the complete data set C={((x1i,δ1i),(x2i,δ2i)):i=1,,n} the log-likelihood function can be obtained as follows.

l(θ|C)=iI312(2lnα+lnλ1+lnλ2)(λ1+λ3)iI312x1iαλ2iI312x2iα(α+1)iI312(lnx1i+lnx2i)+iI132(2lnα+lnλ2+lnλ3)(λ1+λ3)iI132x1iαλ2iI132x2iα(α+1)iI132(lnx1i+lnx2i)+iI321(2lnα+lnλ1+lnλ2)λ1iI321x1iα(λ2+λ3)iI321x2iα(α+1)iI321(lnx1i+lnx2i)+iI231(2lnα+lnλ1+lnλ3)λ1iI231x1iα(λ2+λ3)iI231x2iα(α+1)iI231(lnx1i+lnx2i). (11)

It can be easily seen that for a given α, the maximization of (11) can be obtained in explicit forms in terms of λ1, λ2 and λ3, and the maximization with respect to α can be obtained by solving a one dimensional optimization problem with respect to α. This is the main motivation of the proposed EM algorithm. We need the following result for developing the EM algorithm. They can be obtained from Table 1.

RESULT 4 :

P(δ1=1,δ2=2|X1<X2)=λ1(λ1+λ3)=a,P(δ1=3,δ2=2|X1<X2)=λ3(λ1+λ3)=1a,P(δ1=1,δ2=2|X1>X2)=λ2(λ2+λ3)=b,P(δ1=1,δ2=3|X1>X2)=λ3(λ2+λ3)=1b.

Let us use the following notations. At the r-th stage of the EM algorithm, the estimates of α, λ1, λ2 λ3 will be denoted by α(r), λ1(r), λ2(r), λ3(r), respectively. Similarly, λ(r)=λ1(r)+λ2(r)+λ3(r), a(r)=λ1(r)/(λ1(r)+λ3(r)) and b(r)=λ2(r)/(λ2(r)+λ3(r)) are also defined. Now, using the idea of Dinse [6], see also Kundu [10], the ‘pseudo’ log-likelihood function at the r-th stage (‘E’-step) of the EM algorithm can be written as

l(θ|θ(r),Data)=(n1a(r)+n2)lnλ1λ1iI1I2x1iα+(n1+n2b(r))lnλ2λ2iI1I2x2iα+(n1(1a(r))+n2(1b(r)))lnλ3λ3[iI1x1iα+iI2x2iα]+2nlnα(α+1)iI1I2(lnx1i+lnx2i). (12)

The ‘M’-step can be obtained by maximizing l(θ|θ(r),Data) with respect to the unknown parameters. For fixed α, the function (12) is maximized at λ1=λ1(r+1)(α), λ2=λ2(r+1)(α), λ3=λ3(r+1)(α), where,

λ1(r+1)(α)=n1a(r)+n2iI1I2x1iα,λ2(r+1)(α)=n1+n2b(r)iI1I2x2iα,λ3(r+1)(α)=n1(1a(r))+n2(1b(r))iI1x1iα+iI2x2iα.

The corresponding α(r+1) can be obtained by maximizing pseudo profile log-likelihood function l(α,λ1(r+1)(α),λ2(r+1)(α),λ3(r+1)(α)|θ(r),Data)=h(r)(α). Without the additive constant,

h(r)(α)=(n1a(r)+n2)ln(iI1I2x1iα)(n1+n2b(r))ln(iI1I2x2iα)(n1(1a(r))+n2(1b(r)))ln(iI1x1iα+iI2x2iα)+2nlnα(α+1)iI1I2(lnx1i+lnx2i).

The following result ensures the existence of a unique maximum of the pseudo profile log-likelihood function at every stage of the iteration.

RESULT 5 :

At the r-th stage the pseudo log-likelihood function h(r)(α) is a unimodal function of α.

Proof.

See in Appendix.

The maximization of the pseudo log-likelihood function which is a one dimensional optimization problem can be done by the Newton Raphson or bisection method. Once α(r+1) can be obtained, we can compute λ1(r+1)=λ1(r+1)(α(r+1)), λ2(r+1)=λ2(r+1)(α(r+1)) and λ3(r+1)=λ3(r+1)(α(r+1)). The process will be continued till convergence.

5. Bayesian inference

In this section we provide the Bayesian analysis of the unknown model parameters. The Bayes estimators are derived based on the squared error loss function although any other loss function also can be easily incorporated. The prior assumption and its posterior analysis are discussed below.

5.1. Prior assumption

Following the approach of Pena and Gupta [18] the following prior assumptions are made on the scale parameters. It is assumed that λ=λ1+λ2+λ3 follows a gamma distribution, say GA(a0,b0) with the density

fGA(λ)=b0a0γa0λa01eb0λ;λ>0,a0,b0>0.

To incorporate dependence among λ1,λ2,λ3, it is assumed that given λ, (λ1/λ,λ2/λ) follows a Dirichlet prior say, D(a1,a2,a3). Note that, a random vector (X1,x2,,Xk1) is said to follow a Dirichlet distribution with parameter a1,a2,,ak, denoted by (X1,X2,,Xk1)D(a1,a2,,ak), with the density of the form

f(x1,,xk1)=1B(a1,a2,,ak)i=1k1xiai1×(1i=1k1xi)ak1;where,0<xi<1i=1,,k1,andB(a1,a2,,ak)=i=1kΓ(ai)Γ(a)andai>0,i=1,,k,a=i=1kak.

The joint prior of (λ1,λ2,λ3) can be obtained as

π1(λ1,λ2,λ3)(λ1+λ2+λ3)a0a1a2a3λ1a11λ2a21λ3a31eb0(λ1+λ2+λ3). (13)

The joint prior with hyper parameters a,b,a1,a2,a3 is called the Gamma-Dirichlet prior, denoted by GD(a0,b0,a1,a2,a3) and it is a very flexible prior, see Pena and Gupta [18] for details. It is further assumed that the shape parameter α has a log-concave prior π2(α) with a positive support on (0,) and π2(α) is independent of π1(λ1,λ2,λ3). Hence the joint prior of α,λ1,λ2,λ3 is obtained as π(α,λ1,λ2,λ3)=π2(α)π1(λ1,λ2,λ3).

5.2. Posterior analysis

In this section we provide the Bayes estimators based on the squared error loss function and the associated credible intervals CRI of the unknown parameters. The joint posterior density of θ=(α,λ1,λ2,λ3) can be obtained as

π(θ|Data){λ1+λ2+λ3λ1+λ2}nλ1n2+a11λ2n1+a21(λ1+λ2)n1(λ2+λ3)n3×e(λ1+λ3)iI1x1iαeλ2iI1x2iαeλ1iI2x1iαe(λ2+λ3)iI2x2iα×(λ1+λ2+λ3)a0a1a2a3eb0(λ1+λ2+λ3)×α2nπ2(α)×{iI1I2x1i(α+1)x2i(α+1)}.

The Bayes estimate of any function of θ say g(θ) can be obtained as

E(g(θ)|Data)=0000g(θ)π(θ|Data)dαdλ1dλ2dλ3,

provided it exists. The estimator cannot be obtained in closed form. For further development simplify the posterior density as follows.

π(θ|Data)h(λ1,λ2,λ3)π11(λ1|α,data)π12(λ2|α,data)π13(λ3|α,Data)π2(α|data)

where,

π11(λ1|α,Data)GA(n2+a1,b0+iI1I2x1iα),π12(λ2|α,Data)GA(n1+a2,b0+iI1I2x2i,α),π13(λ3|α,Data)GA(n+a3,b0+iI1x1iα+iI2x2iα),π2(α|Data){iI1I2x1i(α+1)x2i(α+1)}α2nπ2(α)(b0+iI1I2x1iα)n2+a1(b0+iI1I2x2iα)n1+a2(b0+iI1x1iα+iI2x2iα)n+a3,h(λ1,λ2,λ3)=(λ1+λ2+λ3)n+a0a1a2a3(1+λ1λ3)n1(1+λ2λ3)n2(λ1+λ2)n.

As the posterior cannot be obtained in any standard form we rely on the importance sampling technique to compute the Bayes estimates and the associated credible intervals. The following result is needed for further development.

RESULT 6 :

When π2(α) is log-concave, π2(α|data) is also log-concave.

Proof.

It can be obtained similarly as in Kundu [11].

To apply the importance sampling technique, it is required to generate α,λ1,λ2,λ2 from the posterior density. Based on the method by Devroye [5] α can be generated from the log-concave density function π2(α|data). Here we follow the simpler method suggested in Kundu [11] to generate α from its log-concave posterior density. Once α is generated we can generate λ1,λ2,λ3 from the respective conditional densities. The following algorithm can be used to compute the Bayes estimate and the associated credible intervals.

Algorithm 2

  1. Given data, generate α from π2(α|data).

  2. For a given α, generate λ1,λ2,λ3 from π11(λ1|data), π12(λ2|data) and π13(λ3|data), respectively.

  3. Repeat the process say N times to generate ((α1,λ11,λ21,λ31),,(αN,λ1N,λ2N,λ3N)).

  4. To compute the Bayes estimate of g(α,λ1,λ2,λ3) compute (g1,,gN) and (h1,,hN) where gi=g(αi,λ1i,λ2i,λ3i) and hi=h(λ1i,λ2i,λ3i) for i=1,,N.

  5. The Bayes estimate of g(α,λ1,λ2,λ3) based on the squared error loss function can be computed as i=1Nhigij=1Nhj=i=1Nwigi where wi=hij=1Nhj.

  6. To compute a 100(1γ)% credible interval of g(α,λ1,λ2,λ3), arrange gi's in an ascending order to obtain (g(1),g(N)) and record the corresponding wi as (w[1],,w[N]). Here w[i] is not ordered and they are assigned to the corresponding ordered g(i)'s. A 100(1γ)% credible interval can be obtained as (g(j1),g(j2)) where j1,j2 such that
    j1<j2,j1,j2{1,,N}andi=j1j2w[i]1γ<i=j1j2+1w[i]. (14)

6. Simulation results and data analysis

6.1. Simulation results

In this section we present the results based on the simulation experiments. The experiment is performed to check how the different methods perform for different sample sizes and for different parameter values. We consider n=20, 30, 40 for three different sets of parameter values with α=0.5,1,2 and λ1=1, λ2=1, λ3=1. We compute the average estimates (AE) and the corresponding mean squared errors (MSE) of the MLEs derived through EM algorithm as described in Section 4. In Tables 24 we provide the AEs and MSEs based on 1000 samples. In the EM algorithm we set the initial guesses as the true parameter values and stop the process until the absolute differences of the estimates from two consecutive iteration is less than 104. It may be mentioned that in our simulation experiments we have kept λ's to be constant and we have changed α. In fact we have performed some simulations with different values of λ's also, but we have obtained similar results, hence they have not been reported. The effect of change of α is more evident in the simulation experiments. All the computations are performed using R software, and they are available on request from the authors.

Table 2. AE and MSE of the MLEs and Bayes estimators for different values of n with α=0.5,λ1=1,λ2=1,λ3=1.

    MLE Bayes (IP) Bayes (NIP)
n Parameter AE MSE AE MSE AE MSE
n=20 α 0.530 0.007 0.516 0.005 0.517 0.005
  λ1 1.162 0.418 0.956 0.095 0.914 0.115
  λ2 1.142 0.378 0.956 0.087 0.928 0.113
  λ3 0.948 0.445 1.165 0.107 1.246 0.163
n=30 α 0.520 0.004 0.507 0.003 0.507 0.003
  λ1 1.096 0.267 0.928 0.058 0.889 0.072
  λ2 1.107 0.288 0.939 0.060 0.900 0.075
  λ3 0.993 0.370 1.181 0.084 1.219 0.117
n=40 α 0.513 0.002 0.503 0.002 0.502 0.002
  λ1 1.065 0.174 0894 0.056 0.896 0.058
  λ2 1.069 0.186 0.846 0.060 0.889 0.063
  λ3 0.978 0.268 1.213 0.091 1.213 0.099
n=80 α 0.508 0.001 0.498 0.001 0.499 0.001
  λ1 1.044 0.098 0.888 0.032 0.874 0.036
  λ2 1.027 0.093 0.895 0.031 0.877 0.036
  λ3 0.985 0.160 1.186 0.055 1.196 0.063

Table 4. AE and MSE of the MLEs and Bayes estimators for different values of n with α=2,λ1=1,λ2=1,λ3=1.

    MLE Bayes (IP) Bayes (NIP)
n Parameter AE MSE AE MSE AE MSE
n=20 α 2.940 0.097 2.049 0.073 2.053 0.086
  λ1 1.097 0.502 0.964 0.095 0.893 0.112
  λ2 1.118 0.527 0.954 0.094 0.890 0.116
  λ3 1.027 0.705 1.157 0.102 1.240 0.161
n=30 α 2.076 0.064 2.029 0.047 2.026 0.052
  λ1 1.061 0.382 0.940 0.063 0.891 0.069
  λ2 1.065 0.383 0.931 0.062 0.894 0.075
  λ3 1.645 0.594 1.181 0.086 1.221 0.115
n=40 α 2.062 0.044 2.014 0.032 2.022 0.041
  λ1 1.080 0.295 0.879 0.053 0.887 0.056
  λ2 1.084 0.302 0.849 0.058 0.877 0.061
  λ3 0.991 0.465 1.226 0.101 1.233 0.100
n=80 α 2.032 0.023 2.001 0.017 1.998 0.017
  λ1 1.025 0.161 0.899 0.032 0.877 0.035
  λ2 1.035 0.166 0.897 0.031 0.871 0.036
  λ3 0.992 0.277 1.180 0.054 1.195 0.060

Based on the observed information matrix we compute the asymptotic confidence intervals and record the average lengths (AL) and coverage percentages (CP) based on 1000 replications in Tables 510. In the Bayesian analysis, for the shape parameter α we choose a gamma prior which is a log-concave, i.e. π2(α) GA(c,d). The Bayes estimates based on the squared error loss function is computed both for an informative prior (IP) and non-informative prior (NIP). In the informative prior we choose the values of the hyper parameters equating the prior mean with the true parameter values. When α=0.5,λ1=1,λ2=1,λ3=1 the hyper parameters are: a0=1,b0=13,a1=1,a2=1,a3=1,c=1,d=0.5. When α=1,λ1=1,λ2=1,λ3=1 the hyper parameters are: a0=1,b0=13,a1=1,a2=1,a3=1,c=1,d=1 and for α=2,λ1=1,λ2=1,λ3=1, the hyper-parameters are: a0=1,b0=13,a1=1,a2=1,a3=1,c=2,d=1. In the non-informative prior following the idea from Congdon [3] in every set of parameters, the hyper-parameters are as a=105,b=105,a1=105,a2=105,a3=105,c=105,d=105. We have considered both the informative and non-informative priors to see the effect of priors on the performance of the Bayes estimators.

Table 6. AL and CP of 95% Asymptotic CI and 95% Symmetric CRI for different values of n with α=0.5,λ1=1,λ2=1,λ3=1.

    95% Asymptotic CI 95% Symmetric CRI (IP) 50% Symmetric CRI (NIP)
n Parameter AL CP AL CP AL CP
n=20 α 0.312 95.7% 0.245 91.2% 0.242 91.3%
  λ1 3.129 98.1% 1.275 95.1% 1.285 93.5%
  λ2 3.186 98.7% 1.257 95.6% 1.280 93.1%
  λ3 3.530 99.8% 1.382 98.6% 1.446 98.1%
n=30 α 0.246 96.1% 0.98 93.1% 0.199 91.8%
  λ1 2.645 98.9% 1.019 94.9% 1.038 93.2%
  λ2 2.650 98.7% 1.030 94.5% 1.036 93.3%
  λ3 2.993 99.2% 1.139 98.9% 1.172 98.4%
n=40 α 0.211 96.4% 0.169 91.4% 0.169 91.0%
  λ1 2.376 99.3% 0.877 94.2% 0.887 91.2%
  λ2 2.356 99.5% 0.894 93.5% 0.889 91.2%
  λ3 2.766 99.7% 0.987 97.4% 1.007 96.9%
n=80 α 0.146 95.0% 0.117 91.2% 0.118 91.1%
  λ1 1.812 99.3% 0.619 89.3% 0.626 88.2%
  λ2 1.811 98.9% 0.628 91.8% 0.623 87.8%
  λ3 2.187 98.9% 0.708 92.7% 0.713 91.9%

Table 7. AL and CP of 90% Asymptotic CI and 90% Symmetric CRI for different values of n with α=1,λ1=1,λ2=1,λ3=1.

    90% Asymptotic CI 90% Symmetric CRI (IP) 90% Symmetric CRI (NIP)
n Parameter AL CP AL CP AL CP
n=20 α 0.518 91.8% 0.412 86.4% 0.413 84.4%
  λ1 2.709 97.3% 1.092 92.9 % 1.101 88.8%
  λ2 2.701 96.1% 1.114 92.1% 1.121 88.6%
  λ3 3.019 99.7% 1.226 97.0% 1.287 96.3%
n=30 α 0.415 91.3% 0.329 86.9% 0.330 84.6%
  λ1 2.339 97.5% 0.886 90.2% 0.911 88.7%
  λ2 2.360 98.2% 0.898 91.2% 0.908 88.5%
  λ3 2.669 98.2% 1.024 86.6% 1.063 95.%
n=40 α 0.354 91.7% 0.280 85.9% 0.283 85.7%
  λ1 2.113 98.0% 0.772 88.0% 0.777 87.4%
  λ2 2.081 98.2% 0.760 85.3% 0.772 85.7%
  λ3 2.382 98.7% 0.911 93.4% 0.912 92.4%
n=80 α 0.246 91.1% 0.198 86.4% 0.196 84.4%
  λ1 1.561 97.8% 0.542 83.4% 0.549 81.4%
  λ2 1.559 98.1% 0.542 81.4% 0.548 80.7%
  λ3 1.885 99.6% 0.634 86.6% 0.647 86.0%

Table 8. AL and CP of 95% Asymptotic CI and 95% Symmetric CRI for different values of n with α=1,λ1=1,λ2=1,λ3=1.

    95% Asymptotic CI 95% Symmetric CRI (IP) 95% Symmetric CRI (NIP)
n Parameter AL CP AL CP AL CP
n=20 α 0.618 96.0% 0.490 92.2% 0.498 90.2%
  λ1 3.050 98.2% 1.276 97.4 % 1.306 94.4%
  λ2 3.079 98.7% 1.286 96.3 % 1.291 94.1%
  λ3 3.541 99.7% 1.395 99.3% 1.452 98.2%
n=30 α 0.496 96.2% 0.395 91.6% 0.395 92.8%
  λ1 2.702 99.2% 1.036 96.0% 1.016 94.1%
  λ2 2.641 98.9% 1.034 96.2% 1.027 91.3%
  λ3 3.007 99.1% 1.145 98.4% 1.160 97.4%
n=40 α 0.422 95.1% 0.338 93.3% 0.336 91.4%
  λ1 2.342 98.2% 0.889 93.4% 0.887 91.6%
  λ2 2.336 98.6% 0.886 93.5% 0.889 93.6%
  λ3 2.764 99.2% 0.991 98.1% 1.014 96.8%
n=80 α 0.294 96.2% 0.235 90.3% 0.234 92.1%
  λ1 1.812 99.3% 0.620 88.5% 0.620 87.4%
  λ2 1.811 98.9% 0.628 91.6% 0.617 87.0%
  λ3 2.151 99.2% 0.710 91.3% 0.721 90.2%

Table 9. AL and CP of 90% Asymptotic CI and 90% Symmetric CRI for different values of n with α=2,λ1=1,λ2=1,λ3=1.

    90% Asymptotic CI 90% Symmetric CRI (IP) 90% Symmetric CRI (NIP)
    AL CP AL CP AL CP
n=20 α 1.043 90.3% 0.815 85.7% 0.816 85.1%
  λ1 2.747 96.7% 1.097 91.6% 1.113 88.9%
  λ2 2.720 98.3% 1.104 90.6% 1.131 89.3%
  λ3 3.122 98.8% 1.224 96.0% 1.298 95.4%
n=30 α 0.836 90.3% 0.660 86.0% 0.662 84.9%
  λ1 2.314 97.2% 0.898 91.7% 0.918 89.5%
  λ2 2.321 97.8% 0.893 89.8% 0.913 87.8%
  λ3 2.636 98.6% 1.018 95.0% 1.060 94.3%
n=40 α 0.711 91.8% 0.566 87.9% 0.570 83.1%
  λ1 2.075 97.8% 0.767 88.3% 0.787 87.7%
  λ2 2.072 97.6% 0.752 87.8% 0.782 84.8%
  λ3 2.391 99.2% 0.911 92.7% 0.916 92.9%
n=80 α 0.492 91.0% 0.394 86.9% 0.396 84.8%
  λ1 1.546 96.4% 0.549 85.5% 0.545 81.6%
  λ2 1.539 96.1% 0.551 84.5% 0.542 81.7%
  λ3 1.854 98.2% 0.643 87.2% 0.645 86.8%

Table 5. AL and CP of 90% Asymptotic CI and 90% Symmetric CRI for different values of n with α=0.5,λ1=1,λ2=1,λ3=1.

    90% Asymptotic CI 90% Symmetric CRI (IP) 90% Symmetric CRI (NIP)
n Parameter AL CP AL CP AL CP
n=20 α 0.262 91.3% 0.202 85.2% 0.207 85.3%
  λ1 2.824 98.8% 1.102 92.2% 1.132 89.2%
  λ2 2.855 97.8% 1.109 92.0% 1.116 89.4%
  λ3 3.047 99.8% 1.232 96.9% 1.288 96.2%
n=30 α 0.209 91.0% 0.165 86.2% 0.168 84.2%
  λ1 0.428 98.9% 0.905 90.5% 0.909 89.9%
  λ2 2.420 98.5% 0.886 90.7% 0.903 89.4%
  λ3 2.642 99.0% 1.019 96.6% 1.049 94.5%
n=40 α 0.176 91.9% 0.141 85.6% 0.142 85.5%
  λ1 2.088 98.3% 0.780 88.3% 0.782 86.8%
  λ2 2.089 98.4% 0.755 85.2% 0.768 85.7%
  λ3 2.400 99.8% 0.917 93.0% 0.909 93.7%
n=80 α 0.122 90.4% 0.098 84.4% 0.098 83.2%
  λ1 1.555 98.8% 0.542 84.7% 0.548 81.8%
  λ2 1.556 98.6% 0.543 84.4% 0.550 83.3%
  λ3 1.925 99.8% 0.633 86.7% 0.651 86.4%

Table 10. AL and CP of 95% Asymptotic CI and 95% Symmetric CRI for different values of n with α=2,λ1=1,λ2=1,λ3=1.

    95% Asymptotic CI 95% Symmetric CRI (IP) 95% Symmetric CRI (NIP)
n Parameter AL CP AL CP AL CP
n=20 α 1.246 96.3% 0.976 92.1% 0.981 89.9%
  λ1 3.098 98.3% 1.269 95.8% 1.275 93.8%
  λ2 3.106 97.8% 1.251 94.9% 1.303 94.1%
  λ3 3.514 99.9% 1.371 99.1% 1.438 98.8%
n=30 α 0.998 96.4% 0.784 91.7% 0.785 90.4%
  λ1 2.648 98.2% 1.028 95.0 % 1.027 92.0%
  λ2 2.645 98.1% 1.019 94.0% 1.037 94.1%
  λ3 3.069 98.6% 1.129 97.7% 1.169 98.0%
n=40 α 0.841 96.3% 0.671 91.9% 0.669 90.3%
  λ1 2.349 98.6% 0.898 94.5% 0.881 91.3%
  λ2 2.357 98.8% 0.886 94.3% 0.903 92.1%
  λ3 2.681 99.2% 0.999 97.8% 1.009 96.2%
n=80 α 0.588 94.7% 0.467 90.7% 0.465 91.9%
  λ1 1.789 98.4% 0.622 89.4% 0.618 88.2%
  λ2 1.794 98.6% 0.614 89.4% 0.623 88.4%
  λ3 2.144 99.9% 0.703 92.3% 0.713 91.7%

Both for the informative and the non-informative priors we compute the AEs and the corresponding MSEs of the Bayes estimators based on 1000 replications are provided in Tables 2, 3 and 4 for different sample sizes and for different sets of parameters. In interval estimation we compute both 90% and 95% symmetric credible intervals and report the ALs and CPs based on 1000 replications in Tables 510.

Table 3. AE and MSE of the MLEs and Bayes estimators for different values of n with α=1.0,λ1=1,λ2=1,λ3=1.

    MLE Bayes (IP) Bayes (NIP)
n Parameter AE MSE AE MSE AE MSE
n=20 α 1.060 0.024 1.030 0.021 1.030 0.021
  λ1 1.162 0.496 0.968 0.098 0.925 0.124
  λ2 1.159 0.481 0.946 0.082 0.905 0.111
  λ3 0.957 0.616 1.168 0.096 1.237 0.156
n=30 α 1.039 0.015 1.016 0.012 1.023 0.013
  λ1 1.095 0.333 0.929 0.064 0.894 0.077
  λ2 1.089 0.313 0.929 0.055 0.888 0.080
  λ3 0.998 0.503 1.166 0.076 1.239 0.126
n=40 α 1.026 0.010 1.009 0.008 0.999 0.009
  λ1 1.102 0.246 0.879 0.055 0.890 0.057
  λ2 1.076 0.225 0.856 0.058 0.877 0.064
  λ3 0.953 0.356 1.218 0.095 1.241 0.105
n=80 α 1.014 0.006 0.997 0.004 1.000 0.004
  λ1 1.043 0.138 0.890 0.031 0.878 0.035
  λ2 1.043 0.133 0.886 0.032 0.877 0.036
  λ3 0.968 0.224 1.172 0.051 1.206 0.065

From Tables 24, it is clear that as sample size increases the MLEs perform better in terms of average bias and the mean squared error. This result indicates the consistency of the estimators. Also the Bayes estimator of the shape parameter α is consistent estimator. Again we see that as n increases, the Bayes estimates for the scale parameters become more distant from the true values, though the MSEs decrease. This result indicates the Bayes estimators of the scale parameters converge to some values but not the true parameter values. It is also observed that the Bayes estimators based on the informative prior have lower MSEs than the other two estimators. The MLEs over estimate λ1,λ2 and under estimate λ3, slightly, where as the Bayes estimators under estimate λ1,λ2 and over estimate λ3.

In interval estimation for small to moderate sample sizes the symmetric credible intervals perform better than the asymptotic intervals in terms of average length and CP. As sample size increases, the CPs of the symmetric credible intervals for the scale parameters go lesser than the nominal level. The asymptotic intervals for the scale parameters perform better than the symmetric credible intervals in terms of CP for large sample.

6.2. Data analysis

In this section we analyze a real data set for illustrative purposes. The data set has been taken from Johnson et al. [9]. This is a bivariate data set (X1,X2) on 24 children where X1 represents the bone mineral density in g/cm2 for Dominant Ulna and X2 represents the bone mineral density in g/cm2 for Ulna bones. The data are as follows.

(0.869, 0.964), (0.602, 0.689), (0.765, 0.738), (0.761, 0.698), (0.551, 0.619), (0.753, 0.515), (0.708,0.787), (0.687 0.715), (0.844 0.656), (0.869 0.789), (0.654, 0.726), (0.692, 0.526), (0.670, 0.580), (0.823,0.773), (0.746, 0.729), (0.656, 0.506), (0.693, 0.740), (0.883, 0.785), (0.577, 0.627), (0.802, 0.769), (0.540, 0.498), (0.804, 0.779), (0.570, 0.634), (0.585, 0.640).

To check whether a IW distribution can fit the maximum of X1 and X2, we perform Kolmogorov-Smirnov (K-S) goodness of fit test. The MLEs and the K-S distance between the empirical distribution and the fitted distribution along with the p value are recorded in Table 11. Based on the results in Theorem 3.3 we have fitted also ACBIW distributions both on X1|X1>X2 and X2|X2>X1 and perform the K-S test for both the cases. The results are also provided in Table 11. Based on the results, it is reasonable to assume that the ACBIW may be used to analyze this data set.

Table 11. Goodness of fitting of real data set.

  MLE    
Case shape parameter scale parameter KS distance p value
max(X1,X2) 7.563 10.961 0.110 0.933
X1|X1>X2 6.602 9.355 0.227 0.461
X2|X2>X1 9.999 19.067 0.201 0.813

We consider the classical and the Bayes estimations of the parameters α, λ1, λ2, λ3. In the EM algorithm as the initial guesses of the parameters we consider the MLEs for max(X1,X2). Here, in Table 11, the MLE of shape parameter is 7.563 and the MLE of the scale parameter is 10.961. Therefore, we set α(0)=7 and λ1(0)=3,λ2(0)=3,λ3(0)=3. We start the EM algorithm with these initial guesses and continue until the absolute differences of the estimates in two consecutive iterations is less than 104.

The MLEs and the percentile bootstrap confidence intervals are recorded in Table 12. The Bayes estimates are derived for the non-informative prior based on the squared error loss function. We compute the 90% symmetric credible intervals for the unknown model parameters. All these results are provided in Table 12.

Table 12. Maximum likelihood estimate and Bayesian estimates for Real Data set.

Parameter MLE 90% Asymptotic CI Bayes estimate 90% symmetric CRI
α 5.133 (3.331, 6.381) 4.950 (4.745, 5.140)
λ1 2.142 (0.465, 3.819) 2.147 (1.209,, 3.629)
λ2 1.434 (0.587, 2.281) 1.409 (0.606, 2.924)
λ3 3.097 (1.315, 4.879) 2.971 (1.485,, 4.505)

Here we have performed the data analysis considering (X1,X2) follows a ACBIW distribution. Now we perform K-S tests to check whether the marginal distributions of the ACBIW can fit X1 and X2. The p values are 0.647 and 0.449 for X1 and X2, respectively. Therefore the assumption of ACBIW on the Data set is quite reasonable.

7. Application

In this section we apply the ACBIW distribution as a dependent complementary risks model. It is assumed that n experimental units are put on a life testing experiment and each unit is susceptible to two risk factors with latent failure times X1 and X2. It is further assumed that (X1,X2) ACBIW(α,λ1,λ2,λ3). For each experimental unit we observe T=max(X1,X2). We define a random variable δ with δi=1 if i-th failure occurs due to cause 1 and δi=2 if it is due to cause 2. Therefore we observe the data D={(t1,δ1),,(tn,δn)}. The problem is to estimate the unknown parameters θ=(α,λ1,λ2,λ3) based on the observed data.

The likelihood function of the observation can be written as

L(θ|D)=(λ1+λ2+λ3λ1+λ2)nλ1d1λ2d2αni=1nti(α+1)e(λ1+λ2+λ3)i=1ntiα. (15)

Here d1 = # {i:δi=1} and d2 = # {i:δi=2}. Now to compute the MLEs of the unknown parameters, we can modify the EM algorithm described in Section 4, very easily. Using the notation in Section 4, note that if δi=1, then clearly x1i=ti and x2i<x1i, and it is unknown. Moreover, tiI2. Similarly, if δi=2, then clearly x2i=ti and x1i<x2i, and it is unknown. Here tiI1. Therefore, the EM algorithm as described in Section 4, can be modified as follows. In the set I1, x1i's are replaced by their expectations, similarly, in the set I2, the missing x2i's are replaced by their expectations. These expectations cannot obtained in explicit forms, but they can be obtained in the integration form using Theorem 3.2 as follows:

E(X1|X1<X2=t)=e(λ1+λ3)tα(λ1+λ3)1/α(λ1+λ3)tαu1/αeudu.E(X2|X2<X1=t)=e(λ2+λ3)tα(λ2+λ3)1/α(λ2+λ3)tαu1/αeudu.

In the Bayesian analysis part, we use the same prior as in Section 6. Based on the likelihood in (15), the joint posterior density function can be written as

π(θ|D)h(λ1,λ2,λ3)π11(λ1|α,D)π12(λ2|α,D)π13(λ3|α,D)π2(α|D) (16)

where

π11(λ1|α,D)GA(d1+a1,b+i=1ntiα),π12(λ2|α,D)GA(d2+a2,b+i=1ntiα),π13(λ3|α,D)GA(a3+n,b+i=1ntiα),π2(α|D)αni=1nti(α+1)π2(α)(b+i=1ntiα)2n+a1+a2+a3,h(λ1,λ2,λ3)=(λ1+λ2+λ3)n+aa1a2a3(λ1λ3+λ2λ3)n.

To derive the Bayes estimate of any function g(α,λ1,λ2,λ3), based on the squared error loss function, provided it exists, as well as associated credible intervals we follow the Algorithm 2 from Section 6.

A real data set under complementary risk is analyzed for illustrative purposes. The data set are originally taken from Han [8]. Here we see a small unmanned aerial vehicle (SUAV) is equipped with dual propulsion systems and the SUAV drops from its flying altitude only when both the propulsion systems fail. Table 13 indicates the failure time of the SUAV as well as which propulsion system fails later.

Table 13. real Data in presence of complementary risks.

Failure Time Cause of Failure
2.365 1
3.467 2
5.386 2
7.714 2
9.578 1
9.683 2
11.416 1
11.789 1
12.039 2
14.928 1
14.938 2
15.325 2
15.781 2
16.105 1
16.362 2
17.178 2
17.366 1
17.803 1
19.578 2

We ignore the censored data as given in Han [8] and conduct the analysis based on the first n=19 units. To check whether a IW distribution can fit the failure time data (without the cause of failure) we compute the K-S distance between the empirical and the fitted distributions. The MLEs of the shape and scale parameters become 1.427 and 1.991, respectively. The K-S distance and the associated p value become 0.247 and 0.196, respectively. Therefore, based on the p value we say IW distribution can be used to analyze the SUAV failure data. We have calculated the MLEs and the Bayes estimates along with the asymptotic confidence and symmetric credible intervals. The results are presented in Table 14.

Table 14. Maximum likelihood estimate and Bayes estimate for SUAV Data in presence of complementary risks.

Parameter MLE 90% Asymptotic CI Bayes estimate 90% symmetric CRI
α 1.363 (1.119, 1.607) 1.354 (1.320, 1.546)
λ1 0.488 (0.325,0.651) 0.480 (0.315, 0.628)
λ2 0.907 (0.379, 1.435) 0.947 (0.371, 1.065)
λ3 0.755 (0.300, 1.210) 0.759 ( 0.563, 1.470)

8. Conclusion

In this paper we have introduced a new absolutely continuous bivariate distribution by omitting the singular part of the BIW distribution introduced by Muhammed [17] and Kundu and Gupta [14]. The proposed ACBIW distribution can be used quite effectively if the marginals have heavy tailed distribution and there are no ties in the data set. We have studied different properties of the model. The MLEs of the unknown parameters cannot be obtained in explicit forms, we have provided a very effective EM algorithm to estimate unknown parameters. It is observed based on extensive simulation studies that the proposed EM algorithm works quite well in practice. We further considered the Bayesian inference of the unknown parameters based on a fairly general set of priors. We have provided a very effective method of generating samples from the joint posterior PDFs, and it can be used to compute Bayes estimates and the associated credible intervals. The model has been used as dependent complementary risks model. In this paper we have mainly concentrated on the bivariate set up, but it can be extended to the multivariate case also. Moreover, in most of the times in reliability and survival analysis, the data are censored. Here, we have considered complete sample only. It will be of interest to study the analysis of this model under different censoring schemes. More work is needed along that direction.

Acknowledgments

The authors would like to thank the reviewers and the associate editor for their constructive comments.

Appendix.

Proof of Theorem 3.3 —

(i)

P(X1<X2)=00x2fACBIW(x1,x2)dx1x2=00x2(λ1+λ2+λ3)(λ1+λ2)fIW(x1;α,λ1+λ3)fIW(x2;α,λ2)dx1dx2=(λ1+λ2+λ3)(λ1+λ2)0fIW(x2;α,λ2)e(λ1+λ3)x2αdx2=λ2λ1+λ2.

(ii)

P(max(X1,X2)<x)=P(X1<X2<x)+P(X2<X1<x)=(λ1+λ2+λ3)(λ1+λ2)0x0x2fIW(x1;α,λ1+λ3)fIW(x2;α,λ2)dx1dx2+(λ1+λ2+λ3)(λ1+λ2)0x0x1fIW(x1;α,λ1)fIW(x2;α,λ2+λ3)dx2dx1=(λ1+λ2+λ3)(λ1+λ2)(0xfIW(x2;α,λ2)(e(λ1+λ3)xαdx2+0xfIW(x1;α,λ1)e(λ2+λ3)xαdx1)=(λ1+λ2+λ3)(λ1+λ2)(λ2(λ1+λ2+λ3)e(λ1+λ2+λ3)xα+λ1(λ1+λ2+λ3)e(λ1+λ2+λ3)xα)=e(λ1+λ2+λ3)xα

(iii) and (iv) can be proved in similar ways.

Proof of Result 5 —

First we will prove that if yi>0 for i=1,,n, then g(α)=ln(i=1nyiα) is a concave function. It mainly follows from the fact that d2/dα2g(α)<0 due to Cauchy-Schwartz inequality. It implies that h(r)(α) is a concave function. Now the result follows as h(r)(α) as α0 or α

Funding Statement

Part of the work of the second author has been partially funded by the grant from Science and Engineering Research Board (SERB) MTR/2018/000179.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  • 1.Basu A.P. and Ghosh J.K., Identifiability of distributions under competing risks and complementary risks model, Communications in Statistics -- Theory and Methods 9 (1980), pp. 1515–1525. doi: 10.1080/03610928008827978 [DOI] [Google Scholar]
  • 2.Block H. and Basu A.P., A continuous bivariate exponential extension, J. Am. Stat. Assoc. 69 (1974), pp. 1031–1037. [Google Scholar]
  • 3.Congdon P., Applied Bayesian Modeling, John Wiley & Sons, New York, 2014. [Google Scholar]
  • 4.Crowder M.J., Classical Competing Risks, Chapman & Hall, Boca Raton, Florida, 2011. [Google Scholar]
  • 5.Devroye L., A simple algorithm for generating random variates with a log-concave density, Computing 33 (1984), pp. 247–257. doi: 10.1007/BF02242271 [DOI] [Google Scholar]
  • 6.Dinse G.E., Non-parametric estimation of partially incomplete time and types of failure data, Biometrics 38 (1982), pp. 417–431. doi: 10.2307/2530455 [DOI] [PubMed] [Google Scholar]
  • 7.Feizjavadian S.H. and Hashemi R., Analysis of dependent competing risks in the presence of progressive hybrid censoring using Marshall-Olkin bivariate Weibull distribution, Comput. Stat. Data. Anal. 82 (2015), pp. 19–34. doi: 10.1016/j.csda.2014.08.002 [DOI] [Google Scholar]
  • 8.Han D., Estimation in step-stress life tests with complementary risks from the exponentiated exponential distribution under time constraint and its applications to UAV data, Stat. Methodol. 23 (2015), pp. 103–122. doi: 10.1016/j.stamet.2014.09.001 [DOI] [Google Scholar]
  • 9.Johnson R.A. and Wichern D.W., Applied Multivariate Analysis, 4th ed, Prentice Hall, New Jersey, 1999. [Google Scholar]
  • 10.Kundu D., Parameter estimation of the partially complete time and type of failure data, Biometrical Journal 46 (2004), pp. 165–179. doi: 10.1002/bimj.200210014 [DOI] [Google Scholar]
  • 11.Kundu D., Bayesian inference and life testing plan for Weibull distribution in presence of progressive censoring, Technometrics 50 (2008), pp. 144–154. doi: 10.1198/004017008000000217 [DOI] [Google Scholar]
  • 12.Kundu D. and Gupta R.D., A class of absolutely continuous bivariate distributions, Stat. Methodol. 7 (2010), pp. 464–477. doi: 10.1016/j.stamet.2010.01.004 [DOI] [Google Scholar]
  • 13.Kundu D. and Gupta A.K., Bayes estimation for the Marshall-Olkin bivariate Weibull distribution, Computational Statistics and Data Analysis 57 (2013), pp. 271–281. doi: 10.1016/j.csda.2012.06.002 [DOI] [Google Scholar]
  • 14.Kundu D. and Gupta A.K., On bivariate inverse Weibull distribution, Brazilian Journal of Probability and Statistics 31 (2017), pp. 275–302. doi: 10.1214/16-BJPS313 [DOI] [Google Scholar]
  • 15.Louzada F., Cancho V.G., Roman M. and Leite J.G., A new long-term lifetime distribution induced by a latent complementary risk framework, J. Appl. Stat. 39 (2012), pp. 2209–2222. doi: 10.1080/02664763.2012.706264 [DOI] [Google Scholar]
  • 16.Marshall A.W. and Olkin I., A multivariate exponential distribution, J. Am. Stat. Assoc. 62 (1967), pp. 30–44. doi: 10.1080/01621459.1967.10482885 [DOI] [Google Scholar]
  • 17.Muhammed H.Z., Bivariate inverse Weibull distribution, J. Stat. Comput. Simul. 86 (2016), pp. 2335–2345. doi: 10.1080/00949655.2015.1110585 [DOI] [Google Scholar]
  • 18.Pena E.A. and Gupta A.K., Bayes estimation for the Marshall-Olkin exponential distribution, Journal of the Royal Statistical Society, Ser. B 52 (1990), pp. 379–389. [Google Scholar]
  • 19.Pradhan B. and Kundu D., Bayes estimation for the block and basu bivariate and multivariate Weibull distributions, J. Stat. Comput. Simul. 86 (2016), pp. 170–182. doi: 10.1080/00949655.2014.1001759 [DOI] [Google Scholar]
  • 20.Shen Y. and Xu A., On the dependent competing risks using Marshall-Olkin bivariate Weibull model: parameter estimation with different methods, Communications in Statistics -Theory and Methods 47 (2018), pp. 5558–5572. doi: 10.1080/03610926.2017.1397170 [DOI] [Google Scholar]

Articles from Journal of Applied Statistics are provided here courtesy of Taylor & Francis

RESOURCES