Skip to main content
Computational and Mathematical Methods in Medicine logoLink to Computational and Mathematical Methods in Medicine
. 2012 May 29;2012:639124. doi: 10.1155/2012/639124

Let Continuous Outcome Variables Remain Continuous

Enayatollah Bakhshi 1,*, Brian McArdle 2, Kazem Mohammad 3, Behjat Seifi 4, Akbar Biglarian 1
PMCID: PMC3368309  PMID: 22693539

Abstract

The complementary log-log is an alternative to logistic model. In many areas of research, the outcome data are continuous. We aim to provide a procedure that allows the researcher to estimate the coefficients of the complementary log-log model without dichotomizing and without loss of information. We show that the sample size required for a specific power of the proposed approach is substantially smaller than the dichotomizing method. We find that estimators derived from proposed method are consistently more efficient than dichotomizing method. To illustrate the use of proposed method, we employ the data arising from the NHSI.

1. Introduction

Recently, logistic regression has become a popular tool in biomedical studies. The parameter in logistic regression has the interpretation of log odds ratio, which is easy for people such as physicians to understand. Probit and complementary log-log are alternatives to logistic model. For a covariate X and a binary response variable Y, let π(X) = P(Y = 1 | X = x). A related model to the complementary log-log link is the log-log link. For it, π(x) approaches 0 sharply but approaches 1 slowly. When the complementary log-log model holds for the probability of a success, the log-log model holds for the probability of a failure [1].

These models use a categorical (dichotomous or polytomous) outcome variable. In many areas of research, the outcome data are continuous. Many researchers have no hesitation in dichotomizing a continuous variable, but this practice does not make use of within-category information. Several investigators have noted the disadvantages of dichotomizing both independent and outcome variables [210]. Ragland [11] showed that the magnitude of odds ratio and statistical power depend on the cutpoint used to dichotomize the response variable. From a clinical point of view, binary outcomes may be preferred for some reasons such as (1) setting diagnostic criteria for disease, (2) offering a simpler interpretation of common effect measures from statistical models such as odds ratios and relative risks. However, all advantages come at the lost information. From a statistical point of view, this loss of information means more samples which are required to attain prespecified powers.

Moser and Coombs [12] provided a closed-form relationship that allows a direct comparison between the logistic and linear regression coefficients. They also provided a procedure that allows the researcher to analyze the original continuous outcome without dichotomizing. To date, a method that applies the complementary log-log model without dichotomizing and without loss of information has not been available.

We aim to (a) provide a method that allows the researcher to estimate the coefficients of the complementary log-log model without dichotomizing and without loss of information, (b) show that the coefficient of the complementary log-log model can be interpreted in terms of the regression coefficients, (c) demonstrate that the coefficient estimates from this method have smaller variances and shorter confidence intervals than the dichotomizing method.

2. Methods

2.1. Model

Let y 1, y 2,…, y n be n independent observations on y, and let x 1, x 2,…, x p−1 be p − 1 predictor variables thought to be related to the response variable y. The multiple linear regression model for the ith observation can be expressed as

yi=β0+β1xi1+β2xi2++βp1xip1+Eii=1,2,,n, (1)

or

yi=xiβ+Eii=1,2,,n, (2)

where

xi=(1,xi1,xi2,,xip1). (3)

To complete the model, we make the following assumptions:

  1. E(E i) = 0 for i = 1,2,…, n,

  2. var(E i) = σ 2 for i = 1,2,…, n,

  3. the independent E i follows an extreme value distribution for i = 1,2,…, n.

Writing the model for each of the n observations, in matrix form, we have

[y1y2..yn]=[1x11x12x1p11x21x22x2p1..1xn1xn2xnp1][β0β1..βp1]+[E1E2..En], (4)

or

y=Xβ+E. (5)

The preceding three assumptions on E i and y i  can be expressed in terms of this model:

  1. E(E) = 0,

  2. cov⁡(E) = σ 2 I,

  3. the E i is extreme value (0, σ 2) for i = 1,2,…, n.

2.2. (Largest) Extreme Value Distribution

The PDF and CDF of the extreme value distribution are given by

f(yxβ,σ)=πσ6×exp(yxβkσσ×π6  exp(yxβkσσ×π6))x,σ0,P(yc)=exp(exp(cxβ+kσσ×π6))x,σ,  k0.45. (6)

It is easy to check that

ωj=lnπ1lnπ2=ln(p(ycx))ln(p(ycx(1,j)))=exp(((cxβ+kσ)/σ)×π/6)exp(((cx(1,j)β+kσ)/σ)×π/6)=exp(π6·βjσ)π1=π2exp((π/6)·(βj/σ)), (7)

where

x=(1,x1,,xj,,xp1),x(1,j)=(1,x1,,xj1,,xp1),β=(β0,  β1,,βj,,βp1). (8)

To return to a random sample of observations (y 1, y 2,…, y n), we conclude that the PDF and CDF of each independent y i are given by (6), and the corresponding equality (7) is given by

lnπ^1lnπ^2=exp(πσ^6β^j), (9)

where the estimate β^j  is the (j + 1)th element of vector β^=(β^0,β^1,,β^j,,β^p-1). It is readily shown that the results also hold true for the smallest extreme value distribution (Appendix A).

2.3. The Proposed Confidence Intervals

Let

β^=(β^0,β^1,,β^j,,β^p1)=(XX)1XYj=0,,p1,σ^2=Y(InX(XX)1X)Y(np). (10)

According to the preceding three assumptions on E i and  y i, we obtain

E(β^)=E[(XX)1XY]=(XX)1XEY=(XX)1XXβ=β,E(σ^2)=1npE(Y(InX(XX)1X)Y)=1np{tr[(InX(XX)1X)σ2I]+E(Y)[InX(XX)1X]E(Y)}=1np{σ2tr[InX(XX)1X]    +βX[InX(XX)1X]Xβ}=1np{σ2[ntr(X(XX)1X)]+βXXββXX(XX)1XXβ}=1np{σ2[ntr(X(XX)1X)]+βXXββXXβ}=1npσ2[ntr(IP)]=1npσ2(np)=σ2. (11)

Therefore, β^ and  σ^2  are unbiased estimators of β and  σ 2.

We have assumed that E i is distributed as an extreme value, and we use the approximation of the extreme value distribution of the errors E i by the normal distribution. For normally distributed observations, β^j/(σ^δj) follows a noncentral t distribution with np degree of freedom and noncentrality parameter -<βj/(σδj)<,

1α=P{t1(α/2)[np,βj(σδj)]<β^j(σ^δj)<tα/2[np,βj(σδj)]}, (12)

where t α/2[r, s] represents the 100(1 − (α/2))  percentile point of a noncentral t distribution with r degrees of freedom and noncentrality parameter − < s < , and δ j is the (j + 1)st diagonal element of (XX)−1. We use the approximation of the percentiles of the noncentral t distribution by the standard normal percentiles [13], then

1α=P{βj/(σδj)zα/2[1+(βj2/(σ2δj)zα/22)/2(np)]1/21(zα/22/2(np))  <β^j(σ^δj)<βj/(σδj)+zα/2[1+((βj2/(σ2δj)zα/22)/2(np))]1/21(zα/22/2(np))  },(βjσ)U={β^jσ^[1zα/222(np)]+zα/2[δj(1+((β^j2/σ^2δj)zα/222(np)))]1/2},(βjσ)L={β^jσ^[1zα/222(np)]zα/2[δj(1+((β^j2/σ^2δj)zα/222(np)))]1/2}, (13)

Thus, we obtain an approximate 100(1 − α) percent confidence interval for ω j

{exp[π6(βjσ)L],  exp[π6(βjσ)U]}. (14)

3. Comparison of the Two Methods

Let Y i be a continuous outcome variable. For fixed value of C, we define Y i* such that

Yi={1if  YiC,0if  Yi<C. (15)

Suppose that Y 1*,…, Y n* form a random sample of observations, and we fit a complementary log-log model

πi1=P(Yi=1xi)=exp(exp(xiθ)),πi2=P(Yi=1x(1,i))=exp(exp(x(1,i)θ)), (16)

where x i = (1,x i1,…,x i,p−1)′  is the P × 1 vector of covariates for the ith observation, and θ = (θ 0,…,θ p−1)′ is the P × 1 vector of unknown parameters. The dichotomized ω j* parameter corresponding to the effect θ j is

ωj=ln(π1)ln(π2)=ln(P(Y=1x))ln(P(Y=1x(1,j)))=(exp(xθ))(exp(x(1,j)θ))=exp(θj)j=0,,p1. (17)

In general, maximum likelihood estimation (MLE) can be used to estimate the parameter  θ = (θ 0,…, θ p−1). Let θ^=(θ^0,,θ^p-1) be the P × 1 ML estimate of θ, and let COV(θ^) be the P × P  covariance matrix of θ^. Using COV(θ^) from (23), one can construct confidence intervals. This matrix has as its diagonal the estimated variances of each of the ML estimates. The (j + 1)th diagonal element is given by σθ^j2. Therefore,

ω^j=exp(θ^j), (18)

and for large samples, (θ^jL,θ^jU)=(θ^j-zα/2σ^θ^j,θ^j+zα/2σ^θ^j) is a 100(1 − α) percent confidence interval for the true θ j. Then  (exp(θ^jL),exp(θ^jU)) is a 100(1 − α) percent confidence interval for the true ω j*.

We now compare the ω j from (7) with the ω j* from (17)

ωj=ln(π1)ln(π2)ωj=ln(π1)ln(π2)ωj=ωjexp(π6·βjσ)=exp(θj)π6·βjσ=θjβj,θj,σ. (19)

This show that the coefficient of the complementary log-log model, θ j, can be interpreted in terms of the regression coefficients, β j. Note that β are related to the responses through the general linear regression model

yi=xiβ+Eii=1,,n, (20)

where the independent E i are distributed as an extreme value with mean 0 and variance σ 2 > 0.

4. Covariance Matrix of Model Parameter Estimators

4.1. Derivation of var(ω j*) for Large n

The information matrix of generalized linear models has the form ∫ = XWX [1], where W is the diagonal matrix with diagonal elements w i = (∂μ i/∂η i)2/(var(y i)), y is response variable with independent observations (y 1,…y n), and x ij  denote the value of predictor j,

μi=E(yi),ηi=g(μi)=jθjxij,j=0,1,,p1. (21)

The covariance matrix of θ^ is estimated by  (XW^X)-1.

Maximum likelihood estimation for the complementary log-log model is a special case of the generalized linear models. Let

μi=πi=exp(exp(jθjxij))πi=exp(exp(ηi)),μiηi=(exp(ηi))exp(exp(ηi))=πilnπi,wi=(πIlnπi)2πi(1πi)=πi(lnπi)21πi, (22)

then

XWX=[i=1nπi(lnπi)21πii=1nxi1πi(lnπi)21πii=1nxi,p1πi(lnπi)21πii=1nxi1πi(lnπi)21πii=1nxi12πi(lnπi)21πii=1nx1xi,p1πi(lnπi)21πii=1nxi,p1πi(lnπi)21πii=1nx1ixi,p1πi(lnπi)21πii=1nxi,p12πi(lnπi)21πi]. (23)

It is readily shown that the results hold true for the largest extreme value distribution (Appendix A).

In large samples, var(θ^j) approaches σθj2|θ=θ^ [14] which equals the (j + 1)th diagonal element of (XWX)−1.

By applying the delta method, let f(θ^j)=exp(θ^j), then

var(ω^j)var(exp(θ^j))=var(f(θ^j))=(f(θ^j)θ^j|θ^j=θj)2(var(θ^j))=(exp(θj))2×σθ^j2. (24)

4.2. Derivation of var(ω^j) for Large n

In large samples, from (10) σ^2σ2 [15]. Therefore,

var(ω^j)=var(exp(πβ^jσ^6))var(exp(πβ^jσ6)). (25)

In addition, var(β^j)=σ2δj.

By applying the delta method, let g(β^j)=exp(πβ^j/(σ6)), then

var(ω^j)var(exp(πβ^jσ6))=var(g(β^j))=(g(β^j)β^j|β^j=βj)2×var(β^j)=(πσ6exp(πβjσ6))2σ2δj=π26δj(expπβjσ6)2. (26)

5. Sample Sizes Saving

5.1. The Power for the Dichotomized Method

In large samples, σ^θ^j converges to σθ^j almost surely [14]. Therefore, for a given value of ω j = exp⁡θ j (i.e., ln⁡ω j = θ j), the power is given by

p(ωj)=p{rejection  of  ωj=1ωj1}=p{exp(θjL)>1θj}+p{exp(θjU)<1θj}=p{θ^j>zα/2σθ^jθj}+p{θ^j<zα/2σθ^jθj}=p{Z>zα/2σθ^jlnωjσθ^j}+p{Z<zα/2σθ^jlnωjσθ^j}=p{Z>zα/2lnωjσθ^j}+p{Z<zα/2lnωjσθ^j}=P{Z>z1}+P{Z<z2}, (27)

where

{z1=zα/2lnωjσθ^jz2=zα/2+lnωjσθ^j}. (28)

5.2. The Power for the Proposed Method

In large samples, σ^ converges to σ almost surely [15]. Therefore, for a given value of ωj=exp(πβj/σ6) (i.e., βj=σ(lnωj6/π)), the power is given by

p(ωj)=p{exp(π6(βjσ)L)>1ωj}+p{exp(π6(βjσ)U)<1ωj}=P{βJL>zα/2σδjβj=σlnωj6π}+P{βJU<zα/2σδjβj=σlnωj6π}=p{Z>zα/2σδj(σlnωj6/π)σδj}+p{Z<zα/2σδj(σlnωj6/π)σδj}=p{Z>zα/2lnωj6πδj}+p{Z<zα/2lnωj6πδj}=p{Z>z1}+p{Z<z2}, (29)

where

  {z1=zα/2lnωj6πδjz2=zα/2+lnωj6πδj}. (30)

Our proposed method, since it is based on continuous data rather than dichotomized, is likely to be more powerful. We show that the proposed method can produce substantial sample size saving for a given power. Let

  1. the number of parameters p = 2 (i.e., θ = (θ 0, θ 1)),

  2. x i = (1,x i1)′, x i1 ∈ {−a + (2an/(g − 1)) | n = 0,…, g − 1}, that is, x i1  follows a discrete uniform distribution with range (−a, a). For simplicity, a = 2.

  3. Total samples are n and n* for the proposed and dichotomized methods, respectively. These samples included k and k* set of these g uniformly distributed points for the proposed and dichotomized methods, respectively. That is, n = gk and n* = gk*, then

δj=[ki=1g(x1ix̅1.)2]1,j=1, (31)

and from (23),

σθ^j2=i=1g((πi)(ln(πi))2/ln(1πi))(k){i=1gx1i2((πi)(ln(πi))2/ln(1πi))i=1g((πi)(ln(πi))2/ln(1πi))[i=1gx1i((πi)(ln(πi))2/ln(1πi))]2}. (32)

We consider the same power for two methods:

z1=z1z2=z2{zα/2lnωjσθ^j=zα/2lnωj6πδjzα/2+lnωjσθ^j=zα/2+lnωj6πδjπ6δj=σθ^j,j=1π6[ki=1g(x1ix̅1.)2]1=i=1g((πi)(ln(πi))2/ln(1πi))(k){i=1gx1i2((πi)(ln(πi))2/ln(1πi))i=1g((πi)(ln(πi))2/ln(1πi))[i=1gx1i((πi)(ln(πi))2/ln(1πi))]2} (33)

relative sample size

nn=kk=6σθ^j2π2δj=i=1g(x1ix̅1.)2×i=1g((πi)(ln(πi))2/ln(1πi))(π2/6){i=1gx1i2((πi)(ln(πi))2/ln(1πi))i=1g((πi)(ln(πi))2/ln(1πi))[i=1gx1i((πi)(ln(πi))2/ln(1πi))]2}. (34)

That is, (34) is independent of σ 2 and applies for any power, and any test size α.

Table 1 presents relative sample sizes n*/n for a given fixed parameter ω j*  and an average proportion of success π̅. We consider the situations in which π̅=i=1g(πi/g)=0.1,0.2,0.3,0.4,0.5, g = 9, ω j* = 0.25,0.50,0.75.

Table 1.

Relative sample sizes required to attain any power for the dichotomizing method versus the proposed method.

ω* = exp⁡(θ) Average proportion of successes (π̅)
0.1 0.2 0.3 0.4 0.5
0.25 23.7166 9.5092 7.4954 7.1996 6.8575
0.50 10.6719 5.4176 3.4215 2.5209 2.1784
0.75 7.7088 3.8713 2.5171 1.9380 1.5841

For given fixed ω j* and π̅, the relative sample sizes in Table 1 can be computed by the following step:

  1. compute the value θ j via the equation θ j = ln⁡(ω j*),

  2. calculate the cut-off point C iteratively such that π̅ attained the specified value for the values x i1, using the value of θ j  in (i).

As can be seen from Table 1, all values are greater than 1. The values of n*/n increase as the  ω j* moves farther away from 1. Values of Table 1 immediately highlight the improvement accomplished by the proposed method.

6. Relative Efficiency of ω^j with ω^j

Here, we examine the relative efficiency of the estimate  ω^j to the estimate ω^j.

Using (24) and (26), the relative efficiency is given by

r.e.  (ω^j,ω^j)=var(ω^j)var(ω^j)=6(exp(θj))2×σθ^j2π2δj(exp(λβj/σ))2=6σθ^j2π2δj. (35)

Note that the relative efficiency is independent of n and σ 2 and converges to a constant. Comparing (34) and (35), the relative efficiency equals the relative sample sizes. Therefore, as in Table 1, the proposed method is a consistent improvement over the dichotomizing method with respect to relative efficiencies.

It should be noted that these results hold true under the following assumptions:

  1. the responses y i and β are related through the equation y i = x i β + E i where the independent  E i are distributed as an extreme value with mean 0 and variance σ 2 > 0,

  2. the independent variables x i follow a discrete uniform distribution.

7. Odds Ratio

For values of π larger than 0.90, −ln⁡(π) and π/(1 − π)  are very close. Hence, for large values of π,

ln(π1)ln(π2)π1/1π1π2/1π2=OR. (36)

And from (7), odds ratio is given by

OR=exp(π6·βjσ). (37)

The parameters estimated from the linear regression can be interpreted as an odds ratio.

8. Simulation Study

It should be noted that, as in Table 1, the proposed method is a consistent improvement over the dichotomizing method with respect to relative efficiencies. These results hold true under the assumption that predictor variable has a discrete uniform distribution and that the random variables E i follow an extreme value distribution. To demonstrate the robustness of this conclusion to changes in the distributions of predictor variables, simulations were run under different distributional conditions. The data were sampled 10000 times for three sample sizes {n = 250, 500, 1000}, three average proportions of successes {π̅=  0.10,  0.50,  0.95}, and seven ω j{ω j = 0.75, 0.90, 1.1, 1.2, 1.3, 1.4, 1.5}. The simulated data are generated using the following algorithm

  1. Generate y i, where y i = β 0 + β 1 x i + E i, β1=6lnωj/π through (7) to produce the correct ω j, and for simplicity β 0 = 0, σ 2 = 1.

  2. For fixed π̅, generate cutoff point C using (15).

We simulated the data for two scenarios based on the distribution of the explanatory variable. In the first scenario, the independent variable follows a continuous uniform distribution and range (−2, 2), and in the second, the independent variable follows a truncated normal distribution with mean 0 and range (−2, 2). The relative mean square errors, relative interval lengths, absolute biases, and the probability of coverage were calculated.

Results of the simulations addressing the validity of the proposed method are displayed in Tables 2 and 3.

Table 2.

Simulated relative mean square errors, relative intervals lengths, coverage probabilities, and absolute biases for the proposed and dichotomizing methods (using a continuous uniform distribution for the explanatory variable and an extreme value distribution for the errors).

Sample size ω
Cut off
.75 .9 1.1 1.2 1.3 1.4 1.5
1.15a 1.07 1.09 1.14 1.24 1.47 1.71
1.10b 1.03 1.03 1.07 1.14 1.23 1.35
0.10 0.943c 0.948 0.949 0.949 0.945 0.938 0.933
0.948d 0.947 0.949 0.947 0.951 0.947 0.953
0.05e 0.04 0.12 0.14 0.10 0.15 0.11
0.07f 0.01 0.17 0.13 0.24 0.34 0.58
1.23 1.26 1.27 1.28 1.27 1.24 1.26
2.16 1.13 1.23 1.14 1.15 1.17 1.19
1000 0.50 0.940 0.951 0.951 0.945 0.942 0.937 0.934
0.951 0.949 0.951 0.950 0.948 0.947 0.948
0.04 0.01 0.08 0.10 0.05 0.09 0.04
0.05 0.04 0.15 0.12 0.09 0.12 0.13
12.75 12.44 13.22 12.68 13.14 12.91 12.79
3.67 3.57 3.58 3.63 3.69 3.76 3.84
0.95 0.943 0.951 0.952 0.944 0.944 0.938 0.929
0.952 0.954 0.952 0.952 0.951 0.951 0.951
0.04 0.07 0.11 0.10 0.10 0.17 0.10
0.75 0.68 0.86 1.01 1.21 1.45 1.24

1.30 1.08 1.07 1.17 1.24 1.54 1.95
1.16 1.03 1.04 1.08 1.15 1.25 1.39
0.10 0.942 0.950 0.951 0.95 0.944 0.941 0.936
0.951 0.950 0.949 0.951 0.954 0.954 0.953
0.12 0.07 0.24 0.25 0.21 0.18 0.29
0.23 0.08 0.33 0.39 0.41 0.73 1.21
1.35 1.10 1.27 1.26 1.26 1.25 1.26
1.26 1.03 1.13 1.14 1.16 1.17 1.20
500 0.50 0.940 0.949 0.947 0.948 0.943 0.940 0.933
0.952 0.951 0.949 0.949 0.954 0.950 0.951
0.23 0.34 0.27 0.23 0.26 0.25 0.38
0.48 0.11 0.17 0.18 0.31 0.26 0.42
13.04 13.17 13.8 13.90 14.45 14.48 14.47
3.72 3.65 3.68 3.73 3.82 3.91 3.99
0.95 0.942 0.947 0.951 0.949 0.947 0.938 0.935
0.953 0.952 0.954 0.955 0.955 0.953 0.954
0.05 0.11 0.08 0.08 0.24 0.32 0.27
0.94 1.38 1.78 1.92 2.52 3.00 2.90

13.41 14.46 1.12 1.28 1.52 1.96 2.33
3.78 3.73 1.04 1.09 1.18 1.30 1.45
0.10 0.942 0.949 0.949 0.945 0.942 0.942 0.933
0.957 0.954 0.948 0.949 0.952 0.957 0.953
0.02 0.20 0.38 0.33 0.42 0.41 0.66
2.11 2.74 0.42 0.84 1.18 1.78 2.24
1.27 1.25 1.32 1.28 1.30 1.30 1.29
1.16 1.13 1.13 1.14 1.16 1.18 1.20
250 0.50 0.941 0.948 0.952 0.947 0.945 0.943 0.933
0.951 0.951 0.951 0.950 0.951 0.951 0.951
0.12 0.13 0.35 0.44 0.41 0.53 0.55
0.11 0.22 0.39 0.47 0.51 0.74 0.59
12.98 14.6 15.64 15.46 17.05 16.89 18.33
3.75 3.72 3.82 3.88 4.01 4.12 4.29
0.95 0.945 0.955 0.946 0.948 0.940 0.937 0.932
0.959 0.955 0.955 0.959 0.958 0.957 0.952
0.02 0.16 0.39 0.22 0.46 0.47 0.51
1.22 2.75 3.97 3.98 4.99 5.19 6.19

a: Relative mean square errors, b: Relative intervals lengths, c: Coverage probability (proposed), d: Coverage probability (dichotomized), e: % bias (proposed), f: % bias (dichotomized).

Table 3.

Simulated relative mean square errors, relative intervals lengths, coverage probabilities, and absolute biases for the proposed and dichotomizing methods (using a truncated normal distribution for the explanatory variable and an extreme value distribution for the errors).

Sample size ω Cut off .75 .9 1.1 1.2 1.3 1.4 1.5
1.17a 1.02 1.08 1.13 1.19 1.28 1.36
1.11b 1.03 1.03 1.06 1.10 1.25 1.22
0.10 0.942c 0.948 0.948 0.952 0.944 0.942 0.940
0.951d 0.951 0.950 0.952 0.949 0.951 0.951
0.08e 0.06 0.03 0.14 0.13 0.14 0.16
0.10f 0.11 0.15 0.23 0.30 0.39 0.39
1.26 1.24 1.26 1.28 1.28 1.25 1.28
1.24 1.13 1.13 1.14 1.14 1.15 1.17
1000 0.50 0.944 0.948 0.952 0.947 0.947 0.944 0.941
0.948 0.951 0.949 0.949 0.947 0.950 0.949
0.02 0.09 0.08 0.07 0.18 0.16 0.13
0.03 0.06 0.12 0.16 0.20 0.16 0.14
12.33 13.12 13.03 12.71 12.86 12.55 12.88
3.62 3.59 3.61 3.62 3.64 3.68 3.71
0.95 0.944 0.951 0.948 0.948 0.945 0.945 0.946
0.952 0.948 0.95 0.949 0.949 0.951 0.952
0.10 0.04 0.11 0.04 0.16 0.16 0.20
1.26 1.05 1.56 1.36 1.43 1.80 1.94

1.18 1.09 1.06 1.75 1.23 1.32 1.58
1.11 1.03 1.03 1.06 1.11 1.16 1.23
0.10 0.945 0.95 0.951 0.951 0.949 0.943 0.944
0.953 0.953 0.953 0.950 0.949 0.951 0.950
0.04 0.13 0.31 0.18 0.33 0.36 0.37
0.21 0.08 0.37 0.50 0.62 0.69 0.96
1.25 1.27 1.27 1.29 1.27 1.29 1.25
1.14 1.13 1.13 1.14 1.15 1.16 1.17
500 0.50 0.944 0.948 0.949 0.947 0.948 0.944 0.935
0.951 0.951 0.951 0.948 0.951 0.948 0.949
0.13 0.22 0.35 0.37 0.35 0.30 0.44
0.16 0.19 0.39 0.48 0.44 0.41 0.54
13.11 14.02 14.02 13.5 13.54 13.80 14.32
3.73 3.71 3.73 3.75 3.77 3.81 3.86
0.95 0.944 0.95 0.951 0.950 0.947 0.944 0.944
0.954 0.95 0.951 0.953 0.948 0.956 0.953
0.15 0.10 0.24 0.38 0.32 0.33 0.43
2.50 2.70 2.92 3.10 2.92 3.36 3.89

1.28 1.11 1.12 1.19 1.33 1.54 1.76
1.11 1.03 1.04 1.08 1.13 1.19 1.28
0.10 0.947 0.951 0.950 0.947 0.950 0.950 0.942
0.951 0.950 0.950 0.952 0.954 0.952 0.951
0.40 0.34 0.37 0.64 0.69 0.58 0.81
0.26 0.06 0.69 1.08 1.30 1.55 2.22
1.32 1.30 1.27 1.33 1.31 1.33 1.31
1.15 1.13 1.13 1.14 1.18 1.17 1.18
250 0.50 0.951 0.95 0.953 0.951 0.940 0.945 0.940
0.949 0.951 0.952 0.948 0.948 0.950 0.948
0.22 0.43 0.57 0.69 0.66 0.58 0.66
0.38 0.53 0.64 0.89 0.91 0.82 0.91
14.09 14.51 16.27 15.91 15.89 15.73 15.60
3.86 3.87 3.93 3.92 3.98 4.04 4.11
0.95 0.943 0.95 0.951 0.951 0.947 0.944 0.937
0.953 0.95 0.953 0.956 0.953 0.956 0.952
0.30 0.37 0.57 0.68 0.42 0.62 0.75
4.98 5.52 6.547 5.91 6.17 6.88 7.72

a: Relative mean square errors, b: Relative intervals lengths, c: Coverage probability (proposed), d: Coverage probability (dichotomized), e: % bias (proposed), f: % bias (dichotomized).

The simulations show that the relative mean square errors are all greater than 1, increasing with the average proportion of successes and when the ω j moves farther away from 1. The results in Tables 1 and 2 demonstrate that the proposed method provides confidence intervals which successfully maintain their nominal 95 percent coverage. For the proposed method in first scenario, 51 out of 63 coverage probabilities fell within (0.94, 0.96), and all 63 coverage probabilities are greater than 0.93 and, in the second scenario, almost all coverage probabilities fell within (0.94, 0.96). The absolute biases for proposed method are never greater than a few percent. The proposed method is less biased than the dichotomizing method in 6 of 63 simulations in both two scenarios.

9. An Example

To illustrate the application of the proposed method presented in the previous section, we utilize the data arising from the National Health Survey in Iran. The other analyses using this data appear in many places [16].

In this study, 14176 women aged 20–69 years were investigated. BMI (body mass index), our dependent variable, was calculated as weight in kilograms divided by height in meters squared (kg/m2). Independent variables included place of residence, age, smoking, economic index, marital status, and education level. The independent variables considered were both categorical and continuous. At first, BMI was treated as a continuous variable, and ω^j and 95 percent confidence intervals were calculated using the proposed linear regression method. Then subjects were classified into obese (BMI ≥ 30 kg/m2) and nonobese (BMI <30 kg/m2). A complementary log-log model was used for the binary analysis, with obese or nonobese used as the outcome measure. The ω^j and 95 percent confidence intervals were calculated using the dichotomized method. Table 4 presents the coefficient estimates, estimated confidence intervals, and relative confidence interval lengths. The proposed and dichotomizing methods produced different confidence intervals, although the ω^j and ω^j were similar only varying slightly. The ω^j  estimate from the proposed method had smaller variances and shorter confidence intervals than the dichotomizing method. All relative confidence interval lengths were greater than 2.58.

Table 4.

Adjusted ω^j, ω^j  for obesity and confidence intervals using two methods for the National Health Survey.

Covariates ω^j(ω^j) 95% CIa (proposed) 95% CI (dichotomized) Relativeb length of CI
Place of residence 1.65 (1.97)c 1.58–1.74 1.79–2.18 2.43
Age 1.021 (1.019) 1.018–1.022 1.015–1.022 1.75
Years of education 0.99 (0.98) 0.985–0.997 0.971–0.994 1.92
Smoking 0.76 (0.68) 0.66–0.90 0.51–0.92 1.71
Marital status 1.16 (1.42) 1.10–1.22 1.27–1.58 2.58
Lower-middle economy index 1.24 (1.32) 1.14–1.32 1.18–1.48 1.67
Upper-middle economy index 1.21 (1.26) 1.14–1.29 1.12–1.42 2.0
High economy index 1.20 (1.21) 1.11–1.30 1.08–1.36 1.47

aconfidence interval, bdichotomized/proposed, cproposed (dichotomized).

10. Discussion

When assuming the errors  E i  are distributed as an extreme value distribution, as noted before, the method has several advantages. First, the method allows the researcher to apply the complementary log-log model without dichotomizing and without loss of information. Second, the ω^j from the dichotomizing method is dependent on the chosen cutoff point C and will vary with c. However, the proposed ω^j is independent of the c since ω^j is a function of the continuous Y i  and not a function of the dichotomized Y i*  defined through C. Third, we show that the coefficient of the complementary log-log model, θ j, can be interpreted in terms of the regression coefficients, β j. Fourth, when the independent variables x i follow a discrete uniform distribution, the proposed method is a consistent improvement over the dichotomizing method with respect to relative efficiencies. The proposed method can provide sample size saving, smaller variances, and shorter confidence intervals than the dichotomized method. Fifth, when π  is large, the parameters estimated from the linear regression can be interpreted as odds ratios.

Our results were consistent with the findings by Moser and Coombs [12] and Bakhshi et al. [16] showing the greater efficiency of parameter estimates from the regression method that avoids dichotomizing in comparison with a more traditional dichotomizing method using the logistic regression.

Our main recommendation is to let continuous response remain continuous. Do not throw away information by transforming the data to binary. This means that if the objective is to estimate and/or test coefficients when responses are continuous, please resist dichotomizing your response variable.

Conflict of Interests

The authors have declared no conflict of interests.

Appendix

A. Largest Extreme Value Distribution

(a) The PDF and CDF are given by

f(yxβ,σ)=πσ6×exp(yxβkσσ×π6  exp(yxβkσσ×π6))x,σ0,P(yc)=1exp(exp(cxβkσσ×π6))x,σ0, (A.1)

where Y is a continuous outcome variable, x = (1, x 1,…, x p−1) is the p × 1 vector of known independent variables, β = (β 0, β 1,…, β p−1) is the p × 1 vector of unknown parameters, and k ≈ 0.45.

It is easy to check that

ωj=ln(1π1)ln(1π2)=ln(1p(ycx))ln(1p(ycx(1,j)))=exp(((cxβkσ)/σ)×(π/6))exp(((cx(1,j)βkσ)/σ)×(π/6))=exp(π6·βjσ)1π1=(1π2)exp((π/6)·(βj/σ)),   (A.2)

where

x=(1,x1,,xj,,xp1),x(1,j)=(1,x1,,xj1,,xp1),β=(β0,β1,,βj,,βp1). (A.3)

(b) Suppose that E i is distributed as a largest extreme value with mean 0 and variance σ 2 > 0. We conclude that the PDF and CDF of each independent Y i are given by (A.1), and the corresponding equality (A.2) is given by

ω^j=ln(1π^1)ln(1π^2)=exp(π6·β^jσ^). (A.4)

(c) Similar to largest extreme value distribution

μi=πi=1exp(exp(jθjxij))πi=1exp(exp(ηi)),μiηi=(exp(ηi))exp(exp(ηi))=(1πi)ln(1πi)wi=((1πi)ln(1πi))2πi(1πi)=(1πi)(ln(1πi))2πi, (A.5)

then

XWX=[i=1n(1πi)(ln(1πi))2πii=1n(1πi)(ln(1πi))2πii=1nxi,p1(1πi)(ln(1πi))2πii=1nxi1(1πi)(ln(1πi))2πii=1nxi12(1πi)(ln(1πi))2πii=1nx1xi,p1(1πi)(ln(1πi))2πii=1nxi,p1(1πi)(ln(1πi))2πii=1nxixi,p1(1πi)(ln(1πi))2πii=1nxi,p12(1πi)(ln(1πi))2πi]. (A.6)

References

  • 1.Agresti A. Categorical Data Analysis. 2nd edition. New York, NY, USA: Wiley; 2002. [Google Scholar]
  • 2.Zhao LP, Kolonel LN. Efficiency loss from categorizing quantitative exposures into qualitative exposures in case-control studies. American Journal of Epidemiology. 1992;136(4):464–474. doi: 10.1093/oxfordjournals.aje.a116520. [DOI] [PubMed] [Google Scholar]
  • 3.MacCallum RC, Zhang S, Preacher KJ, Rucker DD. On the practice of dichotomization of quantitative variables. Psychological Methods. 2002;7(1):19–40. doi: 10.1037/1082-989x.7.1.19. [DOI] [PubMed] [Google Scholar]
  • 4.Cohen J. The cost of dichotomization. Applied Psychological Measurement. 1983;7(3):249–253. [Google Scholar]
  • 5.Greenland S. Avoiding power loss associated with categorization and ordinal scores in dose-response and trend analysis. Epidemiology. 1995;6(4):450–454. doi: 10.1097/00001648-199507000-00025. [DOI] [PubMed] [Google Scholar]
  • 6.Austin PC, Brunner LJ. Inflation of the type I error rate when a continuous confounding variable is categorized in logistics regression analyses. Statistics in Medicine. 2004;23(7):1159–1178. doi: 10.1002/sim.1687. [DOI] [PubMed] [Google Scholar]
  • 7.Vargha A, Rudas T, Delaney HD, Maxwell SE. Dichotomization, partial correlation, and conditional independence. Journal of Educational and Behavioral Statistics. 1996;21(3):264–282. [Google Scholar]
  • 8.Maxwell SE, Delaney HD. Bivariate median splits and spurious statistical significance. Psychological Bulletin. 1993;113(1):181–190. [Google Scholar]
  • 9.Streiner DL. Breaking up is hard to do: the heartbreak of dichotomizing continuous data. Canadian Journal of Psychiatry. 2002;47(3):262–266. doi: 10.1177/070674370204700307. [DOI] [PubMed] [Google Scholar]
  • 10.Chen H, Cohen P, Chen S. Biased odds ratios from dichotomization of age. Statistics in Medicine. 2007;26(18):3487–3497. doi: 10.1002/sim.2737. [DOI] [PubMed] [Google Scholar]
  • 11.Ragland DR. Dichotomizing continuous outcome variables: dependence of the magnitude of association and statistical power on the cutpoint. Epidemiology. 1992;3(5):434–440. doi: 10.1097/00001648-199209000-00009. [DOI] [PubMed] [Google Scholar]
  • 12.Moser BK, Coombs LP. Odds ratios for a continuous outcome variable without dichotomizing. Statistics in Medicine. 2004;23(12):1843–1860. doi: 10.1002/sim.1776. [DOI] [PubMed] [Google Scholar]
  • 13.Johnson NL, Welch H, Wei CZ. Application of the non-central t distribution. Biometrika. 1940;31(3-4):362–389. [Google Scholar]
  • 14.Serfling RJ. Approximation Theory of Mathematical Statistics. New York, NY, USA: Wiley; 1980. [Google Scholar]
  • 15.Lai TL, Robbins H, Wei CZ. Strong consistency of least squares estimates in multiple regression. Proceedings of the National Academy of Sciences of the United States of America. 1978;75(7):3034–3036. doi: 10.1073/pnas.75.7.3034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Bakhshi E, Eshraghian MR, Mohammad K, Seifi B. A comparison of two methods for estimating odds ratios: results from the National Health Survey. BMC Medical Research Methodology. 2008;8, article 78 doi: 10.1186/1471-2288-8-78. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Computational and Mathematical Methods in Medicine are provided here courtesy of Wiley

RESOURCES