Skip to main content
Biometrika logoLink to Biometrika
. 2010 Aug 30;97(4):825–838. doi: 10.1093/biomet/asq048

Noncrossing quantile regression curve estimation

Howard D Bondell 1, Brian J Reich 1, Huixia Wang 1
PMCID: PMC3371721  PMID: 22822254

Summary

Since quantile regression curves are estimated individually, the quantile curves can cross, leading to an invalid distribution for the response. A simple constrained version of quantile regression is proposed to avoid the crossing problem for both linear and nonparametric quantile curves. A simulation study and a reanalysis of tropical cyclone intensity data shows the usefulness of the procedure. Asymptotic properties of the estimator are equivalent to the typical approach under standard conditions, and the proposed estimator reduces to the classical one if there is no crossing. The performance of the constrained estimator has shown significant improvement by adding smoothing and stability across the quantile levels.

Some key words: Crossing quantile curve, Heteroscedastic error, Quantile regression, Robustness, Smoothing spline, Tropical cyclone

1. Introduction

Quantile regression has become a useful tool to complement a typical least-squares regression analysis (Koenker, 2005). Modelling of the median as opposed to the mean is much more robust to outlying observations. Additionally, examining the effect of the predictors on other quantiles can yield a clearer picture regarding the overall distribution of a response. In particular, in numerous instances, interest focuses on the effect of the predictors on the tails of a distribution, in addition to, or instead of, the centre. Quantile regression is used to model the conditional quantiles of a response. Applications of quantile regression come from diverse areas, including economics, public health, meteorology and surveillance.

However, when an investigator wishes to use quantile regression at multiple percentiles, the quantile curves can cross, leading to an invalid distribution for the response. Given a set of covariates, it may turn out, for example, that the predicted 95th percentile of the response is smaller than the 90th percentile, which is impossible.

Consider a recent application of quantile regression to model tropical cyclone intensity (Jagger & Elsner, 2009). The goal is to model the maximum wind speed for near-coastal tropical cyclones occurring near the US coastline based on climatological variables. Of particular interest is the upper tail of the distribution, as these are the storms that may cause major damage. Jagger & Elsner (2009) used quantile regression to examine the upper tail behaviour as a function of four large-scale climate conditions. A sample of 422 tropical cyclones is used to model the maximum wind speed in terms of the climate covariates: the North Atlantic Oscillation Index, the Southern Oscillation Index, the Atlantic sea-surface temperature and the average sunspot number.

With the focus being on the upper quantiles of the wind speed distribution, those corresponding to category 4 and 5 hurricanes, consider using these data to fit a quantile regression at the set of percentiles (0.25, 0.5, 0.75, 0.9, 0.95, 0.99). The fitted slopes of the quantile functions give the effects of the covariates at the various levels of cyclone intensity. One particular issue with fitting the upper quantiles is the lack of data, so fitting individual quantile curves can be even more problematic.

For example, if one were to examine the fitted quantiles for these data, the upper quantiles cross not far from the mean. Due to the crossing, for instance, the practitioner would be forced to claim that when the North Atlantic Oscillation Index is one standard deviation below its mean and the other three are one standard deviation above, the 90th percentile of the distribution of wind speeds is larger than the 95th percentile. In addition, as further discussed in the analysis later, inference regarding the significant predictors changes dramatically from one quantile to the next. For example, predictors that appear highly significant in both the 90th percentile and the 99th percentile may not appear significant at the 95th percentile. Although this is possible, much of it can be attributed to the fact that the quantile functions are estimated separately.

This problem is well known, but no simple and general solution currently exists. For linear quantile regression, Koenker (1984) considered parallel quantile planes to avoid the crossing problem. Cole (1988) and Cole & Green (1992) assumed that a suitable transformation would yield normality of the response and proceeded to obtain nonparametric estimates of the transformation along with the location and scale, which then fully determine the quantile functions.

Similarly, He (1997) proposed a method to estimate the quantile curves while ensuring non-crossing. However, the approach assumes a heteroscedastic regression model for the response, which allows the predictors to affect the distribution of the response via a location and scale change of an underlying base distribution. Although this is a flexible model, the predictors may affect the response distribution in a less structured manner, which may not be captured by this model. Furthermore, since the procedure is a sequential algorithm, the distributional properties of the estimator are unclear. Simulation has also shown that even when the assumed heteroscedastic model is correct, the estimation procedure does not necessarily improve upon the unconstrained quantile regression estimator in finite samples. Wu & Liu (2009) have recently proposed an algorithm to ensure noncrossing, by fitting the quantiles sequentially and constraining the current curve to not cross the previous curve. One drawback of this algorithm is its dependence on the order that the quantiles are fitted. Neocleous & Portnoy (2008) discuss interpolation of the typical regression quantiles to ensure that, asymptotically, the probability of crossing will tend to zero for the full quantile process.

The crossing problem persists for nonlinear quantile curves. Several authors have proposed to first estimate the conditional cumulative distribution function via local weighting, and then invert it to obtain the quantile curve. Hall et al. (1999), Dette & Volgushev (2008) and Chernozhukov et al. (2009) enforce the noncrossing via this approach by modifying the estimate of the conditional distribution function. This indirect approach is used if interest is purely in estimation of the conditional quantile. However, when interest also focuses on quantifying the effects of the predictors, the quantile curves are typically modelled via a parametric form, such as linear predictor effects, and a direct estimation approach is required.

In this paper, a direct correction to the quantile regression optimization problem is used to ensure noncrossing quantile curves for any given sample. The approach is also extended to nonparametric quantile curves.

2. Noncrossing quantile regression

2.1. Alleviating the crossing problem

Let x = (x1, . . . , xp)T, and denote z = (1, xT)T. Let 𝒟Rp, be a closed convex polytope, represented as the convex hull of N points in p-dimensions. Interest focuses on ensuring that the quantile curves do not cross for all values of the covariate x𝒟. Although the region of interest for noncrossing is assumed to be bounded, the covariate space itself may still be unbounded. If noncrossing were desired in linear quantile regression on an unbounded domain in any covariate direction, the result will be parallel lines, yielding the constant slope, location-shift model.

Assuming a linear quantile model, the τth conditional quantile of the response is given by ziTβτ, i.e. pr(yiziTβτ|xi)=τ. The classical estimator of the regression coefficients for this quantile function is given by

β^τ=argminβi=1nρτ(yiziTβ), (1)

where ρτ (u) = u{τI (u < 0)} is the so-called check function.

A typical quantile regression analysis will solve (1) separately for each of the q desired quantile levels, τ1 < … < τq, to get β^(τ)=(β^τ1T,,β^τqT)T. Without any restriction, these resulting regression functions will often cross in finite samples, and hence the resulting conditional quantile curve for a given x will not be a monotonically increasing function of τ. Recall that z = (1, x)T. Then, formally, nonmonotonicity of the resulting estimated quantile function at a given point x is given by zTβ^τt<zTβ^τt1 for at least one t ∈ (2, . . . , q). For a simple example, it often turns out to be the case that the intercept is not an increasing function of τ, hence at x = 0 the quantile function is nonmonotone. This becomes even more problematic with a larger number of predictors as the curves have a much larger space in which they may cross. When q increases, crossing also becomes much more likely.

To alleviate the crossing issue, it is proposed to estimate the quantiles simultaneously under the noncrossing restriction. Specifically, the optimization problem becomes

β^(τ)=argminβt=1qw(τt)i=1nρτt(yiziTβτt)subjecttozTβτtzTβτt1,x𝒟,t=2,,q, (2)

for some weight function, w(τt), such that w(τt) > 0 for all t = 1, . . . , q.

Without the restriction, the solution to (2) is exactly the solution to (1) regardless of the choice of weight function w(τt). In fact, a direct consequence of this formulation is that if the classical estimator obeys the noncrossing constraint for a given dataset, the resulting estimator will be the classical estimator. Furthermore, the asymptotic distribution of the estimator, as discussed in the next section, will not depend on the choice of weight function. A convenient practical choice of weight function is to equally weight the terms, i.e. w(τt) = 1 for all t. This is the choice used in the examples.

Since the domain is the convex hull of N points, it suffices to enforce the noncrossing restriction at each vertex. Letting (z1, . . . , zN) denote the set of vertices, then any point in the region can be expressed as k=1Nckzk with k=1Nck=1 and each ck ⩾ 0. If zkTβτtzkTβτt1 for all k = 1, . . . , N, it follows that k=1NckzkTβτtk=1NckzkTβτt1. This can then be solved via standard linear programming with N * (q − 1) linear constraints, where N is the number of vertices and q is the number of quantiles. This may result in a large number of constraints.

However, it will now be shown that if the domain of interest can be reduced to [0, 1]p, constraints at each of the 2p vertices are unnecessary, and only a total of q − 1 constraints are needed rather than N (q − 1). Any domain of interest, for which there exists an invertible affine transformation that maps to [0, 1]p, can be used so that the transformation is performed, and then transformed back after the estimation, while retaining the noncrossing property. This has the potential to simplify computation a great deal. Hence, one may wish to approximate the desired domain by one of this form. The remainder of the article will focus on the domain 𝒟 = [0, 1]p.

To ensure noncrossing everywhere, consider the transformation from βτ1, . . . , βτq to γτ1, . . . , γτq, where γτ1 = (γ0,τ1, γ1,τ1, . . . , γp,τ1)T = βτ1 and γτt = (γ0,τt, γ1,τt, . . . , γp,τt)T = βτtβτt−1 for t = 2, . . . , q.

The constraint in (2) is equivalent to zTγτt ⩾ 0 for all x𝒟 and all t = 2, . . . , q. Break each γj,τt into its positive and negative parts, so that γj,τt=γj,τt+γj,τt, where both γj,τt+ and γj,τt are nonnegative, and only one may be nonzero. With this reparameterization, noncrossing can be easily ensured on 𝒟 = [0, 1]p. Using this parameterization, the constraint in (2) becomes simply that

γ0,τtj=1pγj,τt0(t=2,,q). (3)

This gives the point that is the worst-case scenario for each t, having xj = 1 when γj,τt < 0, and xj = 0 when γj,τt > 0. Since this point is in 𝒟, noncrossing must be enforced there, and hence (3) is a necessary condition. Since this point is the worst case, all other points in 𝒟 = [0, 1]p will then automatically satisfy the constraint for each t = 2, . . . , q, and hence (3) is also a sufficient condition.

After reparameterization, the problem is thus reduced to a linear programming problem, which can be solved via standard software. The linear program is extremely sparse and thus the use of a sparse matrix representation is more efficient. For computation, the linear programming has been implemented via use of the SparseM package (Koenker & Ng, 2003) in the R software platform (R Development Core Team, 2010), and the code is available from the first author.

2.2. Asymptotic properties

Consider a set of percentiles τ1 < … < τq, such that τt ∈ [ε, 1 − ε] for all t, with 0 < ε < 1/2. For the asymptotic properties, the set (τ1, . . . , τq) is allowed to change with n, if desired. In particular, one may wish to consider a denser set as the sample size increases. Assume that the linear quantile regression model holds with true parameter β0(τ), so that the τtth quantile of the conditional distribution for the response is given by zTβτt0. Specifically, FYi|x1(τt)=zTβτt0 for i = 1, . . . , n, where FYi|x denotes the conditional distribution function for observation i. Let β̃(τ) be the classical quantile regression estimator obtained by solving (1) separately for each τt, and let β̂(τ) be the constrained noncrossing version obtained via (2).

The following theorem shows that, regardless of the choice of a weight function, the estimator obtained via (2) is asymptotically equivalent to the typical quantile regression estimator. The following conditions are assumed.

  • Condition 1. The weights w(τt) > 0 for all t = 1, . . . , q.

  • Condition 2. The matrix n1i=1nxixiT is positive definite.

  • Condition 3. The conditional distributions have densities, fYi|x, that are differentiable with respect to Yi for every x and each i = 1, . . . , n.

  • Condition 4. There exist constants a > 0, b < ∞ and c < ∞ such that
    afYi|x{FYi|x1(τ)}b,|fYi|x(1){FYi|x1(τ)}|c,
    uniformly for x𝒟, ετ ⩽ 1 − ε, and i = 1, . . . , n, where fYi|x(1) denotes the derivative of fYi|x with respect to Yi.

The first condition ensures that the chosen weight function is appropriate to estimate the desired quantile curves. The remaining conditions allow a uniform Bahadur representation of the classical quantile regression estimator, as in Neocleous & Portnoy (2008), which ensures that β˜τβτ0=Op(n1/2) uniformly in ετ ⩽ 1 − ε.

Theorem 1. Let β̂(τ) and β̃(τ) be the constrained and unconstrained estimators, respectively, for the set of quantiles τ1 < … < τq, such that n1/2 mint (τt+1τt) → ∞. Then for any u ∈ 𝔕pq,

|pr[n1/2{β^(τ)β0(τ)}u]pr[n1/2{β˜(τ)β0(τ)}u]|0,

so that the constrained estimator has the same limiting distribution as the classical quantile regression estimator.

Based on Theorem 1, inference for the n1/2-consistent constrained quantile regression can be achieved by using the known asymptotic results for classical quantile regression. In particular, appropriate asymptotic standard errors can be computed via the quantile regression sandwich formula (Koenker, 2005, Sec. 3.2.3).

3. Extension to nonparametric quantile curves

3.1. Quantile curves

Consider the model with a single predictor x ∈ [0, 1]. Without assuming that the quantiles vary linearly with x, a nonparametric fit is often performed via quantile smoothing splines. When the quantiles are curves, the crossing problem becomes even more pronounced, as the curves are more likely to cross.

Taking the approach of quantile smoothing splines (Koenker et al., 1994), the constrained joint quantile smoothing spline estimate can be formulated as the set of functions ĝτ1, . . . , ĝτq𝒢 that minimizes

t=1qw(τt)i=1nρτt{yigτi(xi)}+t=1qλτtV(gτt)subjecttogτt(x)gτt1(x),x[0,1],t=2,,q, (4)

where V (g′) is the total variation of the derivative of the function g. For twice continuously differentiable g, we have V(g)=01|g(x)|dx. In general,

V(g)=supPi=0NP1|g(xi+1)g(xi)|dx, (5)

where the supremum is taken over the set of all possible partitions P of [0, 1], and NP denotes the number of endpoints that defines the partition P.

Following Pinkus (1988) and Koenker et al. (1994), consider the expanded second-order Sobolev space,

𝒢={g:g(x)=a0+a1x+01(xy)+dμ(y),V(μ)<,ai,i=0,1},

where μ is a measure with finite total variation. This space includes the usual Sobolev space of functions having second derivative with finite L1 norm and absolutely continuous first derivative, while also including the limiting piecewise linear functions. As discussed in Pinkus (1988), this expansion is necessary to ensure that the interpolating function that minimizes the total variation resides in the function space 𝒢. For piecewise linear functions, the supremum in (5) occurs when the partition coincides with the breakpoints of the function.

Theorem 2. The set of functions ĝτ1, . . . , ĝτq𝒢 minimizing (4) subject to the noncrossing constraint consists of noncrossing linear splines with knots at the data points.

This implies that it suffices to consider the problem in terms of a linear spline basis, for which the total variation is a linear function of the coefficients. Hence, as in Koenker et al. (1994), the smoothing spline problem is a linear programming problem. It will now be shown that the noncrossing quantile constraint can be directly incorporated into this framework while retaining the linear programming problem.

Let {Bj (x)} for j = 1, . . . , kn + 1 denote the linear B-spline basis with kn internal knots and endpoints at 0 and 1. Then g^τt(x)=β^0,τt+j=1kn+1β^j,τtBj(x). Using the analogous parameterization as in the previous section in terms of differences in the coefficients across quantile levels, i.e. γτ1 = βτ1 and γτt = βτtβτt−1 for t = 2, . . . , q, it follows that g^τt(x)g^τt1(x)=γ^0,τt+j=1kn+1γ^j,τtBj(x) for any x. Hence the differences across quantiles is simply written in terms of a linear B-spline basis.

Considering the differences across successive quantiles as a linear spline with knots at each data point, it follows that it is necessary and sufficient to enforce the nonnegative constraint at the knots. Nonnegativity at each knot will imply nonnegativity between the knots due to the linearity. The form of the linear B-spline basis allows for a convenient parameterization, since at each knot a single basis function takes the value 1 while the remaining are 0. Hence at xj, where xj is the value at knot j, the difference across the successive quantiles is given by ĝτt (xj) − ĝτt−1 (xj) = γ̂0,τt + γ̂j,τt. Hence, letting γj,τt=γj,τt+γj,τt be the coefficients parameterized as above, it is necessary and sufficient to improve γ0,τtmaxj(γj,τt)0 for each t = 2, . . . , q. This can be turned into a linear programming problem using the constraints γ0,τtγj,τt0 for each t = 2, . . . , q and j = 1, . . . , kn + 1.

Using the linear B-spline basis, the total variation penalty is simply a linear function of the basis coefficients. Hence, the full optimization problem given by (4) is again a linear programming problem, and computation can be done efficiently using the sparse matrix representation as before.

3.2. Tuning the procedure

Koenker et al. (1994) and He & Ng (1999) suggest the use of a Schwartz-type information criterion for choosing the regularization parameter in quantile smoothing splines. For each individual quantile curve, this criterion is

SIC(λτt)=log[n1i=1nρτt{yig^τtλτt(xi)}]+(2n)1pλτtlog(n), (6)

where gτtλτt denotes the estimated function for that choice of λτt, while pλτt is the number of interpolated data points, i.e. those with zero residuals, and serves as a natural measure of the complexity of the model. The full set of tuning parameters for the individual quantile curves minimizes the joint Schwartz-type criterion,

SICJ(λ)=t=1qw*(τt)log[n1i=1nρτt{yig^τtλτt(xi)}]+(2n)1log(n)t=1qpλτt, (7)

for any choice of weights w*(τt) > 0. The weights, w*(τt), may differ from those in the loss function due to the log scale. However, in the case of w(τt) = 1, as in the examples, it is natural to also choose w*(τt) = 1.

For the proposed joint noncrossing quantile smoothing spline, the joint Schwartz-type criterion (7) is used to select the tuning parameters. If the individual quantile smoothing splines chosen via the separate sic criteria (6) do not cross, then the joint quantile smoothing spline will be exactly this set.

One could instead choose the logarithm of the full objective function, i.e. taking the logarithm outside the double summation, instead of summing the individual pieces. However, the proposed criterion in (7) will guarantee that the individual quantile smoothing splines agree with the joint noncrossing smoothing spline when the former are noncrossing. The alternative criterion will not necessarily maintain this desirable property. In addition, simulation results have shown that (7) seems to exhibit better empirical performance.

In the joint quantile smoothing spline, there are separate tuning parameters for each quantile curve, resulting in a q-dimensional tuning parameter selection problem. However, experience, as shown in the simulations, has shown that a single tuning parameter performs sufficiently well in controlling the overall smoothness of the set of curves, while still allowing for varying degrees of smoothness among the curves themselves.

If a full q-dimensional tuning is desired, a directed search can proceed as follows.

Algorithm 1.

  • Step 1. Fit the joint smoothing spline with a single tuning parameter, i.e. set λτ1 = … = λτq and find the value that minimizes the sicJ criterion.

  • Step 2. Vary λτ1 on a grid around the current value and minimize the sicJ criterion keeping the remaining λτt values fixed at the current value. Take this as the new tuning parameter vector.

  • Step 3. Continue to update sequentially in this fashion, until no further improvement is possible.

While this algorithm is not guaranteed to find the global minimizer, it will at least converge to a local minimizer, starting from the optimal solution for the single tuning parameter case.

4. Simulation study

4.1. Linear quantile regression

The proposed method is now compared to classical quantile regression without the noncrossing assumption. Additionally, the method is compared to the approach of He (1997), which assumes a location-scale heteroscedastic error model. Each of the examples is a special case of the heteroscedastic error model as in He (1997):

yi=β0+βTxi+(γ0+γTxi)εi,xij~U(0,1),εi~N(0,1).

For all examples, the intercepts are set to β0 = γ0 = 1. In each example, six quantile curves, τ = 0.1, 0.3, 0.5, 0.7, 0.9, 0.99, are fitted to the data, either separately for the classical quantile regression approach or simultaneously for the proposed approach. The method of He (1997) estimates the location and scale parameters in the model directly, and then uses them to estimate the quantile curves. For each example, 500 datasets are simulated. The three examples are as follows.

  • Example 1. The sample size is n = 100, with p = 4 predictors, and parameters β = (1, 1, 1, 1)T, and γ = (0.1, 0.1, 0.1, 0.1)T.

  • Example 2. The sample size is n = 100, with p = 10 predictors, and parameters β = (1, 1, 1, 1, 0T)T, and γ = (0.1, 0.1, 0.1, 0.1, 0T)T.

  • Example 3. The sample size is varied with n = (100, 200, 500), with p = 7 predictors, and parameters β = (1, 1, 1, 1, 1, 1, 1)T and γ = (1, 1, 1, 0, 0, 0, 0)T.

To illustrate the crossing problem, for the first example, out of the 500 generated datasets, 491 of them had at least one crossing in the domain.

The results from the examples are given in Table 1. Presented are the empirical root mean integrated squared errors in estimation of the curves for each of τ = 0.5, 0.9, 0.99 given by RMISE=[n1i=1n{g^τ(xi)gτ(xi)}2]1/2, where ĝτ is the estimated function and gτ is the true function. In this case, the functions g and ĝ are linear, while in the next section they are nonlinear. The table presents the average root mean integrated squared errors over the 500 datasets along with their estimated standard errors. The results for the other quantiles are similar, and are thus omitted. However, by fitting simultaneously, it is guaranteed that they will not cross.

Table 1.

Average root mean integrated squared error (×100) over 500 simulated datasets, with standard error in parentheses

Example 1 Example 2
τ = 0.5 τ = 0.9 τ = 0.99 τ = 0.5 τ = 0.9 τ = 0.99
NCRQ 30.1 (0.44) 40.7 (0.59) 72.9 (0.88) 42.9 (0.43) 53.2 (0.52) 89.7 (0.84)
RQ 31.2 (0.46) 42.9 (0.65) 86.1 (0.96) 47.9 (0.45) 66.5 (0.64) 121.3 (0.95)
RRQ 31.2 (0.46) 48.6 (0.70) 92.7 (2.01) 47.9 (0.45) 76.3 (0.70) 147.3 (2.34)
Example 3, n = 100 Example 3, n = 200
τ = 0.5 τ = 0.9 τ = 0.99 τ = 0.5 τ = 0.9 τ = 0.99
NCRQ 75.9 (0.92) 99.8 (1.19) 179.7 (2.04) 56.4 (0.66) 74.6 (0.91) 132.3 (1.62)
RQ 82.1 (0.99) 116.7 (1.36) 226.5 (1.85) 60.0 (0.69) 82.1 (0.99) 162.2 (1.74)
RRQ 82.1 (0.99) 133.0 (1.52) 250.7 (4.72) 60.0 (0.69) 91.5 (1.04) 158.9 (2.50)
Example 3, n = 500
τ = 0.5 τ = 0.9 τ = 0.99
NCRQ 35.8 (0.41) 47.0 (0.55) 92.5 (1.14)
RQ 37.1 (0.42) 49.7 (0.57) 109.0 (1.28)
RRQ 37.1 (0.42) 56.7 (0.65) 96.0 (1.33)

NCRQ, proposed noncrossing regression quantiles; RQ, classical regression quantiles; RRQ, restricted regression quantiles of He (1997).

In each of the settings considered, the proposed constrained approach gives significantly better estimation for all quantiles. This is probably because the constraints add some smoothness and stability across the quantile curves. Since the true curves do not cross, it is expected that the constrained estimates would perform better. This improvement is greater in the tails, as there is less data, hence the smoothing effect of the constraint allows for the borrowing of strength. In addition, the improvement is much more pronounced for Example 2, due to the presence of the extra irrelevant predictors.

By varying the sample size in Example 3, as expected, the differences become smaller as the sample size grows, due to the consistency of the classical estimators. However, even with n = 500 some statistically significant improvement still remains, as seen by the differences in root mean integrated squared errors relative to their reported standard errors. The estimator of He (1997), although based on the heteroscedastic model, does not exhibit better performance than the typical estimator except in the extreme quantiles in the larger sample case. This phenomenon was also observed by Wu & Liu (2009), and probably occurs because it requires implicit nonparametric estimation of the quantiles of the residuals, which may be unstable for smaller sample sizes.

Examples with other set-ups, including covariates generated from Gaussian distributions, gave similar results, and are thus omitted.

4.2. Nonparametric quantile regression

For nonparametric quantile regression, the proposed method is compared with quantile smoothing splines and quantile regression splines. Quantile smoothing splines result in penalized linear splines, as in the proposed method, but with each curve fitted independently. Quantile regression splines start with linear splines and perform knot selection. Both methods are implemented in the R package COBS (Constrained B-Spline Smoothing, He & Ng, 1999; Ng & Maechler, 2007). This package also allows the user to specify qualitative constraints on each individual quantile curve, such as monotonicity or convexity. This constraint on the curves refers to constraining the quantile curve for a particular value of τ, not as a function across values of τ as would be needed to ensure noncrossing. To ease the computational burden, the COBS package, by default, implements the quantile smoothing splines with 25 knots instead of knots at each data point. The user may specify more or less knots, if desired. The same choice of a reduced set of knots is used for computational convenience for the proposed joint smoothing spline in the simulation study.

Each of the two examples is again a special case of the heteroscedastic error model,

yi=f(xi)+g(xi)εi,

for some functions f and g. The covariate is again generated as U (0, 1) and εiN (0, 1) with n = 100. The two examples are given as follows.

  • Example 4. The mean function is f(x) = 0.5 + 2x + sin(2πx − 0.5), and the variance function is g(x) = 1.

  • Example 5. The mean function is f(x) = 3x, and the variance function is g(x) = 0.5 + 2x + sin(2πx − 0.5).

Example 4 results in the quantile curves simply being a shift in the intercept as shown in Fig. 1(a). Example 5 results in the quantile curves having various degrees of smoothness, with the median being linear, and more curvature exhibited in the extreme quantile curves, as shown in Fig. 1(b).

Fig. 1.

Fig. 1

Plot of the true conditional quantile functions for Example 4 (a) and Example 5 (b). The five curves from bottom to top represent τ = 0.1, 0.3, 0.5, 0.7, 0.9.

For the nonparametric fits, the 0.99 quantile was not used, as it is too extreme for a nonparametric estimate based on a sample of size 100. Hence the set of quantiles, τ = 0.1, 0.3, 0.5, 0.7, 0.9 was fitted. Table 2 shows the results for τ = 0.5, 0.7, 0.9 for the two examples over the 500 simulated datasets for each example. The lower quantiles are analogous due to symmetry. Overall, the proposed method compares favourably to the traditional quantile splines, in terms of integrated squared error. For comparison, the proposed method is computed using a single tuning parameter as well as using a separate tuning parameter for each quantile level. In both examples, the single tuning parameter exhibits a performance very similar to the full q-dimensional tuning.

Table 2.

Average root mean integrated squared error (×100) over 500 simulated datasets, with standard error in parentheses

Example 4 Example 5
τ = 0.5 τ = 0.7 τ = 0.9 τ = 0.5 τ = 0.7 τ = 0.9
NCRQ 25.7 (0.21) 25.9 (0.21) 31.8 (0.31) 26.4 (0.33) 32.2 (0.34) 48.5 (0.60)
NCRQ (single) 24.6 (0.19) 25.3 (0.19) 32.3 (0.34) 26.6 (0.34) 31.7 (0.32) 49.9 (0.62)
RQ (RS) 29.8 (0.18) 30.6 (0.21) 35.2 (0.31) 24.7 (0.42) 36.3 (0.43) 52.2 (0.73)
RQ (SS) 27.1 (0.19) 27.6 (0.21) 34.7 (0.31) 21.8 (0.35) 34.2 (0.39) 53.4 (0.90)

NCRQ, proposed noncrossing regression quantiles; NCRQ (single), proposed approach with a single tuning parameter; RQ (RS), classical regression splines with knot selection; RQ (SS), classical regression smoothing splines via regularization.

4.3. Varying the choice of quantiles

Assume that the linear quantile regression model holds for a given quantile of interest. Since the proposed noncrossing approach is based on simultaneous estimation of a set of quantile curves, the estimate for the quantile of interest will change depending on the included quantiles. For example, if interest focuses on the median, one can add any number of additional quantiles to the median. Based on the results of Theorem 1, the asymptotic distribution will not be affected by the number of quantiles added. However, adding additional quantiles can improve the finite sample results via adding stability to the estimation.

Figure 2 plots mean squared error in estimating the slope at the median for a univariate model as a function of the number of included quantiles for each of sample sizes n = 50, 100 and 200. The median regression estimate was computed on 5000 datasets based on an increasingly dense sequence of equally spaced quantiles. The y-axis represents the ratio of mean squared error to that of using only the median. The use of more quantiles seems to help stabilize the estimation.

Fig. 2.

Fig. 2

Plot of mean squared error in estimation of the slope at the median as a function of the number of included quantiles, for sample sizes n = 50 (solid line), n = 100 (dashed line), and n = 200 (dotted line). Each curve is scaled so that the mean squared error is reported as a ratio relative to that of using only median regression.

It is natural to assume that if all quantiles are linear, adding more would give better results. Figure 2 was generated via the model in Example 5, so that only the median is linear. However, even with the misspecification, there is a gain in accuracy, probably because quantiles near the median are still close to linear, and help stabilize the estimation. This phenomenon was observed in other scenarios, including those focused on extreme quantiles. In practice, we recommend adding quantiles in a neighbourhood of the quantile of interest until the estimation appears to stabilize.

5. Analysis of hurricane data

The noncrossing quantile regression approach is now applied to the tropical cyclone data. The data consist of a sample of 422 tropical cyclones occurring near the US coastline over the period 1899–2006. Jagger & Elsner (2009) used linear quantile regression to model the maximum wind speed of each cyclone as a function of four climate covariates: the North Atlantic Oscillation Index, the Southern Oscillation Index, the Atlantic sea-surface temperature and the average sunspot number. The climate covariates are constant within a single year to represent the yearly large-scale climate conditions. The North Atlantic Oscillation Index is the preseason and early-season average of the May and June values. The other three are obtained by averaging over the peak season of August through October. The particular focus is on the upper quantiles, as these extreme hurricane-strength storms are of considerable importance.

Following Jagger & Elsner (2009), quantile regression is applied to these data at the quantiles 0.25, 0.5, 0.75, 0.9, 0.95, 0.99. Table 3 shows the parameter estimates for the intercept along with the four slope parameters for the three upper quantiles from both a classical quantile regression fit and the proposed noncrossing fit. As in Jagger & Elsner (2009), included are pointwise 90% confidence intervals. In the other three quantiles, the results for both methods are similar.

Table 3.

Coefficient estimates for the intercept and the four climate covariates for the hurricane data at upper quantiles, with 90% confidence intervals

Unconstrained Constrained
τ = 0.9 τ = 0.95 τ = 0.99 τ = 0.9 τ = 0.95 τ = 0.99
INT 109.15* (85.01, 133.29) 120.06* (88.26, 151.86) 134.65* (119.41, 149.89) 107.83* (84.56, 131.11) 117.95* (91.16, 144.74) 140.63* (121.44, 159.83)
NAO −5.03* (−9.91, −0.14) −0.95 (−6.38, 4.49) 1.43 (−1.58, 4.44) −4.93* (−9.61, −0.26) −3.06 (−8.04, 1.93) −3.06 (−6.54, 0.43)
SOI 5.74* (0.40, 11.09) 1.23 (−5.75, 8.21) −3.98* (−7.44, −0.51) 5.21* (0.02, 10.39) 3.32 (−2.62, 9.27) −2.80 (−7.10, 1.50)
SST 6.16* (1.34, 10.97) 4.19 (−1.65, 10.03) −0.73 (−3.52, 2.06) 5.17* (0.57, 9.77) 5.17* (0.12, 10.22) 3.21 (−0.48, 6.90)
SUN 3.48 (−1.21, 8.17) 1.95 (−2.60, 6.50) 6.19* (2.86, 9.52) 3.16 (−1.43, 7.75) 3.16 (−1.17, 7.49) 3.16 (−0.52, 6.84)

INT, intercept; NAO, North Atlantic Oscillation Index; SOI, Southern Oscillation Index; SST, Atlantic sea surface temperature; SUN, average sunspot number;

*,

statistically significant coefficients at α = 0.1.

Of particular note is the smoothing of the coefficients in the extreme quantiles. This smoothing stabilizes the inference and avoids some of the possibly spurious associations, such as the significance of the sunspot number in the 99th percentile but not in any of the other upper quantiles. The confidence intervals for both methods were obtained using the asymptotic normality and using the kernel method to estimate the inverse density needed for the standard errors (Powell, 1991; Koenker, 2005).

As the estimates will change depending on the number and location of included quantiles, a sensitivity analysis is performed. The coefficient estimates for the median and 0.99 quantile regressions are examined as a function of an increasing number of quantiles. Figure 3 plots the estimated coefficients for each of the four slopes for the median (a) and the 0.99 quantile (b). Initially, only the two quantiles were fitted. Then quantiles were sequentially added until a grid spacing of 0.05 was obtained, with a grid spacing of 0.01 bracketing the median, from 0.45 to 0.55, and a spacing of 0.01 from 0.9 to 0.99, to ensure a more saturated region around the quantiles of interest. The median is much less sensitive, while the 0.99 quantile is clearly highly sensitive. Once 14 quantiles are included, the inference regarding significant predictors remains unchanged upon adding more quantiles, and is the same as that reported in Table 3. This is at the point in the sensitivity analysis that the 0.95 quantile is first added between the 0.9 and the 0.99 quantiles.

Fig. 3.

Fig. 3

Plot of the estimated slope coefficients at the median (a) and the 0.99 quantile (b) as the number of included quantiles is increased. NAO, North Atlantic Oscillation Index (solid line); SOI, Southern Oscillation Index (dashed line); SST, Atlantic sea surface temperature (dotted line); SUN, average sunspot number (dotted/dashed line).

Acknowledgments

The authors are grateful to the editor, an associate editor and two anonymous referees for their valuable comments. This research was sponsored by the National Science Foundation, U.S.A. and the National Institutes of Health, U.S.A.

Appendix

Proof of Theorem 1

Let n and n denote n1/2{β̂(τ) − β0(τ)} and n1/2{β̃(τ) − β0(τ)}, respectively. Then |pr(nu) − pr(nu)| = |pr(nu|nn) − pr(nu|nn)| pr(nn).

Since the first term in the product is bounded by 1, it suffices to show that pr(nn) → 0, or pr(n = n) → 1.

Due to the formulation of the estimator, the event n = n is equivalent to the event that the classical quantile regression estimator maintains its appropriate ordering. To show that the probability of this event goes to 1, consider the difference in the classical estimator at successive quantiles, n1/2(zTβ̃τt+1zTβ̃τt). It will be shown that this difference must be positive with probability tending to 1 for every t = 1, . . . , q.

The difference can be written as

n1/2(zTβ˜τt+1zTβτt+10)n1/2(zTβ˜τtzTβτt0)+n1/2(zTβτt+10zTβτt0).

Under the assumed regularity conditions, the classical quantile regression estimator, β̃(τ), is n1/2-consistent. Hence, it follows that the first two terms are Op(1) for any t.

By the mean value theorem, it follows that zTβτt+10zTβτt0=(τt+1τt)τzTβτ0|τ=τ*, where τtτ*τt+1. Now, regularity condition (3) yields that τzTβτ0={fYi|x(FYi|x1(τ))}11/b for any τ ∈ (0, 1). Hence n1/2(zTβτt+10zTβτt0)n1/2(τt+1τt)/b. By assumption, the right-hand side diverges. This leads to the third term dominating in the difference with probability tending to 1, and thus the difference will be positive.

Proof of Theorem 2

Assume that τ1 , . . . , τq𝒢 is the minimizing set of functions. Now consider any particular τt. The value of the loss function, the first term in (4), only depends on the values of gτt at the data points. Hence any other gτt such that gτt (x) = τt (x) at the data points will yield the same value for the loss function. So it suffices to consider the problem of finding the interpolator of the set of points {τt (xi)}, which minimizes the total variation. The solution to this interpolation problem is given by a linear spline with knots at the given set of xi (Fisher & Jerome, 1975; Pinkus, 1988). Denote this solution by ĝ. Clearly, since the set of are noncrossing, they maintain the proper ordering at each of the data points, so the interpolating splines ĝ will not cross.

References

  1. Chernozhukov V, Fernandez-Val I, Galichon A. Improving point and interval estimators of monotone functions by rearrangement. Biometrika. 2009;96:559–75. [Google Scholar]
  2. Cole TJ. Fitting smoothed centile curves to reference data (with Discussion) J. R. Statist. Soc. A. 1988;151:385–418. [Google Scholar]
  3. Cole TJ, Green PJ. Smoothing reference centile curves: the LMS method and penalized likelihood. Statist Med. 1992;11:1305–19. doi: 10.1002/sim.4780111005. [DOI] [PubMed] [Google Scholar]
  4. Dette H, Volgushev S. Non-crossing non-parametric estimates of quantile curves. J. R. Statist. Soc. B. 2008;70:609–27. [Google Scholar]
  5. Fisher SD, Jerome JW. Spline solutions to L1 external problems in one and several variables. J. Approx. Theory. 1975;13:73–83. [Google Scholar]
  6. Hall P, Wolff RCL, Yao Q. Methods for estimating a conditional distribution function. J Am Statist Assoc. 1999;94:154–63. [Google Scholar]
  7. He X. Quantile curves without crossing. Am. Statistician. 1997;51:186–92. [Google Scholar]
  8. He X, Ng P. COBS: qualitatively constrained smoothing via linear programming. Comp Statist. 1999;14:315–37. [Google Scholar]
  9. Jagger TH, Elsner JB. Modeling tropical cyclone intensity with quantile regression. Int J Climatol. 2009;29:1351–61. [Google Scholar]
  10. Koenker R. A note on L-estimators for linear models. Statist Prob Lett. 1984;2:323–5. [Google Scholar]
  11. Koenker R. Quantile Regression. Cambridge: Cambridge University Press; 2005. [Google Scholar]
  12. Koenker R, Ng P. SparseM: a sparse matrix package for R. J Statist Software. 2003;8 http://www.jstatsoft.org/v08/i06. [Google Scholar]
  13. Koenker R, Ng P, Portnoy S. Quantile smoothing splines. Biometrika. 1994;81:673–80. [Google Scholar]
  14. Neocleous T, Portnoy S. On monotonicity of regression quantile functions. Statist Prob Lett. 2008;78:1226–9. [Google Scholar]
  15. Ng P, Maechler M. A fast and efficient implementation of qualitatively constrained quantile smoothing splines. Statist Mod. 2007;7:315–28. [Google Scholar]
  16. Pinkus A. On smoothest interpolants. SIAM J Math Anal. 1988;19:1431–41. [Google Scholar]
  17. Powell JL. Estimation of monotonic regression models under quantile restrictions. In: Barnett W, Powell J, Tauchen G, editors. Nonparametric and Semiparametric Methods in Econometrics. Cambridge: Cambridge University Press; 1991. [Google Scholar]
  18. R Development Core Team . R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing; 2010. URL: http://www.R-project.org. [Google Scholar]
  19. Wu Y, Liu Y. Stepwise multiple quantile regression estimation using non-crossing constraints. Statist. Interface. 2009;2:299–310. doi: 10.1080/10485252.2010.537336. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Biometrika are provided here courtesy of Oxford University Press

RESOURCES