Abstract
Often of primary interest in the analysis of multivariate data are the copula parameters describing the dependence among the variables, rather than the univariate marginal distributions. Since the ranks of a multivariate dataset are invariant to changes in the univariate marginal distributions, rank-based estimators are natural candidates for semiparametric copula estimation. Asymptotic information bounds for such estimators can be obtained from an asymptotic analysis of the rank likelihood, i.e. the probability of the multivariate ranks. In this article, we obtain limiting normal distributions of the rank likelihood for Gaussian copula models. Our results cover models with structured correlation matrices, such as exchangeable or circular correlation models, as well as unstructured correlation matrices. For all Gaussian copula models, the limiting distribution of the rank likelihood ratio is shown to be equal to that of a parametric likelihood ratio for an appropriately chosen multivariate normal model. This implies that the semiparametric information bounds for rank-based estimators are the same as the information bounds for estimators based on the full data, and that the multivariate normal distributions are least favorable.
Keywords: copula model, local asymptotic normality, multivariate rank statistics, marginal likelihood, rank likelihood, transformation model
1 Rank likelihood for copula models
Recall that a copula is a multivariate CDF having uniform univariate marginal distributions. For any multivariate CDF F(y1, . . . , yp) with absolutely continuous margins F1, . . . , Fp, the corresponding copula C(u1, . . . , up) is given by
Sklar's theorem [Sklar, 1959] shows that C is the unique copula for which F(y1, . . . , yp) = C(F1(y1), . . . , Fp(yp)).
In this article we consider models consisting of multivariate probability distributions for which the copula is parameterized separately from the univariate marginal distributions. Specifically, the models we consider consist of collections of multivariate CDFs such that ψ parameterizes the univariate marginal distributions and θ parameterizes the copula, meaning that for a random vector Y = (Y1, . . . , Yp)T with CDF F(y|θ, ψ),
We refer to such a class of distributions as a copula-parameterized model. For such a model, it will be convenient to refer to the class of copulas {C(u|θ) : θ ∈ Θ} as the copula model, and the class {F1(y|ψ), . . . , Fp(y|ψ ∈ Ψ)} as the marginal model.
As an example, the copula model for the class of p-variate multivariate normal distributions is called the Gaussian copula model, and is parameterized by letting Θ be the set of p × p correlation matrices. The marginal model for the p-variate normal distributions is the set of all p-tuples of univariate normal distributions. The copula-parameterized models we focus on in this article are semiparametric Gaussian copula models [Klaassen and Wellner, 1997], for which the copula model is Gaussian and the marginal model consists of the set of all p-tuples of absolutely continuous univariate CDFs.
Let Y be an n × p random matrix whose rows Y1, . . . , Yn are i.i.d. samples from a p-variate population. We define the multivariate rank function so that Ri,j, the (i, j)th element of R(Y), is the rank of Yi,j among {Y1,j, . . . , Yn,j}. Note that the ranks R(Y) are invariant to strictly increasing transformations of the columns of Y, and therefore the probability distribution of R(Y) does not depend on the univariate marginal distributions of the p variables. As a result, for any copula parameterized model and data matrix with ranks R(y) = r, the likelihood L(θ, ψ : y) can be decomposed as
| (1) |
where p(y|θ, ψ) is the joint density of Y and p(y|θ, ψ, r) is the conditional density of Y given R(Y) = r. The function L(θ : r) = Pr(R(Y) = r|θ) is called the rank likelihood function. In situations where θ is the parameter of interest and ψ a nuisance parameter, inference for θ can be obtained from the rank likelihood function without having to estimate the margins or specify a marginal model. A univariate rank likelihood function was proposed by Pettitt [1982] for estimation in monotonically transformed regression models. Asymptotic properties of the rank likelihood for this regression model were studied by Bickel and Ritov [1997], and a parameter estimation scheme based on Gibbs sampling was provided in Hoff [2008]. Rank likelihood estimation of copula parameters was studied in Hoff [2007], who also extended the rank likelihood to accommodate multivariate data with mixed continuous and discrete marginal distributions.
The rank likelihood is constructed from the marginal probability of the ranks and can therefore be viewed as a type of marginal likelihood. Marginal likelihood procedures are often used for estimation in the presence of nuisance parameters (see Section 8.3 of Severini [2000] for a review). Ideally, the statistic that generates a marginal likelihood is “partially sufficient” in the sense that it contains all of the information about the parameter of interest that can be quantified without specifying the nuisance parameter. Notions of partial sufficiency include G-sufficiency [Barnard, 1963] and L-sufficiency [Rémon, 1984], which are motivated by group invariance and profile likelihood, respectively. Hoff [2007] showed that the ranks R(Y) are both a G- and L-sufficient statistic in the context of copula estimation.
Although rank-based estimators of the copula parameter θ may be appealing for the reasons described above, one may wonder to what extent they are efficient. The decomposition given in (1) indicates that rank-based estimates do not use any information about θ contained in L(θ, ψ : [y|r]), i.e. the conditional density of the data given the ranks. For at least one copula model, this information is asymptotically negligible: Klaassen and Wellner [1997] showed that for the bivariate normal copula model, a rank-based estimator is semiparametrically efficient and has asymptotic variance equal to the Cramér-Rao information bound in the bivariate normal model, i.e. the bivariate normal model is the least favorable submodel. Genest and Werker [2002] studied the efficiency properties of pseudo-likelihood estimators for two-dimensional semiparametric copula models and showed that the pseudo-likelihood estimators (which are functions of the bivariate ranks) are not in general semiparametrically efficient for non-Gaussian copulas. Chen et al. [2006] proposed estimators in general multivariate copula models that achieve semiparametric asymptotic efficiency but are not based solely on the multivariate ranks. It remains unclear whether estimators based solely on the ranks can be asymptotically efficient in general semiparametric copula models. In particular, it is not yet known if maximum likelihood estimators based on rank likelihoods for Gaussian semiparametric copula models are semiparametrically efficient.
The potential efficiency loss of rank-based estimators can be investigated via the limiting distribution of an appropriately scaled rank likelihood ratio. Generally speaking, the local asymptotic normality (LAN) of a likelihood ratio plays an important role in the asymptotic analysis of testing and estimation procedures. For semiparametric models, the asymptotic variance of a LAN likelihood ratio can be related to efficient tests [Choi et al., 1996] and information bounds for regular estimators [Begun et al., 1983, Bickel et al., 1993]. In particular, the variance of the limiting normal distribution of a LAN rank likelihood ratio provides information bounds for locally regular rank-based estimators of copula parameters.
In this article we obtain the limiting normal distributions of the rank likelihood ratio for Gaussian copula models with structured and unstructured correlation matrices. In the next section we give sufficient conditions under which the rank likelihood is LAN. The basic result is that the rank likelihood is LAN if there exists a good rank-measurable approximation to a LAN submodel. For Gaussian copulas, the natural candidate submodels are multivariate normal models, for which the log likelihood is quadratic in the observations. In Section 3, we identify sufficient conditions for a normal quadratic form to have a good rank-measurable approximation. This result allows us to identify multivariate normal submodels with likelihood ratios that asymptotically approximate the rank likelihood ratio. In Section 4 we show that for any smoothly parameterized Gaussian copula, the rank likelihood ratio is LAN with an asymptotic variance equal to that of the likelihood ratio for the corresponding multivariate normal model with unequal marginal variances. Since the parametric multivariate normal model is a submodel of the semiparametric Gaussian copula model, and in general the semiparametric information bound based on the full data is higher than that of any parametric submodel, our results imply that the bounds for rank-based estimators are equal to the semiparametric bounds for estimators based on the full data, and that the multivariate normal models are least favorable. These bounds can be compared to the asymptotic variance of an estimator to assess its asymptotic efficiency. Via two examples, in Section 5 we show that pseudo-likelihood estimators are asymptotically efficient for some but not all Gaussian copula models. This is discussed further in Section 6.
2 Approximating the rank likelihood ratio
The local log rank likelihood ratio is defined as
where L(θ : r) is defined in (1). Studying λr is difficult because L(θ : r) is the integral of a copula density over a complicated set defined by multivariate order constraints. However, in some cases it is possible to obtain the asymptotic distribution of λr by relating it to the local log likelihood ratio λy of an appropriate parametric multivariate model, where
| (2) |
This method of identifying the asymptotic distribution of λr is analogous to the approach taken by by Bickel and Ritov [1997] in their investigation of the rank likelihood ratio for a univariate semiparametric regression model.
In this section, we will show that if we can find a sufficiently good rank-measurable approximation to λy, then the limiting distribution of λr will match that of λy. Specifically, we prove the following theorem:
Theorem 2.1. Let{F(y|θ, ψ) : θ ∈ Θ, ψ ∈ Ψ} be an absolutely continuous copula parameterized model where for given values of θ and s there exists values of ψ and t such that under i.i.d. sampling from F(y|θ, ψ),
λy(s, t) is LAN, so that , a normal random variable, and
there exists a rank-measurable approximation λŷ(s, t) such that .
Then as n → ∞ under i.i.d. sampling from any population with copula C(u|θ) equal to that of F(y|θ, ψ) and arbitrary absolutely continuous marginal distributions.
Proof. Let L(θ, ψ : y) be the (parametric) likelihood function for a given dataset . The lack of dependence of the rank likelihood on the marginal distributions leads to the following identity relating λr(s) to λy(s, t):
Now suppose we would like to describe the statistical properties of λr(s) when the matrix r is replaced by the ranks R(Y), where the rows of Y are i.i.d. samples from a population with copula C(u|θ). Since the distribution of the ranks of Y is invariant with respect to the univariate marginal distributions, the particular marginal model and values of ψ and t are immaterial and can be chosen to facilitate analysis. For each θ and s, our strategy will be to select ψ and t such that the replacement of y by a rank-measurable approximation ŷ in Equation 2 results in an accurate rank-based approximation λŷ(s, t) of λy(s, t). Because the resulting λŷ is rank-measurable, we can write
If the approximation of λy(s, t) by λŷ(s, t) is sufficiently accurate to make the remainder term, log Eθ[eλy(s,t)–λŷ(s,t)|R(Y)], converge in probability to zero as n → ∞, then the asymptotic distribution of λr(s) is determined by that of λŷ(s, t). Note that λr(s) does not depend on t, which implies that the value of t for which such an approximation is available will depend on s and θ.
Let λy be LAN and Y1, . . . , Yn ~ i.i.d. F(y|θ, ψ). For given s and t, we will show that if , then log , where here and in what follows, limits are as n → ∞ and probabilities and expectations are calculated under θ and ψ unless otherwise noted. We note that this result was essentially proven at the end of the proof of Theorem 1 of Bickel and Ritov [1997] in the context of the regression transformation model, although details were omitted. We include the proof here for completeness.
Let Un = eλy, Vn = eλŷ and Rn = R(Y1 , . . . , Yn), so that the exponential of the remainder term can be written as . For any M > 1 we can write
We now show that each of an, bn and cn converge in probability to zero. To do so, we make use of the following facts:
by the continuous mapping theorem;
Un = eλy and are bounded in probability, as λy and λŷ converge in distribution.
is uniformly integrable, since log Un = λy is LAN [Hall and Loynes, 1977];
If E[|Xn|] → 0 and Zn is a random sequence, then .
To see that and , note that both and are bounded random variables that converge in probability to zero, so their conditional expectations given Rn converge in probability to zero as well. For the sequence bn, note that Un is Op(1) as it converges in distribution, and is op(1) as , so is op(1). Now 0 ≤ Ũn ≤ Un for each n, and is uniformly integrable, so is uniformly integrable as well. This and imply that E[|Ũn|] = E[Ũn] → 0, and so . Since , and is Op(1), bn is op(1).
Recall our original identity relating λr(s) to λy(s, t) and λŷ(s, t):
We have shown that if λy is LAN and under i.i.d. sampling from F(y|θ, ψ), then the remainder term goes to zero, and so λy, λŷ and λr all converge to the same normal random variable. If the data are being sampled from a population with the same copula as F(y|θ, ψ) but different margins, then there exists a transformation of the data such that F(y|θ, ψ) is the distribution of the transformed population, and the result follows.
For a given copula model, Theorem 2.1 essentially says that the asymptotic distribution of the log rank likelihood ratio will be the same as that of the log likelihood ratio of any multivariate model with the same copula, as long as the latter admits an asymptotically accurate rank-measurable approximation. The task of identifying the limiting distribution of λr then becomes one of identifying a suitable marginal model for which such an approximation to the log likelihood ratio holds. For multivariate normal models, the log likelihood ratio is quadratic in the observations, and so the existence of a good rank measurable approximation depends on the accuracy of rank-based approximations to normal quadratic forms. In the next section, we identify a class of quadratic forms that admit sufficiently accurate rank-measurable approximations. In Section 4, we relate these forms to multivariate normal models for which the conditions of Theorem 2.1 hold.
3 Rank approximations to normal quadratic forms
Let Y1, . . . , Yn be i.i.d. random column vectors from a member of a class of mean-zero p-variate normal distributions indexed by a correlation parameter θ ∈ Θ and a variance parameter ψ ∈ Ψ. As discussed further in the next section, the local likelihood ratio λy can be expressed as a quadratic function of Y1, . . . , Yn, taking the form
for some matrix A which could be a function of s, t, θ and ψ. A natural rank-based approximation to λy is
where {Ŷi,j : i ∈ {1, . . . , n}, j ∈ {1, . . . , p}} are the (approximate) normal scores, defined by R = R(Y) and . Whether or not λŷ – λy → 0 therefore depends on the convergence to zero of the difference between the quadratic terms of λŷ and λy. In this section we show that this difference converges to zero under certain conditions on A and the covariance matrix C = Cov[Yi]. Specifically, we prove the following theorem:
Theorem 3.1. Let Y1, . . . , Yn ~ i.i.d. Np(0, C) where C is a correlation matrix, and let , where Ri,j is the rank of Yi,j among Y1,j, . . . , Yn,j. Let A be a matrix such that the diagonal entries of AC + ATC are zero. Then
Proof. Let and let à = (A + AT)/2, so that yTÃy = yTAy for all . Then
the latter equality holding since à is symmetric. From this, we can write Sn = Qn + 2Ln where
We can write Qn as
The squared terms converge in probability to zero by Theorem 1 of de Wet and Venter [1972], and the cross term converges in probability to zero by the Cauchy-Schwarz inequality.
We now find conditions on A under which . Note that
where ã1 . . . , ãp are the rows of Ã. This gives
Let cj be the jth row of C, the correlation matrix of Y. We will show that if using an argument based on conditional expectations. Considering Ln,1 for example, recall that E[Y|Y1] = c1Y1 and so
if . The conditional expectation of is given by
The expectations in the second sum are both proportional to , leaving
The conditional expectation can be obtained by noting that if Y ~ Np(0, C), then the conditional distribution of Y given Y1 can be expressed as
where and ε is p-variate standard normal. The desired second moment is then
which is equal to under the condition that . Letting , the conditional variance of Ln,1 given the observations for the first variate is then
Applying Chebyshev's inequality gives
Now as a result of Theorem 1 of de Wet and Venter [1972] and therefore so does c̃n. But as c̃n is bounded, we have E[c̃n] → 0, giving
and so . The same argument can be applied to Ln,j for each j, and so as long as for each j = 1, . . . , p, or equivalently, if the diagonal elements of AC + ATC are zero.
4 LAN for general Gaussian copulas
In this section we use Theorems 2.1 and 3.1 to prove that the limiting distribution of the rank likelihood ratio λr for smoothly parameterized Gaussian copula models is same as that of the likelihood ratio for the corresponding normal model with unequal marginal variances. Specifically, we prove the following theorem:
Theorem 4.1. Let be a collection of positive definite correlation matrices such that C(θ) is twice differentiable. If Y1, . . . , Yn are i.i.d. from a population with absolutely continuous marginal distributions and copula C(θ) for some θ ∈ Θ, then the distribution of the rank likelihood ratio λr(s) converges to a N(–sT Iθθ·ψs/2, sT Iθθ·ψs) distribution, where Iθθ·ψ is the information for θ in the normal model with correlation C(θ) and marginal precisions ψ.
We note that Iθθ·ψ is a function of θ and not of ψ, as will become clear in the proof.
Proof. Consider the class of mean-zero multivariate normal models with inverse-covariance matrix Var[Y|θ, ψ]–1 = D(ψ)1/2B(θ)D(ψ)1/2, where and D(ψ) is the diagonal matrix with diagonal elements . The log probability density for a member of this class is given by
The log-likelihood derivatives are
and straightforward calculations show that
where “○” is the Hadamard product denoting element-wise multiplication. The local log likelihood ratio for this model can be expressed as
which, under independent sampling from Np(0, D(ψ)1/2C(θ)D(ψ)1/2), converges in distribution to a N(–uT Iu/2, uT Iu) random variable, where uT = (sT, tT) and I is the information matrix for (θ, ψ).
We take our rank based approximation λŷ to be equal to λy absent the op(1) term and with each Yi replaced by its approximate normal scores Ŷi. Clearly, we have λŷ(s) – λy(s) = op(1) if
Given θ and s, we now identify a value of t for which the above asymptotic result holds. Let t = Hs, where , so that
where {hk, k = 1, . . . q} are the columns of H. Now l̇θk(y) and l̇ψ(y) are both quadratic in y. Evaluating at ψ = 1, we have l̇θk(y) = [tr(BθkC) – yTBθky]/2 and , and so
Therefore, we can write sT[l̇θ(y) + HTl̇ψ(y)] as
where c(s, H, θ) does not depend on y, and Ak is given by
Substituting this representation of sTl̇θ + tTl̇ψ into λŷ and λy gives
Theorem 3.1 implies that this difference will converge in probability to zero if the diagonal elements of are zero for each k = 1, . . . , q. The value of can be calculated as
The vector diag(D(hk) + BD(hk)C) can be written as
and so our condition on hk becomes
Therefore, setting yields a quadratic form that satisfies the conditions of Theorem 3.1. The result then follows via Theorem 2.1. The value of uT Iu that determines the asymptotic mean and variance of λy(s), λŷ(s) and λr(s) is given by
This result shows that the least favorable submodel of a semiparametric Gaussian copula model is the multivariate normal model with unequal variances, and that the information bound for any regular estimator of θ is given by Iθθψ. However, for some correlation models the value of Iθθψ is equal to the corresponding information for θ in a model with equal marginal variances. In such cases, the least favorable submodel simplifies to the multivariate normal model with equal marginal variances. To identify conditions under which this result holds, consider the log likelihood ratio for a multivariate normal model with equal marginal variances:
Under i.i.d. sampling from Np(0, C(θ)/ψ), λy(s, t) converges in distribution to a N(–uT Iu/2,uT Iu) random variable, where uT = (sT, t) and I is the information matrix for (θ, ψ), for which
Our candidate rank-measurable approximation to λy(s, t) is given by
Recall that if for our given s and θ we can find a t and ψ such that λŷ – λy = op(1), then the conditions of Theorem 2.1 will be met and the asymptotic distribution of λr(s) will be that of λy(s, t). With this in mind, let t = hT s for some , and writeλy(s,hT s) ≡ λy(s). We will find conditions on C(θ) such that there exists an h for which λŷ(s) – λy(s) = op(1), and will show that any such h must be equal to . With t = hT s and ψ = 1, we have
where Ak = –(Bθk + hkB)/2 = (BCθk B – hkB)/2 and c(θ, s, h) does not depend on y. The difference between λŷ and λy is then
Since Ak is symmetric, Theorem 3.1 implies that this difference will converge in probability to zero if the diagonal elements of AkC are zero for each k = 1, . . . , q. This condition can equivalently be written as follows:
The above condition can only be met if, for each k, the diagonal elements of BCθk all take on a common value. If they do, then the convergence in probability of λŷ(s, t) – λy(s, t) to zero can be obtained by setting t = hT s, where hk = tr(BCθk)/p.
Setting ψ = 1, we have , and so setting results in λy, λŷ and λr each converging in distribution to a N(–sT Iθθ·ψs/2, sT Iθθ·ψ s) random variable, where is the information for θ in this parametric model. We summarize this result in the following corollary:
Corollary 4.2. Let be a collection of positive definite correlation matrices such that C(θ) is twice differentiable, and for each k, the diagonal entries of BCθk are equal to some common value. If Y1, . . . , Yn are i.i.d. from a population with absolutely continuous marginal distributions and copula C(θ) for some θ ∈ Θ, then the distribution of the rank likelihood ratio λr(s) converges to a N(–sT Iθθ·ψs/2, s Iθθ·ψs) distribution, where Iθθ·ψ is the information for θ in the normal model with correlation C(θ) and equal marginal precisions ψ.
5 Asymptotic efficiency in some simple examples
Obtaining the maximum likelihood estimator of a copula parameter θ from the rank likelihood is problematic due to the complicated nature of the likelihood. An easy-to-compute alternative estimator is the maximizer in θ of the pseudo-likelihood, which is essentially the probability of the observed data with the unknown marginal CDFs replaced with empirical estimates. Genest et al. [1995] studied the asymptotic properties of this pseudo-likelihood estimator (PLE) and obtained a formula for its asymptotic variance.
For Gaussian copula models, we can compare this asymptotic variance to the information bound obtained from Theorem 4.1 to evaluate the asymptotic efficiency of the PLE. This is most easily done in the case of a one-parameter copula model for which the conditions of Corollary 4.2 hold, as in this case the least favorable submodel is a simple two-parameter multivariate normal model with equal marginal variances. For such models, the value of Iθθ·ψ can be computed from the variance of the efficient influence function ľθ(y):
where l̃ψ(y) is the efficient influence function for ψ, given by (see, for example, Bickel et al. [1993], Chapter 2). This can be compared to the influence function for the PLE, which is given by
where the likelihood derivative and information matrix are based on the multivariate normal likelihood, and Wj(yj) is defined as
By inspection, the two influence functions are equal if , in which case the PLE is asymptotically efficient. To compute Wj(yj) for j = 1, . . . , p, note that for a Gaussian copula model, we have
where y = (Φ–1(u1), . . . , Φ–1(up)), C is the correlation matrix under θ and B = C–1. Straightforward calculations [Shorack, 2000, page 116] give
where D(y ○ y) is the diagonal matrix with elements , and the last line follows from the fact that BθC = –BCθ. Recall that for the models we are considering here, the diagonal elements of BCθ are assumed to all be equal, and so we can write
On the other hand, Iθψ = –tr(BCθ )/2, and so our condition for asymptotic efficiency becomes
| (3) |
We emphasize that this criterion for asymptotic efficiency only applies to one-parameter Gaussian copula models for which the conditions of Corollary 4.2 hold. Such models include the one-parameter exchangeable correlation model {C(θ) : θ ∈ (–(p – 1)–1, 1)}, for which all off-diagonal elements are equal to θ, as well as any model in which the rows of C(θ) are permutations of one another. To see this, note that if ci, the ith row of C(θ), is a permutation of cj, then bi, the ith row of B, is the same permutation of bj. Therefore for each i and j, and so the conditions of Corollary 4.2 are satisfied. Subclasses of such correlation matrices include circular correlation models, often used for seasonal data [Olkin and Press, 1969, Khattree and Naik, 1994], and any model in which the rows of C are permutations of circular matrices.
Exchangeable correlation model: Consider the p = 4 exchangeable correlation matrix, for which
This gives
and
so that when ψ = 1, we have
and so finally
and so our criterion (3) for asymptotic efficiency is met.
Circular correlation model: Consider the correlation model such that
For this model, we have
Letting and t2 = 2(y1y3 + y2y4), we have
Further calculations give
and so our criterion for asymptotic efficiency is not met. Additional calculations (available from the authors) show that the asymptotic variance of the PLE is given by
The first panel Figure 1 plots the asymptotic variance of the PLE with the information bound, and the second panel plots their difference. The PLE is very nearly asymptotically efficient in this example, but this small discrepancy indicates that the PLE is not generally asymptotically efficient for Gaussian copula models.
Figure 1.
Asymptotic variances for the circular copula model. The left panel gives the information bound (dashed black line) and the asymptotic variance of the PLE (gray line). and the right panel gives the difference between these two quantities as a function of θ.
6 Discussion
In this article, we have shown that the existence of a sufficiently accurate rank measurable approximation to the localized log likelihood of a copula parameterized model implies the local asymptotic normality of the log rank likelihood. We have also shown that such approximations exist for every smoothly parameterized Gaussian copula model. For such a copula model, the asymptotic information bound implied by the rank likelihood matches that of the corresponding parametric multivariate normal submodel. This result suggests the possibility of semiparametrically efficient rank-based estimators for Gaussian copula models: Generally speaking, the information Ir based on the ranks is less than or equal to the the semiparametric information If based on the full data, as the ranks are functions of the full data [Le Cam and Yang, 1988]. Furthermore, the semiparametric information based on the full data is less than or equal to Ip, the infimum of information functions over all parametric submodels, and so Ir ≤ If ≤ Ip in general. On the other hand, for Gaussian copula models we have shown that Ir is equal to the information for a particular parametric submodel, the corresponding multivariate normal model. This implies that for a given Gaussian copula model, the corresponding multivariate normal model is least favorable, that Ir = Ip and therefore Ir = If = Ip.
Based on this result, and the partial sufficiency of the multivariate ranks in semiparametric copula models in general, we conjecture that maximum likelihood estimators based on rank likelihoods are asymptotically efficient for Gaussian copula models, and possibly more generally whenever information bounds based on the complete data for the semiparametric model in question exist. However, the rank likelihood involves a multivariate integral over a set of order constraints, the number of which grows with the sample size, making it difficult to use or study. An alternative to the rank likelihood estimator is the pseudo-likelihood estimator [Genest et al., 1995], which is a very explicit function of the copula density, making optimization and asymptotic analysis tractable. For the one-parameter bivariate Gaussian copula model, the rank-based pseudo-likelihood estimator is asymptotically equivalent to the normal scores correlation coefficient, which Klaassen and Wellner [1997] showed to be asymptotically efficient. However, Genest and Werker [2002] showed with a non-Gaussian example that the pseudo-likelihood estimator is not generally asymptotically efficient, and in this article we have shown that this estimator is not generally asymptotically efficient for the restricted class of Gaussian copula models. However, this does not rule out the possibility that other rank-based estimators, such as the maximizer of the rank likelihood, are asymptotically efficient.
Acknowledgments
Peter Hoff's research was supported in part by NI-CHD grant 1R01HD067509-01A1. Jon Wellner's research was supported in part by by NSF Grants DMS-0804587 and DMS-1104832, by NI-AID grant 2R01 AI291968-04, and by the Alexander von Humboldt Foundation.
Contributor Information
Peter D. Hoff, Professor of Statistics and Biostatistics University of Washington Seattle, WA 98195-4322
Xiaoyue Niu, Research Assistant Professor of Statistics Penn State University University Park, PA 16802.
Jon A. Wellner, Professor of Statistics and Biostatistics University of Washington Seattle, WA 98195-4322
References
- Barnard GA. Logical aspects of the fiducial argument. Bull. Inst. Internat. Statist. 1963;40:870–883. [Google Scholar]
- Begun Janet M., Hall WJ, Huang Wei-Min, Wellner Jon A. Information and asymptotic efficiency in parametric–nonparametric models. Ann. Statist. 1983;11(2):432–452. ISSN 0090-5364. doi: 10.1214/aos/1176346151. URL http://dx.doi.org/10.1214/aos/1176346151. [Google Scholar]
- Bickel PJ, Ritov Y. Festschrift for Lucien Le Cam. Springer; New York: 1997. Local asymptotic normality of ranks and covariates in transformation models. pp. 43–54. [Google Scholar]
- Bickel Peter J., Klaassen Chris A. J., Ritov Ya'acov, Wellner Jon A. Johns Hopkins Series in the Mathematical Sciences. Johns Hopkins University Press; Baltimore, MD: 1993. Efficient and adaptive estimation for semiparametric models. ISBN 0-8018-4541-6. [Google Scholar]
- Chen Xiaohong, Fan Yanqin, Tsyrennikov Viktor. Efficient estimation of semipara-metric multivariate copula models. J. Amer. Statist. Assoc. 2006;101(475):1228–1240. ISSN 0162-1459. doi: 10.1198/016214506000000311. URL http://dx.doi.org/10.1198/016214506000000311. [Google Scholar]
- Choi Sungsub, Hall WJ, Schick Anton. Asymptotically uniformly most powerful tests in parametric and semiparametric models. Ann. Statist. 1996;24(2):841–861. ISSN 0090-5364. doi: 10.1214/aos/1032894469. URL http://dx.doi.org/10.1214/aos/1032894469. [Google Scholar]
- de Wet T, Venter JH. Asymptotic distributions of certain test criteria of normality. South African Statist. J. 1972;6:135–149. ISSN 0038-271X. [Google Scholar]
- Genest C, Ghoudi K, Rivest L-P. A semiparametric estimation procedure of dependence parameters in multivariate families of distributions. Biometrika. 1995;82(3):543–552. ISSN 0006-3444. [Google Scholar]
- Genest Christian, Werker Bas J. M. Distributions with given marginals and statistical modelling. Kluwer Acad. Publ.; Dordrecht: 2002. Conditions for the asymptotic semiparametric efficiency of an omnibus estimator of dependence parameters in copula models. pp. 103–112. [Google Scholar]
- Hall WJ, Loynes RM. On the concept of contiguity. Ann. Probability. 1977;5(2):278–282. [Google Scholar]
- Hoff Peter D. Extending the rank likelihood for semiparametric copula estimation. Ann. Appl. Stat. 2007;1(1):265–283. ISSN 1932-6157. [Google Scholar]
- Hoff Peter D. Rank likelihood estimation for continuous and discrete data. ISBA Bulletin. 2008;15(1):8–10. URL http://bayesian.org/sites/default/files/fm/bulletins/0803.pdf. [Google Scholar]
- Khattree Ravindra, Naik Dayanand N. Estimation of interclass correlation under circular covariance. Biometrika. 1994;81(3):612–617. ISSN 0006-3444. doi: 10.1093/biomet/81.3.612. URL http://dx.doi.org/10.1093/biomet/81.3.612. [Google Scholar]
- Klaassen Chris A. J., Wellner Jon A. Efficient estimation in the bivariate normal copula model: normal margins are least favourable. Bernoulli. 1997;3(1):55–77. ISSN 1350-7265. [Google Scholar]
- Le Cam Lucien, Yang Grace L. On the preservation of local asymptotic normality under information loss. Ann. Statist. 1988;16(2):483–520. ISSN 0090-5364. doi: 10.1214/aos/1176350817. URL http://dx.doi.org/10.1214/aos/1176350817. [Google Scholar]
- Olkin I, Press SJ. Testing and estimation for a circular stationary model. Ann. Math. Statist. 1969;40:1358–1373. ISSN 0003-4851. [Google Scholar]
- Pettitt AN. Inference for the linear model using a likelihood based on ranks. J. Roy. Statist. Soc. Ser. B. 1982;44(2):234–243. ISSN 0035-9246. [Google Scholar]
- Rémon M. On a concept of partial sufficiency: L-sufficiency. Internat. Statist. Rev. 1984;52(2):127–135. ISSN 0306-7734. [Google Scholar]
- Severini Thomas A. Likelihood methods in statistics, volume 22 of Oxford Statistical Science Series. Oxford University Press; Oxford: 2000. ISBN 0-19-850650-3. [Google Scholar]
- Shorack Galen R. Springer Texts in Statistics. Springer-Verlag; New York: 2000. Probability for statisticians. ISBN 0-387-98953-6. [Google Scholar]
- Sklar M. Fonctions de répartitionà n dimensions et leurs marges. Publ. Inst. Statist. Univ. Paris. 1959;8:229–231. [Google Scholar]

