Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 Jul 31.
Published in final edited form as: Biometrics. 2014 May 20;70(3):762–770. doi: 10.1111/biom.12186

Partial Correlation Matrix Estimation using Ridge Penalty Followed by Thresholding and Reestimation

Min Jin Ha 1,, Wei Sun 2,
PMCID: PMC4239206  NIHMSID: NIHMS589713  PMID: 24845967

Summary

Motivated by the problem of construction gene co-expression network, we propose a statistical framework for estimating high-dimensional partial correlation matrix by a three-step approach. We first obtain a penalized estimate of a partial correlation matrix using ridge penalty. Next we select the non-zero entries of the partial correlation matrix by hypothesis testing. Finally we reestimate the partial correlation coefficients at these non-zero entries. In the second step, the null distribution of the test statistics derived from penalized partial correlation estimates has not been established. We address this challenge by estimating the null distribution from the empirical distribution of the test statistics of all the penalized partial correlation estimates. Extensive simulation studies demonstrate the good performance of our method. Application on a yeast cell cycle gene expression data shows that our method delivers better predictions of the protein-protein interactions than the Graphic Lasso.

Keywords: Co-expression network, Empirical null distribution, Graphical model, Partial correlation matrix, Ridge regression

1. Introduction

The expression of multiple genes can be studied through a network perspective, where the set of genes of interest are vertices and the relations among the genes are undirected/directed edges. The gene co-expression network analysis is a popular approach to dissect gene expression regulation patterns and to detect functionally related genes (Stuart et al., 2003; de Jong et al., 2012). In this paper we study the (undirected) co-expression network of a group of genes constructed through their partial correlation matrix.

We denote the expression of p genes by a p-dimensional random vector: X = (X1, …, Xp)T ∈ ℝp with an unknown, but positive definite covariance matrix Σ. Let Ω = Σ−1 be the inverse of the covariance matrix Σ, with its element at a-th row and b-th column denoted by Ωab. Ω is also called concentration matrix or precision matrix. The partial correlation between Xa and Xb is a measure of the linear relationship between Xa and Xb after accounting for the linear effects of all the remaining variables (Christensen, 2002). The partial correlations can be obtained by the off diagonal elements of the negative definite matrix −scale(Ω):

R=[ρab]p×p=scale(Ω), (1)

where the scale is an operator defined for a square matrix. Let diag(A) be a diagonal matrix constructed by the diagonal elements of A, then scale(A) = diag(A)−1/2Adiag(A)−1/2. The derivation of equation (1) is presented in the Section A of the Supplementary Materials. The zero structure of the partial correlation matrix of p random variables can be represented by an undirected graph

G=(Γ,E),

where Γ = {1, ⋯, p} is the set of vertices and E is a set of edges in Γ × Γ such that any edge between vertices a and b belongs to E if and only if ρab ≠ 0. We refer to such an undirected graph G as a partial correlation graph. Under multivariate Gaussian distribution assumption for X, zero partial correlation between two variables is equivalent to the conditional independence of these two variables given the remaining variables.

Although many methods have been developed for partial correlation matrix estimation in high dimensional problems where p > n, we find that a simple penalized estimation using ridge penalty has favorable properties. The advantage of this ridge penalization approach has not been appreciated in the existing literature, partly because it does not provide sparse estimates, i.e., none of the partial correlation is estimated exactly as 0. We propose a novel approach to threshold the ridge estimates to select non-zero entries of the partial correlation matrix. Finally we reestimate the partial correlation coefficients on the none-zero entries of the partial correlation matrix. Thresholding ridge estimates is desirable because it leads to parsimonious and more interpretable partial correlation matrix estimate, and further reduces estimation error as well.

Next we briefly review the existing works for estimating concentration matrix or partial correlation matrix and related statistical inference. Suppose there are n independent samples of p random variables X = (X1, …, Xp)T, and let X be the p × n data matrix. Schäfer et al. (2005) proposed to estimate covariance matrix by

Ŝ=(1λ)S+λT, (2)

where S is sample covariance matrix, and T is a target matrix. Schäfer et al. (2005) derived optimal choice of λ to minimize squared loss of covariance matrix estimate for six commonly used target matrices T’s. Then they propose to estimate partial correlation matrix by the inverse of their correlation matrix estimate. Assume X = (X1, …, Xp)T follows multivariate Gaussian distribution with zero mean, denoted by N(0, Σ). The log-likelihood of concentration matrix Ω is proportional to

l(Ω)=log|Ω|tr(SΩ), (3)

where tr(·) is trace of a square matrix and S is the sample covariance matrix. When np, S is positive definite with probability 1 and S−1 is the maximum likelihood estimate (MLE) of Ω (Lauritzen, 1996). However, this approach fails when p > n, and may perform poorly unless n is much larger than p. Therefore MLE with certain constraints or penalized MLE are often used for high dimensional problems when p is lager or much larger than n. Examples include covariance selection from positive definite matrices (Dempster, 1972) or iterative partial maximization based on deviance tests (Speed and Kiiveri, 1986). More general linear restrictions on edges are enabled by colored graph models (Højsgaard and Lauritzen, 2008). Recently, many penalized MLE of Ω have been proposed for high dimensional problems (Yuan and Lin, 2007; Rothman et al., 2008; Banerjee et al., 2008; Friedman et al., 2008; Fan et al., 2009). One of the most widely used methods is the graphic Lasso (Friedman et al., 2008), which maximizes the following penalized log likelihood:

l(Ω)=log|Ω|tr(SΩ)κa,b|Ωab|, (4)

where κ is a tuning parameter.

With a focus on determining the partial correlation graph, rather than precise estimation of the partial correlation coefficients, Meinshausen and Bühlmann (2006) proposed a regression-based approach called neighborhood selection. The neighborhood for each vertex was estimated by penalized regression of the corresponding variable versus the remaining variables. Banerjee et al. (2008); Friedman et al. (2008) showed that estimating the penalized MLE (with L1 penalty) of Ω could be viewed as p-coupled iterative versions of the p separate neighborhood selections. More recent methodology developments related with neighborhood selection include Yuan (2010) and Zhou et al. (2011).

Statistical inference of partial correlation estimates is another topic related with our method development, particularly the second step of our method for thresholding partial correlations. Given a partial correlation estimate, denoted by ρ̂, one may test H0 : ρ = 0 against HA : ρ ≠ 0 using a test statistic constructed by Fisher’s Z-transformation: ψ(ρ̂) = 0.5 log {(1 + ρ̂) / (1 − ρ̂)}. Specifically, one may reject the null hypothesis at level α if (np − 1)1/2|ψ(ρ̂)| > Φ−1(1 − α/2) for standard normal c.d.f. Φ (Anderson, 2003). However, this testing procedure assumes the sample size n is substantially greater than p. For high dimensional problems with p > n, Schäfer and Strimmer (2005) proposed an inference approach by assuming partial correlation estimates across all variables followed a mixture of null and alternative distributions where the null was Hotelling distribution with unknown degree of freedom and the alternative was uniform (−1,1). Magwene et al. (2004) and Wille et al. (2004) proposed to use low-order partial correlations to avoid singularity problem when p > n. Subsequently, Wille and Bühlmann (2006) discussed more formal statements on Gaussian graphical model inference using low-order partial correlations. Castelo and Roverato (2006) generalized the 0–1 partial correlation graph to an arbitrary qp − 2 order partial correlation graph.

The remaining parts of the paper are organized as follows. We present our method in Section 2, demonstrate the effectiveness of our method by simulations and real data analysis in Section 3, and conclude this paper by some discussions in Section 4.

2. Method

2.1 Estimation of partial correlation matrix using ridge penalty

Without loss of generality, we assume each row of the p × n data matrix X has been standardized to have mean 0 and standard deviation 1 so that S = XXT/n is the sample correlation matrix. Then a straightforward estimate of the off-diagonal elements of a partial correlation matrix can be obtained from

=scale(S1).

However, when n < p, S is not invertible. To solve the singularity problem of inverting a sample correlation matrix, we add a positive constant to the diagonal elements of the sample correlation matrix:

(λ)=scale((S+λIp)1), (5)

where λ ≥ 0 and Ip is a p × p identity matrix. We call S+(λ) = (S + λIp)−1 as the ridge inverse in the analogy to ridge regression (Hoerl and Kennard, 1970). The modified sample covariance matrix S + λIp guarantees full rank for any λ > 0, and has been used as an initial covariance matrix estimate in the coordinate descent algorithms in Banerjee et al. (2008) and Friedman et al. (2008).

Next we show that as λ varies from 0 to ∞, (λ) varies from a scaled generalized inverse to an identity matrix. Let X/n=UDVT be a singular value decomposition with rank(X) = kmin(n, p), where U and V are, respectively p × p and n × n orthogonal matrices, D is p × n diagonal matrix with its first k nonzero diagonal elements d1, …, dk and all other elements being zero. Since S+(λ) = U(D + λIp)−1UT, it is obvious that

scale(S+(λ))scale(S)asλ0, (6)

where S is Moore-Penrose generalized (MPG) inverse of S if k < p (Schott, 2005). By the invariance of the scale operator under scalar product,

scale(S+(λ))=scale(λS+(λ))Ipasλ. (7)

Since the estimates of regression coefficients using MPG inverse is minimum L2 solution (proposition 1 of Lv and Fan (2009)), S+(λ) goes to k rank ridge inverse when λ goes to 0 by (6). From (7) the partial correlation matrix shrinks toward the identity matrix as λ goes to infinity. In practice, the optimal performance of this ridge estimate relies on an appropriate choice of λ, which will be addressed after we introducing the other two steps of our method.

2.2 Thesholding

We propose a hypothesis testing approach to threshold the ridge estimate of partial correlations (λ)=[ρ̂abλ]p×p, where λ is the tuning parameter for ridge estimate. We first apply Fisher’s Z-transformation on partial correlations estimate, denoted by {ψ(ρ̂abλ):aΓ, bΓ, and ab}. We assume these z-statistics follow a mixture of null and alternative distributions, corresponding to the cases where the true partial correlations are zero or non-zero, respectively. By assuming the vast majority of the z-statistics in the central part of this mixture distribution (i.e., a region around 0) arise from the null distribution, we estimate the null distribution using Efron’s central matching method (Efron, 2004). Specifically, we assume the observed z-statistics follow a mixture distribution

f(ψ)=ηf0(ψ)+(1η)fa(ψ), (8)

where the null distribution f0(ψ) is a normal distribution N(μ0,σ02), the alternative distribution fa(ψ) is left un-specified, and η is the proportion of the z-statistics arising from the null distribution. Using Efron’s central matching method (Efron, 2004), we estimate the null distribution N(μ0,σ02) by matching the mixture distribution and the null distribution at the central part of the distributions. Specifically, assuming f(ψ) = f0(ψ) around ψ = 0 gives

logf(ψ)=12(ψμ0σ0)2+C (9)

for a constant C.

We estimate f(ψ) (the density of the mixture distribution) using polynomial Poisson regression. The range of the p(p−1)/2 observed ψ values is partitioned into K equal intervals with interval k having mid point xk and sk observed ψ values. sk’s (k=1,…,K) are assumed to be independently distributed following Poisson distributions with mean νk’s. We fit a q degree polynomial Poisson regression on νk,

log(νk)=log(f(xk)/c)=j=1qθj(xk)j, (10)

for k = 1, …, K and a normalizing constant c making the marginal density f(ψ) integrated to 1. The estimates of {θj : j = 1, …, q} are used to estimate log ((ψ)/c)=j=1qθ̂jψj. Then using equation (9), we can obtain the estimates of μ0 and σ0:

μ̂0=arg maxψ{(ψ)},andσ̂0=[d2dψ2log(ψ)]ψ=μ̂012. (11)

With the estimate of null distribution N(μ̂0, σ̂0), we can calculate p-values for each partial correlation estimate. The degree of polynomial regression, q, is a nuisance parameter. Based on the sparsity assumption that most p-values arise from the null, we choose the q so that the p-values are most uniformly distributed. The empirical distribution function of the p-values, {πab(q)|abΓ} given q, is

Fq(π)=2p(p1)a,bΓ,abI(πab(q)π). (12)

We suggest to estimate q by

=arg minq[sup0<π<1|Fq(π)F0(π)|], (13)

where F0(π) is uniform distribution between 0 and 1, and Dq = sup0<π<1 |Fq(π) − F0(π)| is a distance measure used in Kolmogorov-Smirnov statistic. Figure S6 (Supplementary Materials Section F) displays the average and one standard deviation of Dq values over 100 simulation data sets for p = 500, n = 30 and η = 1 or η = 0.9997, which corresponds to 38 non-zero partial correlations. Adding 38 nonzero partial correlations to the null needs 3 or 4 higher polynomial order on average to estimate the null distribution. Finally, a threshold α is needed to select non-zero entries of the partial correlation matrix. We select α by cross-validation, and we defer the discussion of details to section 2.4. Given this threshold, we can estimate the sparsity η. An upper-bound of η can also be estimated following (Efron, 2004) (Supplementary Materials Section C). From our simulations, the estimate of η based on our cross-validation selected threshold is more accurate.

2.3 Re-estimation of partial correlation coefficients

Given the partial correlation graph structure estimated in the previous step, we re-estimate the partial correlation coefficients at the non-zero entries of the partial correlation matrix. Suppose that the covariance matrix Σ and the concentration matrix Ω are partitioned according to random variables Xa and Xa where Xa is a (p − 1) × 1 random vector except for a random variable Xa. The blocks are denoted by Σa,a, Σa,a, Σa,−a, Σa,−a and Ωa,a, Ωa,a, Ωa,−a, Ωa,−a. Consider the best linear predictor of Xa by XaTβa for any aΓ. Let εa=XaXaTβa. It is easy to show that βa=Σa,a1Σa,a and Var(XaXaTβa)=Var(εa)=Σa,aΣa,aΣa,a1Σa,a. From inverse formula for block matrix and Ω = Σ−1,

Ωa,a=(Var(εa))1,Ωa,a=(Var(εa))1βa. (14)

From Ê(λ, α) estimated in the thresholding step, we know all the variables adjacent to aΓ, denoted by ne^a. Based on the sparsity assumption, we assume |ne^a|<n, then we can have the following refined estimates of the concentration matrix:

Ω̂a,a=(n|ne^a|)/XaXne^aβ̂a,ne^a22,Ω̃a,a=Ω̃a,aβ̂a,ne^a, (15)

where β̂a,ne^a=(Xne^aXne^aT)1Xne^aXa,Xne^a is |ne^a|×n submatrix of X corresponding to ne^a. Since this solution is not symmetric in general, we set the final estimates of the off-diagonal elements of Ω as

Ω̂a,b=Ω̂b,a=sign(β̂ba,ne^a)|Ω̃a,bΩ̃b,a|forab, (16)

and obtain the estimate the partial correlation coefficients from −scale(Ω̂).

2.4 Tuning parameter selection

Our method has two tuning parameters. One is λ for ridge estimate of partial correlation matrix and the other is α, the p-value cutoff. We perform a two-grid search of the combination of λ and α by minimizing sum squared prediction errors of p separate ridge regressions. Specifically, given λ and α, let Nane^a be the neighborhood of Xa after thresholding. Let Xa be an n × 1 vector of the measurements of variable Xa, and let XNa be an n×|ne^a| matrix of the measurements of the variables XNa = {Xk : kNa}. Then the sum squared prediction error is

a[XaXNaβ̂a(λ)]T[XaXNaβ̂a(λ)],

where

β̂a(λ)=arg minβp11n(XaXNaβ)T(XaXNaβ).

We present the details of this tuning parameter selection method and its justifications into Supplementary Materials Section B.

As an alternative, one may conduct p separate ridge regressions with different λ’s to estimate a partial correlation matrix. This is equivalent to a general form of the ridge estimation of partial correlation matrix:

(λ)=scale((S+diag(λ1,,λp))1),

where λ1 > 0, …, λp > 0 and diag1, …, λp) is a p × p diagonal matrix with diagonal elements λ1, …, λp. For all i = 1, …, p, we choose the tuning parameter λi by minimizing 10-fold cross validation estimates of the prediction errors of the ridge regression with Xi as response variable and all the other variables as covariates. We refer this method as separate estimation, and refer the method with λ = λ1 = … = λp as joint estimation. We show that the joint estimation, although suffers from larger bias, does provide more stable ordering of the partial correlation coefficients than the separate estimate, and hence more accurate estimate of partial correlation matrix after thresholding and re-estimation, which eliminate the bias of initial estimates (Supplementary Materials, Section D).

3. Results

3.1 Simulation I

We first use a simple simulation to demonstrate that the p-values calculated using central matching method follow the expected uniform distribution under null, while the p-values calculated using asymptotic distribution can lead to inflated type I error. We simulated data from multivariate Gaussian distribution N(0, Ip×p) with p = 100 and n = 1000 or 110. All pairwise partial correlations were calculated by inverting the sample correlation matrix, and then the test statistics were calculated by Fisher’s Z transformation of the partial correlations. The p-values of the test statistics were calculated using theoretical null distribution N(0, 1/np − 1), and the empirical null distribution estimated by central matching method. As shown in the qq-plots of Figure 1, when p = 100 and n = 1000, p-values calculated using either the theoretical null distribution or the empirical null distribution followed the expected uniform distribution. However when the sample size was decreased to n = 110, the p-values calculated from the empirical null distribution were still uniformly distributed but the p-values calculated from the theoretical null distribution were severely inflated.

Figure 1.

Figure 1

QQ-plots for p-values calculated using theoretical null distribution (larger points) or empirical null distribution estimated by central matching method (smaller points) against the expected uniform distribution on [0,1]. (a) p = 100 and n = 1000, (b) p = 100 and n = 110. The dotted lines are the 90% confidence limits of the expected values.

3.2 Simulation II

We consider random networks where both the network structure and the partial correlation coefficients are random. The only restriction is that the partial correlation matrix is diagonally dominant, so that R is a strictly negative definite matrix. The simulation datasets were generated following similar approach of Schäfer and Strimmer (2005). We simulated a p×n data matrix X composed of n independent random samples from p dimensional multivariate Gaussian distribution Np(0, Σ), where Σ is determined by a simulated concentration matrix Ω. We initialized Ω by a p × p matrix with all elements being 0’s. Given η, the proportion of null edges among all the p(p − 1)/2 edges, we randomly selected 100(1 − η)% of the off-diagonal elements of Ω and filled in values from uniform distribution on [−1,1]. To ensure that Ω is a positive definite matrix, the diagonal elements of Ω were filled by column-wise sums of absolute values plus a small constant. Finally Σ was calculated by scale(Ω−1).

Let |E| be the number of edges in set E. The Gaussian simulation settings are

  1. p = 50, n = 100, and |E| = 45, 55, 65, 75 (1 − η ≈ 0.037, 0.045, 0.053, 0.061)

  2. p = 200, n = 100, and |E| = 160, 200, 220, 240 (1 − η ≈ 0.008, 0.01, 0.011, 0.012).

    Because our proposed method does not require X to follow multivariate normal distribution, we assumed that each column of X is independently following multivariate central t-distribution with degrees of freedom 2 after specifying Σ. The t simulation setting is

  3. p = 194, n = 75, and |E| = 160 (1 − η ≈ 0.008)

This simulation setting has the median level of dimensionality and sparsity of the gene clusters studied in our real data example in section 3.3.

We first evaluated the accuracy of partial correlation graph using ROC curves, and compared our method with one of the most widely used method for partial correlation matrix estimation, the Graphic Lasso (GLasso) (Friedman et al., 2008). For our method, the ROC curve is drawn by selecting optimal ridge parameter λ using 10-fold cross-validation, and then calculating sensitivity and specificity across different p-value cutoffs α. This approach is less ideal than the two-grid search of λ and α, but the ROC curve more smooth and is easier to interpret since the sensitivity is a monotone function of α. For GLasso, we used the implementation in R package GLasso, and the sensitivity and specificity were calculated across different values of the tuning parameter κ ranging from 0.5 to 1000. For each p-value cutoff for our method and κ for GLasso, 1-specificity and sensitivity are averaged across 100 simulation data. As shown in Figure 2, our method has uniformly better sensitivity and specificity than the GLasso in estimating network structure.

Figure 2.

Figure 2

The ROC curves for identifying zero entries of partial correlation matrix using our method or GLasso for three simulation settings: (a) Mutivariate Gaussian for n = 100, p = 50, and |E| = 45. (b) Mutivariate Gaussian for n = 100, p = 200, and |E| = 160. (c) Mutivariate t-distribution for n = 75, p = 194, and |E| = 160.

These simulation studies also demonstrate that our density estimates of the test statistics ψ (the Fisher’s Z-transformation of ridge estimates of partial correlations) fit the observed distribution well (Figure 3). To estimate the sparsity η (the proportion of partial correlations that are zero), there are two possible methods: (1) the proportion of zero partial correlations obtained from our thresholding step; and (2) an estimate of the upper bound of η by (Efron, 2004) (see Supplementary Materials Section C for details). For one typical simulation in the low dimension setting with true sparsity 0.963 (Figure 3 (a)), the estimated null density is 0 = N(−0.002, 0.0252) and the estimated sparsity level is η̂ = 0.960 (Efron’s upper bound is 0.981). For one typical simulation in the high dimension case with true sparsity 0.992 (Figure 3 (b)), the estimated null density is 0 = N(−0.0002, 0.0112) and the estimated sparsity level is η̂ = 0.988 (Efron’s upper bound is 0.995). An alternative choice to threshold partial correlation estimates is the hypothesis testing approach by Schäfer and Strimmer (2005), which is an EM algorithm assuming null and alternative distributions are Hotelling distribution and uniform distribution, respectively. However the density curves of null and mixture distributions of their method do not fit the empirical distribution well and thus we do not further pursue it here (Supplementary Materials Section E).

Figure 3.

Figure 3

Histograms of ψ values and density curves of the estimated null and the marginal density. (a) p = 50, n = 100 and |E| = 45 (b) p = 200, n = 100 and |E| = 160.

We evaluated the accuracy of partial correlation coefficient estimates using sum squared error (SSE), and compared our method with GLasso and the method of Schäfer et al. (2005), where partial correlation matrix was estimated by inverse of an optimal estimate of correlation matrix (implemented by function pcor.shrink of R package corpcor). Given the set of vertices Γ = {1, 2, …, p}, The SSE was calculated as

L(R,)=abΓ(ρ̂abρab)2, (17)

where = [ρ̂ab]p×p was the estimates of R. The mean values of SSE from 100 replicates of each simulation setting were reported. As shown in Figure 4, our method consistently show better performance than GLasso and the method of Schäfer et al. (2005).

Figure 4.

Figure 4

Sparsity (proportion of zero partial correlation estimates) versus log(SSE) for three settings: (a) Gaussian, n = 100, p = 50, and |E| = 45; (b) Gaussian n = 100, p = 200, and |E| = 160; (c) t-distribution n = 75, p = 194, and |E| = 160. The horizontal black line is log(SSE) values when a p × p identity matrix is used and the vertical black line indicates the sparsity of the true network.

The ROC curves and SSE results of other simulation settings for our method and GLasso are displayed in Figures S7–S12 (Supplementary Materials Section F). As number of edges |E| increases, the performance of our method deteriorates while the performance of GLasso improves. For denser graph, the assumption that the null density match with the mixture density at the central part may be violated, and thus the estimate of null density may be biased.

3.3 Application

We applied our method to estimate the partial correlation graph of the expression of 6178 genes from yeast cell cycle data (Spellman et al., 1998). The gene expression data were downloaded from http://genome-www.stanford.edu/cellcycle/data/rawdata/. After removing the samples with more than 20% missing values, 75 samples remained for further analysis. We imputed the remaining missing values of the expression data using nearest neighbor averaging. Then the expression data of each sample were normalized by quantile normalization. In this analysis we did not account for the time-dependent nature of the data and treated the expression of each gene across 75 samples as independent observations.

Denote the observed gene expression data as a matrix X of dimension 6178×75. Each gene is a variable, and thus Γ is {1, …, 6178}. We first grouped the 6178 genes into h clusters. Let Ci be the genes belonging to the i-th cluster, then i=1h|Ci|=6178. We separately constructed the partial correlation graph within each cluster. The graph of ith cluster is denoted by GCi = (Ci, ECi) for i = 1, …, h. We assumed the genes from different clusters were independent, so that the edge set E of the whole graph was estimated by Ê=i=1hÊCi. Specifically, we clustered the 6178 genes using hierarchical clustering with Ward’s minimum variance method and the distance between two genes a and b was defined as 1−|ρ̂ab|∅| where ρ̂ab|∅ denoted marginal Pearson correlation. We chose the number of clusters to be 25 based on the gap statistic (Tibshirani et al., 2001). Cluster sizes varied from 32 to 1370, with 25 percentile, median, and 75 percentile being 139, 194, and 254, respectively. Among all the 19,080,753 gene pairs of the 6178 genes, 1,556,154 belonged to the same cluster, and hence could be connected based on the partial correlation graph estimates.

The density curves estimated by our method fit the observed data well. For example, Figure S13 (Supplementary Materials Section F) shows the results for one cluster including 124 genes. Here the null distribution is N(0.003, 0.0282), and the density of the mixture distribution is estimated using the 5 degrees of polynomial Poisson regression.

We compared the performance of our method with the GLasso (Friedman et al., 2008). We chose the tuning parameter κ of GLasso based on the extended Bayesian information criterion (BIC) (Foygel and Drton, 2010):

BICγ(κ)=log|Ω̂(κ)|+tr(Ω̂(κ)S)+1n|Ê(κ)|logn+4n|Ê(κ)|γlogp, (18)

where Ω̂(κ) was the estimate of the inverse covariance matrix using GLasso with tuning parameter κ, Ê(κ) was the edge set obtained from Ω̂(κ), and γ ∈ [0, 1] is a tuning parameter of the extended BIC. If γ = 0, the classical BIC used in Yuan and Lin (2007) was recovered. Given a fixed γ value, we applied GLasso with tuning parameter selected by the extended BIC (18) to construct the partial correlation graphs for all 25 clusters. An exception is that for two clusters with low dimension such that p < n/2, we always used the classical BIC by setting γ = 0. Different choices of γ led to the different model selection results.

The estimates of partial correlation graphs were evaluated by comparing the edge set Ê with yeast protein-protein interaction database at http://thebiogrid.org/download.php. Table 1 displays the number of directed edges, the number of undirected edges (after omitting the directions) and the number of vertices in each of 20 protein-protein interaction dataset. We considered two genes are truly connected if they belonged to the same cluster and the corresponding proteins had interaction according to at least one of the protein-protein interaction datasets. Among 1,556,154 gene pairs belonging to the same cluster, 9,382 were connected and 1,546,772 were not connected. Many gene pairs may be connected according to their partial correlation, but may not be connected by this standard of protein protein interaction. Therefore by defining true connections by protein-protein interaction, we would over-estimate false discovery rate. However, such over-estimation does not interfere with comparison across different methods. Given this imperfect, but biologically meaningful definition of true/false connections, we evaluated our method and GLasso by ROC curves. The ROC curve of our method was generated across different p-value cutoffs α and the ROC curve of the GLasso was generated across different γ values in the extended BIC. Our method had uniformly higher sensitivity and specificity than the GLasso (Figure 5).

Table 1.

Summary of the protein-protein interaction database

ID Experiment system (type) no. of directed edges no. of undirected edges no. of vertices
1 Affinity Capture-MS (physical) 72767 42538 4613
2 Affinity Capture-Western (physical) 13105 7795 2727
3 Dosage Rescue (genetic) 4812 4022 2161
4 Reconstituted Complex (physical) 5110 3946 1988
5 Synthetic Lethality (genetic) 13870 10965 2915
6 Two-hybrid (physical) 13986 10827 3392
7 Biochemical Activity (physical) 5703 5220 1946
8 Co-crystal Structure (physical) 387 337 421
9 FRET (physical) 142 119 117
10 Protein-peptide (physical) 673 643 353
11 Co-localization (physical) 527 484 441
12 Affinity Capture-RNA (physical) 5895 5888 3702
13 Protein-RNA (physical) 408 399 377
14 PCA (physical) 5117 4845 1663
15 Co-purification (physical) 1675 1309 933
16 Co-fractionation (physical) 777 725 663
17 Dosage Lethality (genetic) 971 945 786
18 Phenotypic Enhancement (genetic) 6449 4803 2153
19 Phenotypic Suppression (genetic) 5287 3965 1729
20 Synthetic Haploinsufficiency (genetic) 262 262 262

Figure 5.

Figure 5

Comparing our method (Ridge+thresholding) with GLasso in terms partial correlation graph estimation by ROC curves, while the underlying true connections are defined as gene pairs belonging the the same cluster and their proteins having protein-protein interaction.

4. Discussion

Motivated by the gene co-expression network estimation problem, we have developed a new framework for estimation and statistical inference of partial correlation matrix. Both simulation and real data analysis have demonstrated the effectiveness of our method. For real data analysis where p is much larger than n, we cluster the genes and then estimate partial correlation matrix within each cluster. This is based on an reasonable assumption that the partial correlation matrix of gene expression has a block diagonal structure. We used the hierarchical clustering to group genes. There are many other clustering method available Monti et al. (2003); Zhang et al. (2005), though a careful study of which clustering method can better identify the block diagonal structure is beyond the scope of this paper. Our method does not require multivariate Gaussian distribution assumption. However, without this assumption, partial correlation being zero may not imply the two variables are conditionally independent with each other.

The first step of our method, ridge estimation of partial correlation matrix is closely related with Schäfer et al. (2005)’s method. However the objective of our method is to estimate partial correlation matrix while Schäfer et al. (2005)’s method is designed to estimate correlation matrix, and the optimal estimate of correlation matrix does not guarantee the optimal estimate of partial correlation matrix, as shown in our simulation studies. In terms of implementation, Schäfer et al. (2005) estimated covariance matrix in one step using a single tuning parameter, which is selected to minimize the squared loss of correlation matrix estimate. In contrast, we estimate partial correlation matrix in three steps with two tuning parameters, which are selected by a two-grid search to minimize the squared loss of partial correlation matrix estimate.

While the work in the field of partial correlation matrix estimation is moving toward more sophisticated penalized estimation, our work shows that a conceptually simple approach of ridge estimation + thresholding + reestimation can deliver surprisingly good results. One key of the success of our method is that we borrow information across variables to reduce estimation variance, while scarifying estimation bias. This point is more clear when comparing our method with neighborhood selection by ridge regression. In neighborhood selection, different tuning parameter λ’s will be used for different variables, which will reduce the bias of ridge estimation, but increases variance. In contrast, our method borrows information across all the variables by utilizing a common tuning parameter λ across all variables, and this approach increase the bias of the partial correlation estimate, but reduces the variance. The reduced variance leads to better discrimination of zero and non-zero parietal correlations in the thresholding step, and the inated bias can be corrected in the final reestimation step.

Supplementary Material

Supp Material

Acknowledgements

We are grateful for constructive comments from the anonymous reviewers, and the associate editor. These comments lead to significant improvement of the work presented in this paper. We would like to acknowledge the support from EPA grant RD-83382501, and NIH grant P01 CA142538-03, R01 CA149569-05, and R01 MH101819-01.

Footnotes

Supplementary Materials

Web Appendices, Tables, Figures referenced in Sections 1, 2.2., 2.4, 3.3, and 3.4 and the GGMridge R package are available with this paper at the Biometrics website on Wiley Online Library.

Contributor Information

Min Jin Ha, Email: mjha@live.unc.edu, Department of Biostatistics, UNC Chapel Hill, North Carolina, U.S.A..

Wei Sun, Email: weisun@email.unc.edu, Department of Biostatistics, Department of Genetics, UNC Chapel Hill, North Carolina, U.S.A..

References

  1. Anderson T. An introduction to multivariate statistical analysis. Wiley Series in Probability and Statistics; 2003. [Google Scholar]
  2. Banerjee O, El Ghaoui L, d'Aspremont A. Model selection through sparse maximum likelihood estimation for multivariate gaussian or binary data. The Journal of Machine Learning Research. 2008;9:485–516. [Google Scholar]
  3. Castelo R, Roverato A. A robust procedure for gaussian graphical model search from microarray data with p larger than n. The Journal of Machine Learning Research. 2006;7:2621–2650. [Google Scholar]
  4. Christensen R. Plane Answers to Complex Questions: The Theory of Linear Models. Springer; 2002. [Google Scholar]
  5. de Jong S, Boks M, Fuller T, Strengman E, Janson E, de Kovel C, Ori A, Vi N, Mulder F, Blom J, et al. A gene co-expression network in whole blood of schizophrenia patients is independent of antipsychotic-use and enriched for brain-expressed genes. PloS one. 2012;7:e39498. doi: 10.1371/journal.pone.0039498. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Dempster A. Covariance selection. Biometrics pages. 1972:157–175. [Google Scholar]
  7. Efron B. Large-scale simultaneous hypothesis testing. Journal of the American Statistical Association. 2004;99:96–104. [Google Scholar]
  8. Fan J, Feng Y, Wu Y. Network exploration via the adaptive lasso and scad penalties. The Annals of Applied Statistics. 2009;3:521. doi: 10.1214/08-AOAS215SUPP. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Foygel R, Drton M. Extended bayesian information criteria for gaussian graphical models. arXiv preprint arXiv:1011.6640. 2010 [Google Scholar]
  10. Friedman J, Hastie T, Tibshirani R. Sparse inverse covariance estimation with the graphical lasso. Biostatistics. 2008;9:432–441. doi: 10.1093/biostatistics/kxm045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Hoerl A, Kennard R. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics. 1970;12:55–67. [Google Scholar]
  12. Højsgaard S, Lauritzen S. Graphical gaussian models with edge and vertex symmetries. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 2008;70:1005–1027. [Google Scholar]
  13. Lauritzen S. Graphical models. ume 17. USA: Oxford University Press; 1996. [Google Scholar]
  14. Lv J, Fan Y. A unified approach to model selection and sparse recovery using regularized least squares. The Annals of Statistics. 2009;37:3498–3528. [Google Scholar]
  15. Magwene P, Kim J, et al. Estimating genomic coexpression networks using first-order conditional independence. Genome Biol. 2004;5:R100. doi: 10.1186/gb-2004-5-12-r100. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Meinshausen N, Bühlmann P. High-dimensional graphs and variable selection with the lasso. The Annals of Statistics. 2006;34:1436–1462. [Google Scholar]
  17. Monti S, Tamayo P, Mesirov J, Golub T. Consensus clustering: a resampling-based method for class discovery and visualization of gene expression microarray data. Machine learning. 2003;52:91–118. [Google Scholar]
  18. Rothman A, Bickel P, Levina E, Zhu J. Sparse permutation invariant covariance estimation. Electronic Journal of Statistics. 2008;2:494–515. [Google Scholar]
  19. Schäfer J, Strimmer K. An empirical bayes approach to inferring large-scale gene association networks. Bioinformatics. 2005;21:754–764. doi: 10.1093/bioinformatics/bti062. [DOI] [PubMed] [Google Scholar]
  20. Schäfer J, Strimmer K, et al. A shrinkage approach to large-scale covariance matrix estimation and implications for functional genomics. Statistical applications in genetics and molecular biology. 2005;4:32. doi: 10.2202/1544-6115.1175. [DOI] [PubMed] [Google Scholar]
  21. Schott J. Matrix analysis for statistics. Wiley; 2005. [Google Scholar]
  22. Speed T, Kiiveri H. Gaussian markov distributions over finite graphs. The Annals of Statistics. 1986;14:138–150. [Google Scholar]
  23. Spellman P, Sherlock G, Zhang M, Iyer V, Anders K, Eisen M, Brown P, Botstein D, Futcher B. Comprehensive identification of cell cycle–regulated genes of the yeast saccharomyces cerevisiae by microarray hybridization. Molecular biology of the cell. 1998;9:3273–3297. doi: 10.1091/mbc.9.12.3273. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Stuart J, Segal E, Koller D, Kim S. A gene-coexpression network for global discovery of conserved genetic modules. Science. 2003;302:249–255. doi: 10.1126/science.1087447. [DOI] [PubMed] [Google Scholar]
  25. Tibshirani R, Walther G, Hastie T. Estimating the number of clusters in a data set via the gap statistic. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 2001;63:411–423. [Google Scholar]
  26. Wille A, Bühlmann P. Low-order conditional independence graphs for inferring genetic networks. Statistical applications in genetics and molecular biology. 2006;5:1–32. doi: 10.2202/1544-6115.1170. [DOI] [PubMed] [Google Scholar]
  27. Wille A, Zimmermann P, Vranová E, Fürholz A, Laule O, Bleuler S, Hennig L, Prelic A, Von Rohr P, Thiele L, et al. Sparse graphical gaussian modeling of the isoprenoid gene network in arabidopsis thaliana. Genome Biol. 2004;5:R92. doi: 10.1186/gb-2004-5-11-r92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Yuan M. High dimensional inverse covariance matrix estimation via linear programming. The Journal of Machine Learning Research. 2010;99:2261–2286. [Google Scholar]
  29. Yuan M, Lin Y. Model selection and estimation in the gaussian graphical model. Biometrika. 2007;94:19–35. [Google Scholar]
  30. Zhang B, Horvath S, et al. A general framework for weighted gene co-expression network analysis. Statistical applications in genetics and molecular biology. 2005;4:1128. doi: 10.2202/1544-6115.1128. [DOI] [PubMed] [Google Scholar]
  31. Zhou S, Rütimann P, Xu M, Bühlmann P. High-dimensional covariance estimation based on gaussian graphical models. The Journal of Machine Learning Research. 2011;12:2975–3026. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supp Material

RESOURCES