Abstract
Many high dimensional and high-throughput biological datasets have complex sample correlation structures, which include longitudinal and multiple tissue data, as well as data with multiple treatment conditions or related individuals. These data, as well as nearly all high-throughput ‘omic’ data, are influenced by technical and biological factors unknown to the researcher, which, if unaccounted for, can severely obfuscate estimation of and inference on the effects of interest. We therefore developed CBCV and CorrConf: provably accurate and computationally efficient methods to choose the number of and estimate latent confounding factors present in high dimensional data with correlated or nonexchangeable residuals. We demonstrate each method’s superior performance compared to other state of the art methods by analyzing simulated multi-tissue gene expression data and identifying sex-associated DNA methylation sites in a real, longitudinal twin study.
Keywords: Batch effects, unwanted variation, confounding, correlation, multi-tissue, cell-type heterogeneity
1. Introduction
The development of high-throughput technologies has provided biologists with a cornucopia of genetic, proteomic and metabolomic data that can help elucidate the genetic components of phenotypes and the mediation of environmental exposures. Many of these ‘omic’ data have complex sample-correlation structure, which includes longitudinal data (Baumgart et al., 2016; McKennan et al., 2018), multi-tissue data (GTEx Consortium, 2017), data with multiple treatment conditions (Knowles et al., 2018) and data with related individuals (Martino et al., 2013; Tung et al., 2015). Longitudinal studies have even been cited as a critical area of future DNA methylation research to assess the stability of methylation marks over time (Breton et al., 2017; Martin and Fry, 2018), so it is crucial that suitable methods exist to analyze these data. An important feature of these high-throughput data is the presence of unmeasured factors that influence the measured response, which include technical factors like batch variables and biological factors like cell composition (Leek et al., 2010; Jaffe and Irizarry, 2014). When unaccounted for, these factors can bias test statistics, reduce power and lead to irreproducible results (Yao et al., 2012; Peixoto et al., 2015).
There have been a number of methods developed by the statistical community to estimate and correct for latent factors in high throughput biological data (Buja and Eyuboglu, 1992; Leek and Storey, 2008; Gagnon-Bartsch and Speed, 2012; Sun et al., 2012; Gagnon-Bartsch et al., 2013; Houseman et al., 2014; Owen and Wang, 2016; Fan and Han, 2017; Lee et al., 2017; Wang et al., 2017; McKennan and Nicolae, 2019). However, these methods make the critical assumption that conditional on both the observed and unobserved covariates, samples are independent with homogeneous residual variances, and tend to perform poorly when these assumptions are violated. The goal of this paper is to provide a provably accurate method to both choose the number of and estimate latent factors from the measured correlated data so that downstream inference on the effects of interest is as accurate as when all latent factors are observed. To the best of our knowledge, this is the first method to estimate and correct for latent covariates in high throughput biological data with correlated samples.
We use DNA methylation quantified in p ≈ 8 × 105 methylation sites (CpGs) in n / 2 = 183 unrelated individuals at birth and age 7 from McKennan et al. (2018) as a motivating data example, although we analyze other methylation data with a more complex covariance structure in Section 6. The aim of that study was to jointly model methylation at birth and age 7 to determine if the effects due to ancestry on methylation levels changed or remained constant over time. If we ignore observed nuisance variables like the intercept, a reasonable model relating ancestry to methylation levels at CpG g at birth, , and age 7, , in the presence of unobserved covariates is
(1) |
where X = A ⊗ A, βg = (βg,0 βg,7)T and βg,0, βg,7 are the effects of interest. is a partition matrix that groups samples by individuals and captures the within-individual variability. and capture the potential difference in residual variance at birth and age 7 and are diagonal matrices with ones in the first n / 2 and second n / 2 diagonal entries, respectively. In order to avoid overestimating the residual variance or biasing our estimates for βg, we must first estimate C and its latent dimension K. One approach would be to use methods from the economics literature that allow for dependent residuals in their theoretical arguments (Bai, 2009; Li et al., 2016; Lu and Su, 2016; Su and Ju, 2018). However, their assumptions on the homogeneity of the effects β1, …, βp are too restrictive to have any application in genetic data. Another strategy would be to use existing methods designed for genetic data to estimate the factors at each age separately. However, this reduces the sample size by 50% and risks underestimating K, and subsequently biasing estimates for βg. To avoid these biases, a second option would be to estimate K and C using all n samples. However, naively estimating K with methods commonly applied to genetic data with independent and identically distributed residuals, like parallel analysis (Buja and Eyuboglu, 1992) and bi-cross validation (Owen and Wang, 2016), will typically overestimate K to be on the order of the sample size, as they will be unable to distinguish the low dimensional factors C from the high dimensional random effect. Even if K were known, the correlation among the residuals obfuscates existing estimates for C.
Our proposed method uses all of the available data to estimate K and C, and subsequently the effects of interest β1, …, βp, in data whose genomic unit-specific covariance can be written as a linear combination of known matrices. This covariance structure is quite general and includes longitudinal, multi-tissue, multi-treatment and twin studies, as well as studies with individuals related through a kinship matrix.
The remainder of the paper is organized as followed. We describe the data and set up the problem in Section 2 and present a detailed description of our method in Section 3. We prove its efficacy in Section 4 by proving its estimate for K is consistent, its estimate for the column space of C is nearly as accurate as when samples are independent (i.e. Vg in (1) is a multiple of the identity) and that inference on the effects of interest is asymptotically equivalent to that when C is known. In Section 5 we analyze simulated multi-tissue gene expression data with a complex, gene-dependent correlation structure that demonstrates our method’s superior performance in both choosing K and estimating C. We lastly apply our method to a longitudinal DNA methylation dataset with measurements made on pairs of twins to identify CpGs with sex-dependent methylation levels in Section 6. An R package called CorrConf that implements our method to estimate K and C is freely available on GitHub. The proofs of all theorems are in the Supplement.
2. Notation and problem set-up
2.1. Notation
We define to be the vector of all ones, to be the identity matrix, [n] = {1, …, n} and [x]i to be the ith element of . For , we let [M]ij be the (i, j) element of M, and define PM and to be the orthogonal projections matrices onto and . If d = dim (M⊥) > 0, we define to be a matrix whose columns form an orthonormal basis for M⊥, i.e. . Lastly, we use the relation if the two random variables, vectors or matrices X and Y have the same distribution.
2.2. A description of and generative probability model for the data
Let be the measured expression or methylation of genomic unit g ∈ [p] in n potentially correlated samples, and let be the covariates of interest. When latent factors influence the data yg, we assume yg is generated as
(2a) |
(2b) |
(2c) |
The gth rows of β, L and E are βg, ℓg and eg, respectively, e1,…, ep are independent, and our goal is to estimate and perform inference on β1, …, βp. We assume C is a random matrix that is independent of E, and whose mean may be a function of X in our theoretical results in Section 4. Therefore, accounting for C can be interpreted as a way to alleviate both the dependence across genomic units and potential biases when estimating and performing inference on β1, …, βp. We postpone discussing the distributional assumptions on C until Section 4 because, other than C being independent of E, they play no part in the methodology presented in Section 3.
The observed matrices describe how the n samples are related, and Model (2b) is a typical model for the variance in mixed effect models in high throughput biological data (Martino et al., 2013; Tung et al., 2015; Chen et al., 2017; Knowles et al., 2018; McKennan et al., 2018). When b = 1 and B1 = In, Model (2) reduces to the model typically used to estimate β1, …, βp in high throughput data with independent samples (Leek and Storey, 2008; Sun et al., 2012; Lee et al., 2017; Wang et al., 2017; Gerard and Stephens, 2018; McKennan and Nicolae, 2019). We assume the unknown variance multipliers vg = (vg,1 … vg,b)T lie in the convex polytope
(3) |
where the matrices and are the equality and inequality constraints on vg. For example, we may know that certain multipliers must be larger than others, or that the sum of two sets of multipliers must be equal to ensure the marginal variances are the same.
In many applications, there are other observed covariates that may influence the response yg but whose effects are not of interest. In that case, the model for yg would be
(4) |
and we can get back to Model (2) by multiplying yg by , provided r does not grow with n:
Therefore, we assume that the only observed covariates are contained in X.
Evidently, C in Model (2) is not identifiable. However, we show in Section 4 that conditional on C, Im (C) is identifiable when β satisfies a modest sparsity assumption. Therefore, our primary goal is to develop a procedure to estimate and perform inference on β1, …, βp that only relies on an accurate estimate for Im (C). Our secondary goal is to simply estimate Im (C) when there are no covariates of interest, i.e. d = 0. Since accomplishing the former implies we can achieve the latter, we only consider the case when d ≥ 1.
We estimate Im (C) by separately computing, and then combining, estimates for the parts of Im (C) in Im (X) and X⊥. Let be a positive definite matrix. We estimate the part of Im (C) in Im (X) using the variation in y1, …, yp explainable by X:
(5a) |
(5b) |
(5c) |
Here is the naive weighted least squares estimator for βg that ignores C and Ω is the weighted least squares regression coefficient for the regression of C onto X. We show how to choose the appropriate G in Section 3.4. Next, we use the variation in y1, …, yp explainable by X⊥ to estimate the part of Im (C) in X⊥ :
(6a) |
(6b) |
(6c) |
Here C⊥ and lie in the space orthogonal to X, where the latter no longer depends on βg. Algorithm 1 below, which we call CBCV-CorrConf, provides a cursory description of how we use the objects defined in (5) and (6) to estimate Im (C) and subsequently β1,…,βp.
Algorithm 1 (CBCV-CorrConf).
For , use Y2 to obtain estimates and using a factor analysis procedure we call ICaSE (Iterative Correlation and Subspace Estimation) (Section 3.2).
Use Y2 and a cross-validation procedure we call CBCV (Correlated Bi-Cross Validation) to define , the estimate for dim {Im (C)} (Section 3.3).
Define G and estimate Ω using Y1 and the estimators , from steps 1 and 2. Let the estimate for C be . For given, we call this method of estimating C CorrConf (Section 3.4).
Use the design matrix [X C] to estimate β1 , …, βp with generalized least sqaures.
The estimates for Im (C⊥) and Im (C) are then Im (C⊥) and Im (C), where the estimate for C in step 3 follows from the fact that for any positive definite G. We show in Section 4 that our estimator in Step 4 only depends on Im (C). If b = 1, B1 = In and K were known, steps 1 and 3 with G = In are similar to methods used by Sun et al. (2012); Wang et al. (2017); Lee et al. (2017); McKennan and Nicolae (2019).
3. Estimating the factor dimension K and Im (C)
3.1. A computationally tractable model for the data
The generative model assumed in (2) is not conducive to estimating Im (C), since it requires jointly estimating Im (C) and all p covariance matrices V1,…,Vp. Instead, we use a simpler, but incorrect, model that assumes to estimate K, L and Im (C):
(7) |
where we introduce δ2 so that V is scale-invariant. We define δ2 in terms of the determinant for reasons discussed in Section 3.3.
Let and V* be the values of δ2 and V that, conditional on C, minimize the KL-divergence between the data generating model in (2) and the incorrect model in (7). Then
(8) |
In Sections 3.2, 3.3 and 3.4 we show that, quite remarkably, we only need to estimate V* and not V1, …,Vp to get accurate estimates for Im (C).
Nevertheless, without any assumptions on V*, this is a challenging, if not an intractable, problem. However, since for all g = 1,…, p and B1,…, Bb are known, then by (8), where
Therefore, we need only estimate τ* to estimate V*, and subsequently Im (C).
3.2. ICaSE: an iterative factor analysis to estimate Im (C⊥)
Here we present the algorithm ICaSE, a method to estimate Im (C⊥), and τ* using Y2. Recall from (6) that the mean of Y2 is and is not dependent on X. Let , which implies and . Therefore,
Since W* is not a multiple of the identity, the span of the first K eigenvectors of S2 will not be an accurate estimate for Im (C⊥), in general. If W* were known, we could instead first estimate as the span of the first K eigenvectors of , and then rescale the estimate by to estimate Im (C⊥). On the other hand, if Im (C⊥) were known, one could easily estimate and τ* using restricted maximum likelihood. ICaSE iterates between these two steps to estimate Im (C⊥), and τ*.
It remains to show how to find a starting point for and τ* that avoids incorporating signal from the random effect into the estimate for Im (C⊥), as imprudent starting points often beget biased estimates for K and Im (C⊥). We Y2 due to C⊥ from the random effect by employing a separate the variation in “warm start” technique often used to solve penalized regression problems. First, we initialize our estimates for and τ* assuming K = 0 (i.e. ). We then use the estimate for τ* when we assume dim {Im (C⊥)} = k − 1 as the starting point when we assume dim {Im (C⊥)} = k. This ensures that we attribute as much variability as possible to the random effect and only ascribe variability to the latent covariates if the observed signal is not amenable to the model for the variance. Algorithm 2 enumerates these steps in ICaSE.
Algorithm 2 (ICaSE).
Let for .
For k = 0, estimate and τ* by maximum likelihood using the model under the restriction that det {W (τ)} = 1 and δ2 τ ∈ Θ.
Let W = W (τ). For k ≥ 1 and given estimates and τ for and τ*, estimate W−1/2 C⊥ as the first k right singular vectors of Y2 W−1/2. Re-scale the estimate by (n − d)1/2 W1/2 to get an estimate for .
- Given C⊥, obtain and τ by restricted maximum likelihood using the model
Iterate between steps 1 and 2 and stop on step 1 of the second iteration. Repeat this for k = 1, …, Kmax, using τ, obtained when dim {Im (C⊥)} = k − 1 as the starting point for step 1 when dim {Im (C⊥)} = k.
Remark 1.
If b = 1 and is a scalar multiple of the first k right singular vectors of Y2 for each k ∈ [Kmax].
Evidently, iterating between steps 1 and 2 is not explicitly maximizing an objective function. However, Theorem 1 in Section 4 proves that ICaSE’s estimates for Im (C⊥) are as accurate as those obtained from principal component analysis when samples are independent.
3.3. Correlated Bi-Cross Validation to estimate K
Here we use Y2 and a cross-validation paradigm developed in Owen and Wang (2016) to estimate K = dim {Im (C)}. Besides having substantially shorter computation time and providing a less variable estimate for K, what differentiates our method from the one developed in Owen and Wang (2016) is we provide an approach amenable to dependent data, while their method is only valid for independent data.
Our primary concern is ensuring our estimates for K are not biased by the correlations in the data (Hastie et al., 2009). An equivalent concern is we want to avoid including factors arising from the random effect in our estimate for K. To describe our procedure, we assume for notational convenience that where the rows e1, …, ep of E are independent and eg ~ Nn (0,Vg) for all g ∈ [p]. Algorithm 3 provides an outline of the algorithm we use to estimate K, which we call Correlated Bi-Cross Validation (CBCV).
Algorithm 3 (CBCV).
Randomly partition the rows of Y (i.e. genes) into F = O (1) folds and let f ∈ [F]. Without loss of generality, assume
where and are the training and test sets.
Let . Obtain and V(− f) from Y(− f) using Algorithm 2.
- Let and . Define the loss for this fold, dimension pair as the leave-one-out cross validation loss:
where and are the ith columns of and , respectively. Lf,(− i) is the ordinary least squares regression coefficient from the regression of onto , where and are submatrices of and with the ith columns removed.(9) Repeat this for folds f = 1,…, F and k = 0,1,…, K max and set to be
(10) |
Since (9) is the leave one out cross validation squared loss using the design matrix and the scaled data , it only depends on Im (C) and V(− f).
Scaling by in (9) is sensible because it places more importance on correctly estimating the portion of C not captured by the model for the residual variance. However, unless proper care is taken, this scaled loss function would underestimate K simply because the estimated residual variance is larger for underspecified models. The restriction that alleviates this issue by making the loss function scale-invariant.
To understand why CBCV gives accurate estimates for K, consider the expected leave-one-out squared error for a particular fold, dimension pair, where is the ith row of :
In standard cross validation, V(− f) = In, meaning the residual variance term (II) would be constant for all k = 0,1,…, Kmax and the cross-term (III) would be 0, since Lf,(− i) would be independent of the ith column of Yf. The minimizer of the squared bias term (I) would then also minimize the expected cross validated error. However, since we must account for the correlation between samples, we now need to ensure that I + II is minimized when and that the cross-term does not contribute to the loss. The restriction that det {V(− f) } = 1 helps ensure this is the case, since , where the inequality holds with equality if and only if V(− f) = Vf * by Jensen’s Inequality. Since an accurate estimate of Im (C) begets an accurate estimate of V(− f)* ≈f *, the minimizer of I + II, and also I + II + III if III ≈ 0, should be close to K. We make this rigorous and prove the consistency of in Theorem 2 in Section 4.
3.4. De-biasing estimates for the main effect in correlated data
Recall from (5b) that for some positive definite , , which helps quantify the proportion of the variability of C explainable by X. While C⊥ enables one to estimate ℓg and Vg, Ω allows us to estimate the effect of X on yg while controlling for the latent factors C. This helps make results reproducible.
To estimate Ω, we must first specify a suitable weighting matrix G. Section 3.1 suggests that a reasonable choice would be , where τ is computed using Algorithm 2. From here on out, we assume K is known and define , Ω and Y1 in (5) by setting G = V.
Our strategy for estimating Ω is to regress Y1 onto the estimate for L obtainable after completing Algorithm 2, where for . One possible estimator for Ω is the ordinary least squares estimate . When samples are independent, this strategy is similar to those used in Sun et al. (2012); Houseman et al. (2014); Lee et al. (2017); Wang et al. (2017); Fan and Han (2017), although the exact estimators differ slightly. However, McKennan and Nicolae (2019) showed that Ωnaive tends to underestimate Ω when samples are independent because L is a noisy estimate for the design matrix L. Therefore, we use the following de-biased estimate of Ω, which is analogous to the estimator used in McKennan and Nicolae (2019) when samples are assumed independent:
(11) |
The term in (11) removes the bias in LT L and reduces to the bias correction used in McKennan and Nicolae (2019) when b = 1 and B1 = In.
Lastly, we can express C as . Therefore, our estimate for C is
(12) |
We subsequently estimate Vg and βg for all g ∈ [p] with restricted maximum likelihood followed by generalized least squares using the design matrix [X C].
4. Theoretical justification of CBCV-CorrConf
For all theoretical results, we assume d, b, Kmax = O (1), and y1,…, yp are generated by (2), where e1,…, ep are independent. The latter assumption is based off of the repeated observation in experimental genetic data that, for the purposes of marginal testing, it is sufficient to assume the dependence between genomic units is driven by a relatively small number of latent variables (Leek and Storey, 2007; Maksimovic et al., 2015; Gerard and Stephens, 2018). It is also a common assumption among the literature that motivates their models with high throughput biological data (Leek and Storey, 2008; Sun et al., 2012; Lee et al., 2017; Wang et al., 2017; Dobriban and Owen, 2019). We also provide simulation evidence in Section S1.3 of the Supplement that suggests our method is robust to dependence among e1,…, ep.
Assumption 1.
and are observed, non-random matrices. Let c1 > 1 be a constant. Define to be [M]ij = n−1Tr (Bi Bj) for all i, j ∈ [b]. For any matrix D, let D(− i) be the sub-matrix of D with the ith row removed. Assume:
The random matrix C is independent of E, where for all i ∈ [n] and k ∈ [K]. Further, , , , for all j ∈ [b] and
Let and . Then for all g ∈ [p], and the first K eigenvalues λ1 > … >λK > λK+1 = 0 of np−1 L Ψ LT satisfy .
, for all g ∈ [p] and j ∈ [b]. Further, as n, p → ∞ and .
- for all k ∈ [K], λ1 / λK ≤ c1 and one of the following holds:
p is a non-decreasing function of n such that n / p, n3/2 / (p λK) → 0 as n → ∞.
Remark 2.
We prove V1,…,Vp ≻ 0 and L Ψ LT are identifiable under Assumption 11 in Proposition S1 in Section S3.1 in the Supplement. The set ΘL is a classic way to identify parameters in factor models when for all g ∈ [p] (Anderson and Rubin, 1956; Wang et al., 2017).
Remark 3.
The assumptions on C imply and , where Ψij is the covariance between the ith and jth rows of C.
The condition on M ensures v1,…, vp are identifiable, and one can easily verify the condition on ║Bj║2 in data from McKennan et al. (2018) and Martino et al. (2013) discussed in Sections 1 and 6, as well as in the simulated multi-tissue data from Section 5. We provide additional data examples to illustrate the ubiquity of the assumption on ║Bj║2 in Section S4.1 of the Supplement. The eigenvalues λ1,…, λK quantify the average variation in y1,…, yp due to C that can be distinguished from that due to X, and we give general sufficient conditions under which in Lemma S1 in the Supplement. Item 5 is standard when for all g ∈ [p] (Wang et al., 2017; McKennan and Nicolae, 2019).
Assumption 2.
Let c2 > 1 be a constant not dependent on n or p. The estimates and τ from step 2 in Algorithm 2 are such that lies in the convex set
where c1 was defined in Assumption 1.
This makes the residual variance parameter space compact and is analogous to Assumption D in Bai and Li (2012) and Assumption 2 in Wang et al. (2017). It is also a standard assumption in likelihood theory (Wald, 1949; Ferguson, 1996; Douc et al., 2004). We now state Theorem 1.
Theorem 1 (Accuracy of ICaSE).
Suppose Assumptions 1 and 2 hold and we apply Algorithm 2 for k = 1,…, Kmax , where K ≤ K max. Then the estimates for , τ* and C⊥ satisfy
where and τ depend on k. Further, and the estimates , τ are invariant of the choice of QX.
Theorem 1 implies that Im (C⊥) is estimated well when k = K and its column space is approximately a subspace of Im (C⊥) when we overestimate K. This result is quite remarkable because besides the additional factor , this is the same rate obtained from principal components analysis when the samples are independent (McKennan and Nicolae, 2019). Theorem 1 makes no assumption on the starting points for and τ (other than the REML estimates are estimated in Θ*), which proves the efficacy of our warm-start technique.
Theorem 2 (Consistency of CBCV).
Let be as defined in (10). Suppose the assumptions of Theorem 1 hold, for some constant ϵ > 0 and we sample QX uniformly over the space of all orthonormal bases for X⊥. Then .
Remark 4.
Let be a non-random matrix such that . Then QX = Q* M, where is independent of C and E, and is sampled uniformly over the space of unitary matrices. This guarantees the maximum leverage score of is oP (1) as n → ∞, and is a common technique to uniformize the leverage scores of a matrix (Mahoney, 2011).
This theorem shows that recovering K is easier when the sample size n is larger, since λ1,…, λK tend to grow with n. This is why it is advisable to use all of the available samples to estimate K, rather than partitioning the data to alleviate correlation, as discussed in Section 1.
The requirement that is tight, as Owen and Wang (2016) demonstrate that is the lower limit of detection when b = 1 and B1 = In. This lower limit of detection is smaller than parallel analysis’, which often fails to recover moderate to small factors (Owen and Wang, 2016). We discuss this in more detail after the statement of Theorem S2 in Section S4 of the Supplement. Theorem 2 is also, to the best of our knowledge, the first result proving that bi-cross validation-like methods are consistent. The authors of Owen and Perry (2009); Owen and Wang (2016) only showed their estimates for K minimize the expected loss when samples are independent.
Assumption 3.
, for some constant c3 > 0 , and for , as n, p→ ∞.
Remark 5.
We prove β is identifiable under Assumptions 11–2, 15 and the sparsity assumption on s in Proposition S2 in the Supplement. Other than the conditions on and S in Assumptions 1 and 3, we make no assumptions about the relationship between C and X.
This provides an explicit relationship between the maximum allowable sparsity on the main effect and the informativeness of the data Y for estimating C: the more signal in , the part of LCT unequivocally distinguishable from βXT, the less stringent Assumption 3 becomes. This sparsity condition is consistent with many modern genetic studies (Liu et al., 2016; Morales et al., 2016; Yang et al., 2017; Zhang et al., 2018), and is the same sparsity assumed in existing theoretical work when b = 1 and B1 = In (Wang et al., 2017; McKennan and Nicolae, 2019). We show through simulation in Section 5 that we can accurately recover C even when this assumption is violated.
Theorem 3 (Inference on β)
Let g ∈ [p]. Suppose Assumptions 1, 2 and 3 hold and we estimate C according to (12). Let and βg be the restricted maximum likelihood (REML) and generalized least squares (GLS) estimates for vg and βg using the design matrix [X C]. Then
as n → ∞ , where Z ~ Nd (0, Id) and . Further, as n → ∞, where Mg is the variance of the GLS estimate for βg when C and Vg are known.
Remark 6.
The REML and GLS estimates V g, βg and Mg only depend on C through Im (C). We prove that conditional on C, Im (C) and Mg are identifiable for all g ∈ [p] in Proposition S3 in Section S5.2 in the Supplement.
5. Simulated multi-tissue gene expression data analysis
5.1. Simulation setup and parameters
We simulated the expression of p = 15, 000 genes from 50 individuals across three tissues with a complicated tissue-by-tissue correlation structure to compare our method against other state of the art methods designed to estimate K and β1,…, βp. We include comparisons of our method to estimate C⊥ in Section S1.1 in the Supplement. We first randomly chose 25 individuals to be in the treatment group and set X ∈ {0,1}n to be the treatment status for the n = 150 samples and Z = 150 ⊗ I3 to be the tissue-specific intercept. For each g ∈ [p], we let Vg = I50 ⊗ Mg be the gene-specific covariance matrix for gene g and simulated in each dataset so the average correlation matrix across all p genes was as given in Table 1.
Table 1.
The correlation matrix corresponding to the covariance matrix .
Tissue 1 | Tissue 2 | Tissue 3 | |
---|---|---|---|
Tissue 1 | 1 | 0.72 | 0.58 |
Tissue 2 | 0.72 | 1 | 0.80 |
Tissue 3 | 0.58 | 0.80 | 1 |
Specifically, we assumed tissues two and three were more similar to one another than to the first and for each g ∈ [p], set Mg to be
The constants , and were simulated from Gamma distributions with means 0.8, 1.25, 0.4, 0.75, 1 and 0.2, respectively, each with coefficient of variation equal to 0.2. We subsequently re-scaled these parameters so that . This complex tissue-by-tissue covariance is amenable to the variance model assumed in (2b), since
(13) |
where has a 1 in the rth and sth coordinates and 0 everywhere else and
We next simulated data with K = 10 latent factors. The parameters X, Z, K and α, defined below, were fixed across all simulations. For , we simulated , the expression of gene g in the n (individual, tissue) pairs, as follows:
(14) |
Here δ0 is the point mass at 0 and T8 is the t-distribution with eight degrees of freedom, and was chosen to emulate data with heavy tails. We chose to simulate a non-sparse main effect β to show that we can violate Assumption 3 and still do inference that is just as accurate as when C is known. The re-scaling of Ξ was simply to make the diagonal elements of n p−1 LT L approximately equal to the eigenvalues λ1,…, λ10 defined in Assumption 1. This re-scaling also caused λ1,…, λ10 to shrink by a factor of 1.5, on average, thus making it harder to recover K and C. The constant α was chosen so that explained approximately 30% of the variability in , on average. We defined α by scaling X, C and Z by because this is what one would do in generalized least squares when the covariance of the observations is V*. The values for πk, ηk and the resulting λ1,…,λ10 are provided in Table 2.
Table 2.
The πk and ηk values used to simulated L and the resulting average λk (k = 1, …, 10).
Factor number (k) | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
---|---|---|---|---|---|---|---|---|---|---|
π k | 0 | 0.45 | 0.60 | 0.71 | 0.79 | 0.85 | 0.90 | 0.92 | 0.94 | 0.96 |
η k | 1 | 0.4 | 0.4 | 0.4 | 0.4 | 0.4 | 0.4 | 0.4 | 0.4 | 0.5 |
λ k | 143 | 12.7 | 9.5 | 6.6 | 4.8 | 3.3 | 2.5 | 1.8 | 1.4 | 1.4 |
For each simulated dataset, we let X be the covariate of interest and the tissue specific intercepts, Z, be nuisance covariates and estimated K using CBCV with F = 3 folds, C⊥ and V* using ICaSE and C using (12). We then estimated Vg and βg for each g ∈ [p] using restricted maximum likelihood and generalized least squares, respectively. We include additional results for data simulated such that conditional on C, expression across genes was correlated in Section S1.3 of the Supplement.
5.2. Comparison of estimators for K and β
Since, as far as we are aware, our methods for estimating K and C are the first methods that account for correlation among samples in high throughput biological data, we compared our estimates for K and β with state of the art methods designed for high throughput biological data with independent samples. We first assessed our estimates for K in 100 simulated datasets and compared them with three widely used methods: the method proposed in Bai and Ng (2002), parallel analysis (Buja and Eyuboglu, 1992) and bi-cross validation (Owen and Wang, 2016). The results are given in Figure 1. We could not compare our method to the deterministic version of parallel analysis designed for data with independent samples proposed in Dobriban and Owen (2019), since the authors did not provide R code to implement their method. However, their simulations show that their method tends to select at least as many factors as parallel analysis.
Fig. 1.
Estimates for K = 10 in 100 simulated datasets using our proposed method (CBCV), the method proposed in Bai and Ng (2002) (Bai & Ng), parallel analysis (PA) and bi-cross validation (BCV).
The fact that CBCV recovers K in all simulated datasets is quite remarkable given that we simulate data with heavy tails and (see Theorem 2). We discuss the behavior of CBCV-CorrConf when λK is smaller than in Sections S1.3 and S4.2 of the Supplement. The method proposed in Bai and Ng (2002) severely underestimates K because it is only able to recover latent factors with overtly large effects (Owen and Wang, 2016). On the other hand, bi-cross validation and parallel analysis overestimate K because both methods are treating the high dimensional random effect as part of the low dimensional effect . We discuss possible adjustments to ameliorate bi-cross validation’s and parallel analysis’ estimates for K in these simulated data in Section S1.2 of the Supplement. Suffice it to say that these adjustments either did not change the estimates, or caused them to consistently underestimate K.
We lastly estimated β via generalized least squares with the design matrix M = [X Z C], where C was estimated with CBCV followed by CorrConf (our estimator specified in (12)), BCconf (McKennan and Nicolae, 2019), Cate-RR (Wang et al., 2017), dSVA (Lee et al., 2017), IRW-SVA (Leek and Storey, 2008), RUV-2 (Gagnon-Bartsch and Speed, 2012), RUV-4 (Gagnon-Bartsch et al., 2013) and when C was known. BCconf, Cate-RR, dSVA, IRW-SVA, RUV-2 and RUV-4 were all applied assuming K = 10 was known. In order to make the estimation of V1,…,Vp computationally tractable, we modeled the simulated residuals e1,…, ep as
where Brs was defined in (13), and estimated v1,…, vp and Trs with restricted maximum likelihood for each of the eight methods using the estimated design matrix M. We then computed the P value for the null hypothesis βg = 0 for all g ∈ [p] by comparing the t-statistics to a t-distribution with n – 4 − K = n − 14 degrees of freedom, used these P values as input into q-value (Storey, 2001) to control the false discovery rate and deemed a gene as being differentially expressed across the two treatment conditions if its q-value was no greater than 0.2. We chose q-value to control the false discovery rate because this is the software most popular among biologists. Figure 2 plots the true false discovery proportion in 100 simulated datasets among genes with a q-value less than or equal to 0.2 for each of the eight methods. The false discovery proportion when we completely ignored C was uniformly greater than that of IRW-SVA’s.
Fig. 2.
The false discovery proportion (FDP) among genes with a q-value no greater than 0.2 in 100 simulated datasets. We randomly chose 90 genes with βg = 0 as control genes in RUV-2 and RUV-4. The FDP when we ignored C was uniformly greater than IRW-SVA’s.
The performance of our method, CorrConf, as the statement of Theorem 3 suggests, is nearly indistinguishable from the generalized least squares estimator when C is known, with nearly identical power. On the other hand, the other six methods tend to introduce more type I errors because their estimates for C do not account for the dependence between the residuals and therefore fail to recover Ω. When K is underestimated, these six methods perform even worse than when K = 10 is known, and we show in Section S1.3 in the Supplement that these methods still fail to control the false discovery rate and, due to the reduction in residual degrees of freedom, exhibit a decrease in power when K is overestimated to the extent suggested by bi-cross validation and parallel analysis in Figure 1.
6. Sex-specific DNA methylation in a longitudinal twin study
We next applied our method to identify sex-specific DNA methylation patterns from a longitudinal twin study using data previously published in Martino et al. (2013). The authors measured the DNA methylation of 10 monozygotic (MZ) and 5 dizygotic (DZ) Australian twin pairs (all DZ twins were both male or both female) at birth and 18 months on the Infinium HumanMethylation450 BeadChip platform in buccal epithelium, a relatively homogeneous tissue. After probe and sample quality filtering and data-normalization, the authors were left with p = 330, 168 methylation sites (CpGs) whose methylation was quantified in 29 male and 24 female (n = 53) samples as the difference between log-methylated and log-unmethylated probe intensity (see Martino et al. (2013) for all pre-processing steps). We then used our proposed method (CBCV-CorrConf), BCconf, Cate-RR, dSVA and IRW-SVA to identify CpGs whose methylation differs in males and females, which we refer to as sex-associated CpGs. We subsequently validated each method’s findings using sex-associated CpGs identified at birth in previous studies with substantially larger sample sizes. We did not compare our method with RUV-2 or RUV-4, since we did not have access to control CpGs.
We first show that we can write the covariance of the 53 observations at each CpG as a linear combination of six matrices. Let ym, t,a be the measured DNA methylation at an arbitrary CpG for twin t ∈{1, 2}, from mother m ∈ [15] at age a ∈{0,18}, where samples with different mothers were assumed to be independent and twin 1 and twin 2 from the same mother were assumed to be exchangeable. A preliminary analysis showed that the correlation between MZ and DZ twins’ methylomes was approximately the same at both ages, which is consistent with the observation that methylation patterns are in large part determined by environmental exposures (Galanter et al., 2017; Martin and Fry, 2018). Therefore, the 4 × 4 covariance matrix for ym = (ym,1,0 ym,1,18 ym,2,0 ym,2,18)T completely determined the n × n covariance matrix for each CpG. We show in Section S2 of the Supplement that by only making assumptions on the pairwise covariances, one can write Cov (ym) as
(15) |
where , . The variance multipliers are such that for t1 ≠ t2,
and lie in a convex polytope defined in (3). We derive these variance multiplier relationships in Section S2 of the Supplement.
Since there was no evidence that the difference in methylation between males and females changed from birth to 18 months, we assumed the methylation at each CpG was a linear combination of the subject’s age (birth or 18 months), sex and other unobserved factors to be estimated, where age was a nuisance covariate and sex was the phenotype of interest. We used CBCV with F = 5 folds and estimated K to be 2, and subsequently estimated V* and C with ICaSE and CorrConf. Our estimates for the six average variance multipliers were all greater than 0, the average residual variance at 18 months was 25% larger than that at birth and the correlation between methylation for twins at 18 months was nearly 20% larger than that at birth. This indicated this set of twin’ s methylomes tended to converge over the first 18 months of life, which is consistent with previous observations that one’s methylome reflects one’s environmental exposure history (Galanter et al., 2017; Martin and Fry, 2018).
We next computed each of the other four method’s estimates for C by first using each method’s default software to choose K: bi-cross validation, the default for Cate-RR and BCconf, or parallel analysis, the default for dSVA and IRW-SVA. These methods estimated K to be 4 and 15, respectively. The fact that both estimated the latent factor dimension to be greater than CBCV’s estimate of 2 is not surprising, as these methods tend to overestimate K when samples are correlated. We subsequently estimated C with each of the four methods using their software’s default settings.
We lastly estimated the effect due to sex on methylation and corresponding q-values to control the false discovery rate, exactly as we did for the simulated data in Section 5. We deemed a CpG a sex-associated CpG if its q-value in that method was no greater than 0.2. Since we did not know the ground truth, we used sex-associated CpGs identified at birth in Yousefi et al. (2015) and Maschietto et al. (2017) as a validation set to help judge the veracity of each method’s findings. Yousefi et al. (2015) and Maschietto et al. (2017) measured DNA methylation in umbilical cord blood on the Infinium HumanMethylation450 BeadChip platform in children born to 111 unrelated Brazilian and 71 unrelated Mexican American mothers, respectively. The authors of both studies measured and corrected for cord blood cellular composition and identified 2,355 and 1,928 sex-associated CpGs that were also among the 330,168 CpGs studied in Martino et al. (2013). Table 3 gives the fraction of sex-associated CpGs identified using C estimated with our method (CBCV-CorrConf), along with the other four methods, that are also among the 3,532 sex-associated CpGs identified in Maschietto et al. (2017) or Yousefi et al. (2015). Since the other four methods are not designed for dependent data, we discuss two adjustments designed to alleviate potential biases in their estimates for both K and C in Section S2 of the Supplement. The overlaps with the modified methods were at least as poor as those presented in Table 3.
Table 3.
The fraction of sex-associated CpGs identified using data from Martino et al. (2013) that are also one of the 3,532 sex-associated CpGs confidently identified in Yousefi et al. (2015) or Maschietto et al. (2017).
CBCV-CorrConf (K = 2) | BCconf (K = 4) | Cate-RR (K = 4) | dSVA (K = 15) | IRW-SVA (K = 15) |
---|---|---|---|---|
38% (278/729) | 23% (424/1839) | 20% (474/2404) | 19% (487/2517) | 28% (341/1240) |
While it may be the case that most of BCconf, Cate-RR, dSVA and IRW-SVA are actual sex-associated CpGs, the results in Table 3 mirror the trends observed in Figure 2. That is, while these four methods nominally identify more sex-associated CpGs, we are less confident in their results because their estimates for C reduce the residual variance but likely do not suitably account for the variation in sex explainable by C, which makes their results less reproducible.
These results also highlight the importance of the choice of K. Estimating K with CBCV, and cross-validation in general, tends to yield more reproducible results because we only include a latent factor if prediction performs suitably well on new, held-out data. When we applied all five methods with K = 2, BCconf, Cate-RR and dSVA performed similarly with overlaps no greater than 30%. However, 272 out of IRW-SVA’s 662 sex-associated CpGs (41%) were in the validation set, which is nearly identical to CorrConf’s results in Table 3. Similarly, when we set K = 4 for all methods, dSVA performed nearly identically to Cate-RR, whereas CorrConf and IRW-SVA had overlaps of 27% and 26%, respectively, and both ostensibly identified approximately 1,500 sex-associated CpGs. This similarity arises because unlike BCconf, Cate-RR and dSVA, IRW-SVA circumvents estimating Ω with a noisy estimate for L and estimates C by performing factor analysis directly on Y, restricted to the genomic units that show little marginal correlation with X. In fact, we use the proof of Theorem 1 in Sections S1.4 and S2 in the Supplement to show that under certain conditions on C, V* and λK, which appear to be satisfied in these data, IRW-SVA can accurately recover C even when residuals are correlated. We believe K = 2 is the most appropriate choice of K for this dataset because CorrConf’s estimate for C appears to explain enough of the variance in methylation to achieve reasonable power, while also accurately recovering Ω to control for false discoveries and ensuring that the results are reproducible.
7. Discussion
To the best of our knowledge, we have developed the first method to account for latent factors in high throughput biological data with correlated samples. We proved that our estimate for K is consistent and that our estimate for Im (C) is accurate enough so that inference on the main effects β1,…, βp is just as accurate as when C is known. We also demonstrated our method’s finite sample properties by analyzing complex, multi-tissue simulated gene expression data, and used a real longitudinal DNA methylation data from a twin study to show our method tends to give more reproducible results compared to existing methods.
Our proposed procedure is certainly not a panacea for data with arbitrary correlation structure, and relies on the residual variance Vg being a linear combination of known matrices. Data with more complex, non-linear sample correlation structure may not be amenable to (2b), since a linear combination of p non-linear functions will not necessarily have an apriori known functional form. This could be an interesting area of future research.
Supplementary Material
Acknowledgments
* The authors gratefully acknowledge the NIH for their support of this research, which was funded by NIH grants R01-HL129735 and R01-MH101820
Contributor Information
Chris McKennan, Department of Statistics, University of Pittsburgh.
Dan Nicolae, Department of Statistics, University of Chicago.
References
- Anderson TW and Rubin H (1956), Statistical inference in factor analysis, in ‘Proceedings of the Third Berkeley Symposium on Mathematical Statistics and Probability, Volume 5: Contributions to Econometrics, Industrial Research, and Psychometry’, University of California Press, Berkeley, California, pp. 111–150. [Google Scholar]
- Bai J (2009), ‘Panel data models with interactive fixed effects’, Econometrica 77(4), 1229–1279. [Google Scholar]
- Bai J and Li K (2012), ‘Statistical analysis of factor models of high dimension’, The Annals of Statistics 40(1), 436–465. [Google Scholar]
- Bai J and Ng S (2002), ‘Determining the number of factors in approximate factor models’, Econometrica 70(1), 191–221. [Google Scholar]
- Baumgart M, Priebe S, Groth M, Hartmann N, Menzel U, Pandolfini L, Koch P, Felder M, Ristow M, Englert C, Guthke R, Platzer M and Cellerino A (2016), ‘Longitudinal RNA-Seq analysis of vertebrate aging identifies mitochondrial complex I as a small-molecule-sensitive modifier of lifespan’, Cell Systems 2(2), 122–132. [DOI] [PubMed] [Google Scholar]
- Breton CV, Marsit CJ, Faustman E, Nadeau K, Goodrich JM, Dolinoy DC, Herbstman J, Holland N, LaSalle JM, Schmidt R, Yousefi P, Perera F, Joubert BR, Wiemels J, Taylor M, Yang IV, Chen R, Hew KM, Freeland DMH, Miller R and Murphy SK (2017), ‘Small-magnitude effect sizes in epigenetic end points are important in children’s environmental health studies: The children’s environmental health and disease prevention research center’s epigenetics working group’, Environmental Health Perspectives 125(4), 511–526. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Buja A and Eyuboglu N (1992), ‘Remarks on parallel analysis’, Multivariate Behavioral Research 27(4), 509–540. [DOI] [PubMed] [Google Scholar]
- Chen LS, Wang J, Wang X and Wang P (2017), ‘A mixed-effects model for incomplete data from labeling-based quantitative proteomics experiments’, The Annals of Applied Statistics 11(1), 114–138. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dobriban E and Owen AB (2019), ‘Deterministic parallel analysis: an improved method for selecting factors and principal components’, Journal of the Royal Statistical Society Series B 81(1), 163–183. [Google Scholar]
- Douc R, Moulines É and Rydén T (2004), ‘Asymptotic properties of the maximum likelihood estimator in autoregressive models with markov regime’, The Annals of Statistics 32(5), 2254–2304. [Google Scholar]
- Fan J and Han X (2017), ‘Estimation of the false discovery proportion with unknown dependence’, Journal of the Royal Statistical Society: Series B 79(4), 1143–1164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ferguson T (1996), A Course in Large Sample Theory, Chapman & Hall. [Google Scholar]
- Gagnon-Bartsch JA, Jacob L and Speed TP (2013), Removing unwanted variation from high dimensional data with negative controls, Technical report, UC Berkeley. [Google Scholar]
- Gagnon-Bartsch JA and Speed TP (2012), ‘Using control genes to correct for unwanted variation in microarray data’, Biostatistics 13(3), 539–552. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Galanter JM, Gignoux CR, Oh SS, Torgerson D, Pino-Yanes M, Thakur N, Eng C, Hu D, Huntsman S, Farber HJ, Avila PC, Brigino-Buenaventura E, LeNoir MA, Meade K, Serebrisky D, Rodríguez-Cintrón W, Kumar R, Rodríguez-Santana JR, Seibold MA, Borrell LN, Burchard EG and Zaitlen N (2017), ‘Differential methylation between ethnic sub-groups reflects the effect of genetic ancestry and environmental exposures’, eLife 6, e20532. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gerard D and Stephens M (2018), ‘Empirical Bayes shrinkage and false discovery rate estimation, allowing for unwanted variation’, Biostatistics. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Consortium GTEx (2017), ‘Genetic effects on gene expression across human tissues’, Nature 550. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hastie T, Tibshirani R and Friedman J (2009), The Elements of Statistical Learning: Data Mining, Inference, and Prediction (Second edition), Springer. [Google Scholar]
- Houseman EA, Molitor J and Marsit CJ (2014), ‘Reference-free cell mixture adjustments in analysis of DNA methylation data’, Bioinformatics 30(10), 1431–1439. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jaffe AE and Irizarry RA (2014), ‘Accounting for cellular heterogeneity is critical in epigenome-wide association studies’, Genome Biology 15(31). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Knowles DA, Burrows CK, Blischak JD, Patterson KM, Serie DJ, Norton N, Ober C, Pritchard JK, Gilad Y and McVean G (2018), ‘ Determining the genetic basis of anthracycline-cardiotoxicity by response QTL mapping in induced cardiomyocytes’, eLife 7, e33480. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee S, Sun W, Wright FA and Zou F (2017), ‘An improved and explicit surrogate variable analysis procedure by coefficient adjustment’, Biometrika 104(2), 303–316. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Leek JT, Scharpf RB, Bravo HC, Simcha D, Langmead B, Johnson WE, Geman D, Baggerly K and Irizarry RA (2010), ‘Tackling the widespread and critical impact of batch effects in high-throughput data’, Nature reviews. Genetics 11(10), 10.1038/nrg2825. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Leek JT and Storey JD (2007), ‘Capturing heterogeneity in gene expression studies by surrogate variable analysis’, PLOS Genetics 3(9), 1724–1735. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Leek JT and Storey JD (2008), ‘A general framework for multiple testing dependence’, Proceedings of the National Academy of Sciences 105(48), 18718–18723. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li D, Qian J and Su L (2016), ‘Panel data models with interactive fixed effects and multiple structural breaks’, Journal of the American Statistical Association 111(516), 1804–1819. [Google Scholar]
- Liu C, Marioni RE, Hedman Å, Pfeiffer L, Tsai P-C, Reynolds LM, Just AC, Duan Q, Boer CG, Tanaka T, Elks CE, Aslibekyan S, Brody JA, Kühnel B, Herder C, Almli LM, Zhi D, Wang Y, Huan T, Yao C, Mendelson MM, Joehanes R, Liang L, Love S-A, Guan W, Shah S, McRae AF, Kretschmer A, Prokisch H, Strauch K, Peters A, Visscher PM, Wray NR, Guo X, Wiggins KL, Smith AK, Binder EB, Ressler KJ, Irvin MR, Absher DM, Hernandez D, Ferrucci L, Bandinelli S, Lohman K, Ding J, Trevisi L, Gustafsson S, Sandling JH, Stolk L, Uitterlinden AG, Yet I, Castillo-Fernandez JE, Spector TD, Schwartz JD, Vokonas P, Lind L, Li Y, Fornage M, Arnett DK, Wareham NJ, Sotoodehnia N, Ong KK, van Meurs JBJ, Conneely KN, Baccarelli AA, Deary IJ, Bell JT, North KE, Liu Y, Waldenberger M, London SJ, Ingelsson E and Levy D (2016), ‘A DNA methylation biomarker of alcohol consumption’, Molecular Psychiatry 23, 422 EP –. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lu X and Su L (2016), ‘Shrinkage estimation of dynamic panel data models with interactive fixed effects’, Journal of Econometrics 190(1), 148–175. [Google Scholar]
- Mahoney MW (2011), Randomized Algorithms for Matrices and Data, now.
- Maksimovic J, Gagnon-Bartsch JA, Speed TP and Oshlack A (2015), ‘ Removing unwanted variation in a differential methylation analysis of illumina humanmethylation450 array data’, Nucleic acids research 43(16), e106–e106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Martin EM and Fry RC (2018), ‘Environmental influences on the epigenome: Exposure- associated DNA methylation in human populations’, Annual Review of Public Health 39(1), 309–333. [DOI] [PubMed] [Google Scholar]
- Martino D, Loke YJ, Gordon L, Ollikainen M, Cruickshank MN, Saffery R and Craig JM (2013), ‘Longitudinal, genome-scale analysis of DNA methylation in twins from birth to 18 months of age reveals rapid epigenetic change in early life and pair-specific effects of discordance’, Genome Biology 14(5), R42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Maschietto M, Bastos LC, Tahira AC, Bastos EP, Euclydes VLV, Brentani A, Fink G, de Baumont A, Felipe-Silva A, Francisco RPV, Gouveia G, Grisi SJFE, Escobar AMU, Moreira-Filho CA, Polanczyk GV, Miguel EC and Brentani H (2017), ‘Sex differences in DNA methylation of the cord blood are related to sex-bias psychiatric diseases’, Scientific Reports 7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McKennan C, Naughton K, Stanhope C, Kattan M, O’Connor G, Sandel M, Visness C, Wood R, Bacharier L, Beigelman A, Lovisky-Desir S, Togias A, Gern J, Nicolae D and Ober C (2018), ‘Longitudinal studies at birth and age 7 reveal strong effects of genetic variation on ancestry-associated DNA methylation patterns in blood cells from ethnically admixed children’, bioRxiv. [Google Scholar]
- McKennan C and Nicolae D (2019), ‘Accounting for unobserved covariates with varying degrees of estimability in high-dimensional biological data’, Biometrika 106(4), 823–840. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morales E, Vilahur N, Salas LA, Motta V, Fernandez MF, Murcia M, Llop S, Tardon A, Fernandez-Tardon G, Santa-Marina L, Gallastegui M, Bollati V, Estivill X, Olea N, Sunyer J and Bustamante M (2016), ‘ Genome-wide DNA methylation study in human placenta identifies novel loci associated with maternal smoking during pregnancy’, International Journal of Epidemiology 45(5), 1644–1655. [DOI] [PubMed] [Google Scholar]
- Owen AB and Perry PO (2009), ‘Bi-cross-validation of the SVD and the nonnegative matrix factorization’, Annals of Applied Statistics 3(2), 564–594. [Google Scholar]
- Owen AB and Wang J (2016), ‘Bi-cross-validation for factor analysis’, Statistical Science 31(1), 119–139. [Google Scholar]
- Peixoto L, Risso D, Poplawski SG, Wimmer ME, Speed TP, Wood MA and Abel T (2015), ‘How data analysis affects power, reproducibility and biological insight of RNA-seq studies in complex datasets’, Nucleic Acids Research 43(16), 7664–7674. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Storey JD (2001), ‘A direct approach to false discovery rates’, Journal of the Royal Statistics Society Series B 63(3), 479–498. [Google Scholar]
- Su L and Ju G (2018), ‘Identifying latent grouped patterns in panel data models with interactive fixed effects’, Journal of Econometrics 206(2), 554–573. [Google Scholar]
- Sun Y, Zhang NR and Owen AB (2012), ‘Multiple hypothesis testing adjusted for latent variables, with an application to the agemap gene expression data’, The Annals of Applied Statistics 6(4), 1664–1668. [Google Scholar]
- Tung J, Zhou X, Alberts SC, Stephens M, Gilad Y and Dermitzakis ET (2015), ‘The genetic architecture of gene expression levels in wild baboons’, eLife 4, e04729. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wald A (1949), ‘Note on the consistency of the maximum likelihood estimate’, The Annals of Mathematical Statistics 20(4), 595–601. [Google Scholar]
- Wang J, Zhao Q, Hastie T and Owen AB (2017), ‘Confounder adjustment in multiple hypothesis testing’, The Annals of Statistics 45(5), 1863–1894. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang IV, Pedersen BS, Liu AH, O’Connor GT, Pillai D, Kattan M, Misiak RT, Gruchalla R, Szefler SJ, Khurana Hershey GK, Kercsmar C, Richards A, Stevens AD, Kolakowski CA, Makhija M, Sorkness CA, Krouse RZ, Visness C, Davidson EJ, Hennessy CE, Martin RJ, Togias A, Busse WW and Schwartz DA (2017), ‘The nasal methylome and childhood atopic asthma’, Journal of Allergy and Clinical Immunology 139(5), 1478–1488. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yao C, Li H, Shen X, He Z, He L and Guo Z (2012), ‘Reproducibility and concordance of differential dna methylation and gene expression in cancer’, PLOS ONE 7(1), e29686–. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yousefi P, Huen K, Davé V, Barcellos L, Eskenazi B and Holland N (2015), ‘Sex differences in DNA methylation assessed by 450K beadchip in newborns’, BMC Genomics 16(1). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang X, Biagini Myers JM, Burleson J, Ulm A, Bryan KS, Chen X, Weirauch MT, Baker TA, Butsch Kovacic MS and Ji H (2018), ‘Nasal DNA methylation is associated with childhood asthma’, Epigenomics 10(5), 629–641. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.