Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2023 Feb 1.
Published in final edited form as: J R Stat Soc Series B Stat Methodol. 2021 Nov 16;84(1):149–173. doi: 10.1111/rssb.12479

Transfer Learning for High-Dimensional Linear Regression: Prediction, Estimation and Minimax Optimality

Sai Li 1, T Tony Cai 2, Hongzhe Li 3
PMCID: PMC8863181  NIHMSID: NIHMS1755759  PMID: 35210933

Abstract

This paper considers estimation and prediction of a high-dimensional linear regression in the setting of transfer learning where, in addition to observations from the target model, auxiliary samples from different but possibly related regression models are available. When the set of informative auxiliary studies is known, an estimator and a predictor are proposed and their optimality is established. The optimal rates of convergence for prediction and estimation are faster than the corresponding rates without using the auxiliary samples. This implies that knowledge from the informative auxiliary samples can be transferred to improve the learning performance of the target problem. When the set of informative auxiliary samples is unknown, we propose a data-driven procedure for transfer learning, called Trans-Lasso, and show its robustness to non-informative auxiliary samples and its efficiency in knowledge transfer. The proposed procedures are demonstrated in numerical studies and are applied to a dataset concerning the associations among gene expressions. It is shown that Trans-Lasso leads to improved performance in gene expression prediction in a target tissue by incorporating data from multiple different tissues as auxiliary samples.

1. Introduction

Modern scientific research is characterized by massive and diverse data sets. It is of significant interest to integrate different data sets to make a more accurate prediction and statistical inference. Given a target problem to solve, transfer learning (Torrey and Shavlik, 2010) aims at transferring the knowledge from different but related samples to improve the learning performance of the target problem. A typical example of transfer learning is that one can improve the accuracy of recognizing cars by using not only the labeled data for cars but some labeled data for trucks (Weiss et al., 2016). Besides classification, another important transfer learning problem is linear regressions with auxiliary samples. In biomedical studies, some clinical or biological outcomes are hard to obtain due to ethical or cost issues, in which case transfer learning can be leveraged to boost the prediction and estimation performance by effectively utilizing information from related studies.

Transfer learning has been applied to problems in medical and biological studies, including predictions of protein localization (Mei et al., 2011), biological imaging diagnosis (Shin et al., 2016), drug sensitivity prediction (Turki et al., 2017), and integrative analysis of “multi-omics” data, see, for instance, Sun and Hu (2016), Hu et al. (2019), and Wang et al. (2019). It has also been applied to natural language processing (Daumé III, 2007) and recommendation systems (Pan and Yang, 2013) in machine learning. The application that motivated the present paper is the integration of the gene expression measurements in different issues for understanding the gene regulations using the Genotype-Tissue Expression (GTEx) data (https://gtexportal.org/). These datasets are always high-dimensional with relatively small sample sizes. When studying the gene regulation relationships of a specific tissue or cell-type, it is possible to incorporate information from other tissues to enhance the learning accuracy. This motivates us to consider transfer learning in high-dimensional linear regression.

1.1. Transfer Learning in High-dimensional Linear Regression

Regression analysis is one of the most widely used statistical methods to understand the association of an outcome with a set of covariates. In many modern applications, the dimension of the covariates is usually very high as compared to the sample size. Typical examples include genome-wide association and gene expression studies. In this paper, we consider transfer learning in high-dimensional linear models. Formally, the target model can be written as

yi(0)=(xi(0))β+ϵi(0),i=1,,n0, (1)

where ((xi(0)),yi(0)), i = 1, . . . , n0, are independent samples, βp is the coefficient vector of interest, and ϵi(0), i = 1, . . . , n0 are independently distributed random noises with E[ϵi(0)xi(0)]=0. In the high-dimensional regime, where p can be larger and much larger than n0, β is often assumed to be sparse such that the number of nonzero elements of β, denoted by s, is much smaller than p.

In the context of transfer learning, we observe additional samples from K auxiliary studies, That is, we observe ((xi(k)),yi(k)) generated from the auxiliary model

yi(k)=(xi(k))w(k)+ϵi(k),i=1,,nk,k=1,,K, (2)

where w(k)p is the regression vector for the k-th study, and ϵi(k) is the random noise such that E[ϵi(k)xi(k)]=0. The regression coefficients w(k) are unknown and different from our target β in general. The number of auxiliary studies, K, is allowed to grow but practically K may not be too large. We will study the estimation and prediction of target model (1) utilizing the primary data ((xi(0)),yi(0)), i = 1, . . . , n0, as well as the data from K auxiliary studies ((xi(k)),yi(k)), i = 1, . . . , nk, k = 1, . . . ,K.

If an auxiliary model is “similar” to the target model, we say that this auxiliary sample/study is informative. In this work, we characterize the informative level of the k-th auxiliary study using the sparsity of the difference between w(k) and β. Let δ(k) = βw(k) denote the contrast between w(k) and β. The set of informative auxiliary samples are those whose contrasts are sufficiently sparse:

Aq={1kK:δ(k)qh}, (3)

for some q ∈ [0, 1]. The set Aq contains the auxiliary studies whose contrast vectors have q-sparsity at most h and is called the informative set. It will be seen later that as long as h is relatively small compared to the sparsity of β, the studies in Aq can be useful in improving the prediction and estimation of β. In the case of q = 0, the set Aq corresponds to the auxiliary samples whose contrast vectors have at most h nonzero elements. We also consider approximate sparsity constraints (q ∈ (0, 1]), which allows all of the coefficients to be nonzero but their magnitude decays at a relatively rapid rate. For any q ∈ [0, 1], smaller h implies that the auxiliary samples in Aq are more informative; larger cardinality of Aq(|Aq|) implies that a larger number of informative auxiliary samples. Therefore, smaller h and larger |Aq| should be favorable. We allow Aq to be empty in which case none of the auxiliary samples is informative. For the auxiliary samples outside of Aq, we do not assume sparse δ(k) and hence w(k) can be very different from β for kAq.

In polygenic risk score (PRS) prediction and gene-expression partial-correlation analysis, this similarity characterization of two different high dimensional regression models is motivated by commonly adopted assumptions. In PRS prediction, for example, high-dimensional sparse regression models are commonly assumed (Mak et al., 2017). In addition, it has been observed that many complex traits have a shared genetic etiology, including various autoimmune diseases (Li et al., 2015; Zhernakova et al., 2009) and psychiatric disorders (Lee et al., 2013; Cross-Disorder Group of the Psychiatric Genomics Consortium, 2019). The similarity characterization we proposed captures the sparse nature of genome-wide association data and shared genetic etiology of multiple genetically related traits. In the gene expression data analysis, one is interested in understanding how a set of genes regulate another gene based on data measured in different tissues. Such an analysis provides useful insights into gene regulatory networks, which are often sparse. In addition, many tissues have shared regulatory relationships among the genes (Pierson et al., 2015; Fagny et al., 2017). In such applications, we also expect sparse and similar regression coefficients for the models assumed for different tissues.

There is a paucity of methods and fundamental theoretical results for high-dimensional linear regression in the transfer learning setting. In the case where the set of informative auxiliary samples Aq is known, there is a lack of rate optimal estimation and prediction methods. A closely related topic is multi-task learning (Ando and Zhang, 2005; Lounici et al., 2009; Agarwal et al., 2012), where the goal is to estimate multiple models simultaneously. The multi-task learning considered in Lounici et al. (2009) estimates multiple high-dimensional sparse linear models under the assumption that the supports of all the regression coefficients are the same. In multi-task learning, different regularization formats have been considered to model the similarity among different studies (Chen et al., 2010; Danaher et al., 2014; Dondelinger et al., 2020).

The goal of transfer learning is however different, as one is only interested in estimating the target model and this remains to be a largely unsolved problem. Cai and Wei (2021) studied the minimax and adaptive methods for nonparametric classification in the transfer learning setting under the assumption that all the auxiliary samples are similar to the target distribution (Cai and Wei, 2021, Definition 5). In a more challenging setting where the set Aq is unknown as is typical in real applications, it is unclear how to avoid the effects of adversarial auxiliary samples. Bastani (2020) studied estimation and prediction in high-dimensional linear models with one informative auxiliary study and q = 1, where the sample size of the auxiliary study is larger than the number of covariates. The current work considers more general scenarios under weaker assumptions. Specifically, the sample size of auxiliary samples can be smaller than the number of covariates and some auxiliary studies can be non-informative, which is more practical in applications. Additional challenges include the heterogeneity among the design matrices, which does not arise in the conventional high-dimensional regression problems and hence requires novel proposals.

The problem we study here is certainly related to the high-dimensional prediction and estimation in the conventional settings where only samples from the target model are available. Several penalized or constrained minimization methods have been proposed for prediction and estimation for high-dimensional linear regression; see, for example, Tibshirani (1996); Fan and Li (2001); Zou (2006); Candes and Tao (2007); Zhang (2010). The minimax optimal rates for estimation and prediction are studied in Raskutti et al. (2011) and Verzelen (2012).

1.2. Our Contributions

In the setting where the informative set Aq is known, we propose a transfer learning algorithm, called Oracle Trans-Lasso, for estimation of the target regression vector and prediction and prove its minimax optimality under mild conditions. The results demonstrate a faster rate of convergence when Aq is non-empty and h is sufficiently smaller than s, in which case the knowledge from the informative auxiliary samples can be optimally transferred to substantially improve estimation and prediction of the regression problem under the target model.

In the more challenging setting where Aq is unknown a priori, we introduce a data-driven algorithm, called Trans-Lasso, to adapt to the unknown Aq. The adaption is achieved by aggregating a number of candidate estimators. The desirable properties of the aggregation methods guarantee that the Trans-Lasso does not perform much worse than the best one among the candidate estimators. We construct the candidate estimators and demonstrate the robustness and the efficiency of Trans-Lasso under mild conditions. In terms of robustness, the Trans-Lasso is guaranteed to be not much worse than the Lasso estimator using only the primary samples no matter how adversarial the auxiliary samples are. In terms of efficiency, the knowledge from a subset of the informative auxiliary samples can be transferred to the target problem under proper conditions. Furthermore, If the contrast vectors in the informative samples are sufficiently sparse, the Trans-Lasso estimator performs as if the informative set Aq is known.

When the distributions of the design matrices are distinct in different samples, the effect of heterogeneous designs in transfer learning is studied. The performance of the proposed algorithm is investigated theoretically and numerically in various settings.

1.3. Organization and Notation

The rest of this paper is organized as follows. Section 2 focuses on the setting where the informative set Aq is known and with the sparsity in (3) measured in 1-norm. A transfer learning algorithm is proposed for estimation and prediction of the target parameter and its minimax optimality is established. In Section 3, we study the estimation and prediction of the target model when Aq is unknown for q = 1. In Section 4, we justify the theoretical performance of our proposals under heterogeneous designs. In Section 5, the numerical performance of the proposed methods is studied in various settings. In Section 6, the proposed algorithms are applied to the GTEx data to investigate the association of one gene with other genes in a target tissue by leveraging data measured on other related tissues or cell types. The proofs and results for q-sparse contrasts with q ∈ [0, 1) are provided in the supplementary materials (Li et al., 2020).

We finish this section with notation. Let X(0)n0×p and y(0)n0 denote the design matrix and the response vector for the primary data, respectively. Let X(k)nk×p and y(k)nk denote the design matrix and the response vector for the k-th auxiliary data, respectively. For a class of matrices Rlnl×p0, lL, we use {Rl}lL to denote Rl, lL. Let nAq=kAqnk. For a generic semi-positive definite matrix Σm×m, let Λmax(Σ) and Λmin(Σ) denote the largest and smallest eigenvalues of Σ, respectively. Let Tr(Σ) denote the trace of Σ. Let ej be a vector such that its j-th element is 1 and all other elements are zero. Let ab denote max{a, b} and ab denote min{a, b}. We use c, c0, c1, . . . to denote generic constants which can be different in different statements. Let an = O(bn) and an ≲. bn denote |an/bn| ≤ c for some constant c when n is large enough. Let anbn denote |an/bn| → c for some constant c as n → ∞. Let an = OP (bn) and anbn denote (|an/bn|c)1 for some constant c < ∞. Let an = oP (bn) denote (|an/bn|>c)0 for any constant c > 0.

2. Estimation with Known Informative Auxiliary Samples

We consider in this section transfer learning for high-dimensional linear regression when the informative set Aq is known. The focus is on the 1-sparse characterization of the contrast vectors. The notation A1 will be abbreviated as A in the sequel without special emphasis. Section C in the supplementary materials generalizes the sparse contrasts from 1-constraint to q-constraint for q ∈ [0, 1) and presents a rate-optimal estimator in this setting.

2.1. Oracle Trans-Lasso Algorithm

We propose a transfer learning algorithm, called Oracle Trans-Lasso, for estimation and prediction when A is known. As an overview, we first compute an initial estimator using all the informative auxiliary samples. However, its probabilistic limit is biased from β as w(k)β in general. We then correct its bias using the primary data in the second step. Algorithm 1 formally presents our proposed Oracle Trans-Lasso algorithm.

Algorithm 1:

Oracle Trans-Lasso algorithm

Input : Primary data (X(0), y(0)) and informative auxiliary samples {X(k),y(k)}kA
Output: β^
Step 1. Compute
w^A=argminwp{12nAkAy(k)X(k)w22+λww1} (4)
 for λw=c1logp/nA with some constant c1.
Step 2. Let
β^=w^A+δ^A, (5)
 where
δ^A=argminδp{12n0y(0)X(0)(w^A+δ)22+λδδ1} (6)
 for λδ=c2logp/n0 with some constant c2.

In Step 1, w^A is realized based on the Lasso (Tibshirani, 1996) using all the informative auxiliary samples. Its probabilistic limit is wA, which can be defined via the following moment condition

E[kA(X(k))(y(k)X(k)wA)]=0.

Denoting E[xi(k)(xi(k))]=Σ(k), wA has the following explicit form:

wA=β+δA (7)

for δA=kAαkδ(k) and αk=nk/nA given that Σ(k) = Σ(0) for all kA. That is, the probabilistic limit of w^A, wA, has bias δA, which is a weighted average of δ(k). Step 1 is related to the approach for high-dimensional misspecified models (Bühlmann and van de Geer, 2015) and moment estimators. The estimator w^A converges relatively fast as the sample size used in Step 1 is relatively large. Step 2 corrects the bias, δA, using the primary samples. In fact, δA is a sparse high-dimensional vector whose 1-norm is no larger than h. Hence, the error of step 2 is under control for a relatively small h. The choice of the tuning parameters λw and λδ will be further specified in Theorem 1.

We compare the proposed Oracle Trans-Lasso method to the multi-task regression methods, say Section 3.4.3 of Agarwal et al. (2012) and Danaher et al. (2014). The Oracle Trans-Lasso does not penalize the differences among the regression coefficients in the auxiliary studies. This is again because the focus of transfer learning is only the target study. Theoretically, extra penalization terms and the joint analysis of multiple estimators may not help improve the estimation accuracy of the parameter of interest.

2.2. Theoretical Properties of Oracle Trans-Lasso

Formally, the parameter space we consider can be written as

Θq(s,h)={B=(β,δ(1),,δ(K)):β0s,maxkAqδ(k)qh} (8)

for Aq{1,,K} and q ∈ [0, 1]. We study the rate of convergence for the Oracle Trans-Lasso algorithm under the following two conditions.

Condition 1.

For each kA{0}, each row of X(k) is i.i.d. Gaussian distributed with mean zero and covariance matrix Σ. The smallest and largest eigenvalues of Σ are bounded away from zero and infinity, respectively.

Condition 2.

For each kA{0}, E[(yi(k))2] is finite and the random noises ϵi(k) are i.i.d. sub-Gaussian with mean zero and variance σk2. For some constant C0, it holds that maxkA{0}E[exp{tϵi(k)}]exp{t2C0} for all t.

Condition 1 assumes Gaussian designs, which provides convenience for bounding the restricted eigenvalues of sample covariance matrices. Moreover, the designs are identically distributed for kA{0}. This assumption simplifies some technical conditions and will be relaxed in Section 4. We mention that the conditions on the eigenvalues of Σ can be replaced with some eigenvalue conditions restricted to a convex cone. Condition 2 assumes sub-Gaussian random noises for primary and informative auxiliary samples and the second moment of the response vector is finite. Conditions 1 and 2 make no assumptions on the non-informative auxiliary samples as they are not used in the Oracle Trans-Lasso algorithm. In the next theorem, we prove the convergence rate of the Oracle Trans-Lasso. Let ηh=hlogp/n0h2.

Theorem 1 (Convergence Rate of Oracle Trans-Lasso).

Assume that Condition 1 and Condition 2 hold true. Suppose that A is known with hslogp/n0 and n0nA. We take λw=maxkAc1E[(yi(k))2]logp/nA and λδ=c2logp/n0 for some sufficiently large constants c1 and c2. If slogp/nA+h(logp/n0)1/2=o(1), then there exists some positive constant c1 such that

infBΘ1(s,h)(1n0X(0)(β^β)22β^β22slogpnA+n0+slogpn0ηh)1exp(c1logp). (9)

where B = {β,w(1), . . . ,w(k)} denotes all the unknown parameters. Theorem 1 provides the convergence rate of β^ for any true parameters in Θ1(s, h) when an informative set A is known. We illustrate Theorem 1 by contrasting to the estimation results of the Lasso. First, the results of Theorem 1 hold under a weaker condition on s, i.e., slogp=O(nA) when nAn0, while s log p = o(n0) is always assumed in the single-task regression. Hence, the Oracle Trans-Lasso can deal with more challenging scenarios with less sparse target parameter. Second, the right-hand side of (9) is sharper than the convergence rate of Lasso, s log p/n0, if hslogp/n0 and nAn0. That is, if the informative auxiliary samples have contrast vectors sufficiently sparser than β and the total sample size is significantly larger than the primary sample size, then the knowledge from the auxiliary samples can significantly improve the learning performance of the target model. The condition for improvement, hslogp/n0, allows a wide range of h. For example, the typical regime for single-task regression is s log p/n0 = O(1) and it implies that slogp/n0 can be as large as n0/logp. Hence, the condition for improvement of Theorem 1 allows h to be as large as n0/logp. Larger the s, weaker the condition for improvement.

The sample size requirement in Theorem 1 guarantees the lower restricted eigenvalues of the sample covariance matrices in use are bounded away from zero with high probability. The proof of Theorem 1 involves an error analysis of w^A and that of δ^A. While wA may be neither 0-nor 1-sparse, it can be decomposed into an 0-sparse component plus an 1-sparse component as illustrated in (7). Exploiting this sparse structure is a key step in proving Theorem 1. Regarding the choice of tuning parameters, λw depends on the second moment of yi(k), which can be consistently estimated by y(k)22/nk. The other tuning parameter λδ depends on the noise levels, which can be estimated by the scaled Lasso (Sun and Zhang, 2012). In practice, cross validation can be performed for selecting tuning parameters.

We now establish the minimax lower bound for estimating β in the transfer learning setup, which shows the minimax optimality of the Oracle Trans-Lasso algorithm in Θ1(s, h).

Theorem 2 (Minimax lower bound for q = 1).

Assume Condition 1 and Condition 2. If max{slogp/(nA+n0),h(logp/n0)1/2}=o(1), then

infβ^supBΘ1(s,h)(β^β22c1slogpnA+n0+c2slogpn0ηh)12

for some positive constants c1 and c2.

Theorem 2 implies that β^ obtained by the Oracle Trans-Lasso algorithm is minimax rate optimal in Θ1(s, h) under the conditions of Theorem 1. To understand the lower bound, the term slogp/(nA+n0) is the optimal convergence rate when w(k) = β for all kA. This is an extremely ideal case where we have nA+n0 i.i.d. samples from the target model. The second term in the lower bound is the optimal convergence rate when w(k) = 0 for all kA, i.e., the auxiliary samples are not helpful at all. Let Bq(r)={up:uqr} denote the q-ball with radius r centered at zero. In this case, the definition of Θ1(s, h) implies that βB0(s)B1(h) and the second term in the lower bound is indeed the minimax optimal rate for estimation when βB0(s)B1(h) with n0 i.i.d. samples (Tsybakov, 2014).

3. Unknown Set of Informative Auxiliary Samples

The Oracle Trans-Lasso algorithm is based on the knowledge of the informative set A. In some applications, the informative set A is not given, which makes the transfer learning problem more challenging. In this section, we propose a data-driven method for estimation and prediction when A is unknown. The proposed algorithm is described in detail in Sections 3.1 and 3.2. Its theoretical properties are studied in Section 3.3.

3.1. The Trans-Lasso Algorithm

Our proposed algorithm, called Trans-Lasso, consists of two main steps. First, we construct a collection of candidate estimators, each of which is based on an estimate of A. Second, we perform an aggregation step (Rigollet and Tsybakov, 2011; Dai et al., 2012, 2018) on these candidate estimators. Under proper conditions, the aggregated estimator is guaranteed to be not much worse than the best candidate estimator under consideration in terms of prediction. For technical reasons, we need the candidate estimators and the samples for aggregation to be independent. Hence, we start with sample splitting. We need some more notation. For a generic estimate of β, b, denote its sum of squared prediction error as

Q^(I,b)=iIyi(0)(xi2(0))b22,

where I is a subset of {1, . . . , n0}. Let ΛL+1={νL+1:νl0,l=0Lνl=1} denote an L-dimensional simplex. The Trans-Lasso algorithm is presented in Algorithm 2.

As an illustration, steps 2 and 3 of the Trans-Lasso algorithm construct some initial estimates of β, β^(G^l). They are computed using the Oracle Trans-Lasso algorithm by treating each G^l as the set of informative auxiliary samples. We construct G^l to be some estimates of A using the procedure provided in Section 3.2. Step 4 is based on the Q-aggregation proposed in Dai et al. (2012) with a uniform prior, a Kullback–Leibler penalty, and a simplified tuning parameter. The Q-aggregation can be viewed as a weighted version of least square aggregation and exponential aggregation (Rigollet and Tsybakov, 2011) and it has been shown to be rate optimal both in expectation and with high probability for model selection aggregation problems.

Algorithm 2:

Trans-Lasso Algorithm

Input : Primary data (X(0), y(0)) and samples from K auxiliary studies {X(k),y(k)}k=1K.
Output: β^θ^
Step 1. Let I be a random subset of {1, . . ., n0} such that |I|c0n0 with some constant 0 < c0 < 1. Let Ic={1,,n0}\I.
Step 2. Construct L + 1 candidate sets of f A, {G^0,G^1,,G^L} such that G^0= and G^1,,G^L are based on (14) using (XI,.(0),yI(0)) and {X(k),y(k)}k=1K.
Step 3. For each 0 ≤ lL, run the Oracle Trans-Lasso algorithm with primary sample (XI,.(0),yI(0)) and auxiliary samples (XI,.(0),yI(0)) and auxiliary samples {X(k),y(k)}kG^l. Denote the output as β^(G^l) for 0 ≤ lL.
Step 4. Compute
θ^=argminθΛL+1{Q^(Ic,l=0Lβ^(G^l)θl)+l=0LθlQ^(Ic,β^(G^l))+2λθn0l=0Lθllog(θl)} (10)
 for some λθ > 0. Output
β^θ^=l=0Lθ^lβ^(G^l). (11)

Model selection aggregation is an effective method for the transfer learning task under consideration. On one hand, it guarantees the robustness of Trans-Lasso in the following sense. Notice that β^(G^0) corresponds to the single-task Lasso estimator and it is always included in our dictionary. The purpose is that, invoking the property of model selection aggregation, the performance of β^θ^ is guaranteed to be not much worse than the performance of the original Lasso estimator under mild conditions. This shows that the performance of Trans-Lasso will not be ruined by adversarial auxiliary samples. Formal statements are provided in Section 3.3. On the other hand, the gain of Trans-Lasso relates to the qualities of G^1,,G^L. If

(G^lA,forsome1lL)1, (12)

i.e., G^l is a non-empty subset of the informative set A, then the model selection aggregation property implies that the performance of β^θ^ is not much worse than the performance of the Oracle Trans-Lasso with kG^lnk informative auxiliary samples. Ideally, one would like to achieve G^l=A for some 1 ≤ lL with high probability. However, it can rely on strong assumptions that may not be guaranteed in practical situations.

To motivate our constructions of G^l, let us first point out a naive construction of candidate sets, which consists of 2K candidates. These candidates are all different combinations of {1, . . . ,K}, denoted by G^1,,G^2K. It is obvious that A is an element of these candidate sets. However, the number of candidates is too large and it can be computationally burdensome. Furthermore, the cost of aggregation can be significantly high, which is of order K/n0 as will be seen in Lemma 1. In contrast, we would like to pursue a much smaller number of candidate sets such that the cost of aggregation is almost negligible and (12) can be achieved under mild conditions. We introduce our proposed construction of candidate sets in the next subsection.

3.2. Constructing the Candidate Sets for Aggregation

As illustrated in Section 3.1, the goal of Step 2 is to have a class of candidate sets, {G^0,,G^L}, that satisfy (12) under certain conditions. Our idea is to exploit the sparsity patterns of the contrast vectors. Recall that the definition of A implies that {δ(k)}kA are sparser than {δ(k)}kAc, where Ac={1,,K}\A. This property motivates us to find a sparsity index R(k) and its estimator R^(k) for each 1 ≤ kK such that

maxkAoR(k)<minkAcR(k)and(maxkAoR^(k)<minkAcR^(k))1, (13)

where Ao is some subset of A. In words, the sparsity indices in Ao are no larger than the sparsity indices in Ac and so are their estimators with high probability. To utilize (13), we can define the candidate sets as

G^l={1kK:R^(k)isamongthefirstlsmallestofall} (14)

for 1 ≤ lK. That is, G^l is the set of auxiliary samples whose estimated sparsity indices are among the first l smallest. A direct consequence of (13) and (14) is that (G^|Ao|=Ao)1 and hence the desirable property (12) is satisfied. To achieve the largest gain with transfer learning, we would like to find proper sparsity indices such that (13) holds for kA1°nk as large as possible. Notice that G^K+1={1,,K} is always included as candidates according to (14). Hence, in the special cases where all the auxiliary samples are informative or none of the auxiliary samples are informative, it holds that G^|A|=A and the Trans-Lasso is not much worse than the Oracle Trans-Lasso. The more challenging cases are 0<|A|<K.

As {δ(k)}kAc are not necessarily sparse, the estimation of δ(k) or functions of δ(k), 1 ≤ kK, is not trivial. As an example, an intuitive sparsity index can be ∥δ(k)1 and its estimate is β^(G^0)w^(k)1, where w^(k) is the Lasso estimate of w(k) based on the k-th study. However, such a Lasso-based estimate is not guaranteed to converge to the oracle ∥δ(k)1 when δ(k) is non-sparse. Therefore, we consider using R(k)=δ(k)22, which is a function of the population-level marginal statistics, as the oracle sparsity index for k-th auxiliary sample. The advantage of R(k) is that it has a natural unbiased estimate even when δ(k) is non-sparse. Let us relate R(k) to the sparsity of δ(k) using a Bayesian characterization of sparse vectors assuming Σ(k) = Σ for all 0 ≤ kK. If δj(k) are i.i.d. Laplacian distributed with mean zero and variance νk2 for each k, then it follows from the properties of Laplacian distribution (Liu and Kozubowski, 2015) that E[δ(k)1]E1/2[δ(k)22]. Hence, the rank of E[Σδ(k)22] is the same as the rank of E[δ(k)1]. As maxkAδ(k)1<minkAcδ(k)1, it is reasonable to expect maxkAδ(k)22<minkAcδ(k)22. The above derivation holds for many other zero mean prior distributions besides Laplacian. This illustrates our motivation for considering R(k) as the oracle sparsity index.

We next introduce the estimated version, R^(k), based on the primary data {(xi(0)),yi(0)}iI (after sample splitting) and auxiliary samples {X(k),y(k)}k=1K. We first perform a SURE screening (Fan and Lv, 2008) on the marginal statistics to reduce the effects of random noises. We summarize our proposal for Step 2 of the Trans-Lasso as follows (Algorithm 3). Let n*=min0kKnk.

Algorithm 3:

Step 2 of the Trans-Lasso Algorithm

Step 2.1. For 1 ≤ kK, compute the marginal statistics
Δ^(k)=1nki=1nkxi(k)yi(k)1|I|iIxi(0)yi(0). (15)
 For each k ∈ {1, . . ., K}, let T^k be obtained by SURE screening such that
  T^k={1jp:|Δ^j(k)| is among the first t* largest of all}
 for a fixed t*=n*α, 0 ≤ α < 1.
Step 2.2. Define the estimated sparse index for the k-th auxiliary sample as
R^(k)=Δ^T^k(k)22. (16)
Step 2.3. Compute G^l as in (14) for l = 1, . . ., L.

One can see that Δ^(k) are empirical marginal statistics such that E[Δ^(k)]=Σδ(k) for kA. The set T^k is the set of first t* largest marginal statistics for the k-th sample. The purpose of screening the marginal statistics is to reduce the magnitude of noise. Notice that the un-screened version Δ^(k)22 is a sum of p random variables and it contains noise of order p/(nkn0), which diverges fast as p is much larger than the sample sizes. By screening with t* of order n*α, α < 1, the errors induced by the random noises is under control. In practice, the auxiliary samples with very small sample sizes can be removed from the analysis as their contributions to the target problem is mild. Desirable choices of T^k should keep the variation of Σδ(k) as much as possible. Under proper conditions, SURE screening can consistently select a set of strong marginal statistics and hence is appropriate for the current purpose. In Step 2.2, we compute R^(k) based on the marginal statistics which are selected by SURE screening. In practice, different choices of t* may lead to different realizations of G^l. One can compute multiple sets of {R^(k)}k=1K with different t* which give multiple sets of {G^l}l=1K. It will be seen from Lemma 1 that a finite number of choices on t* does not affect the rate of convergence.

3.3. Theoretical Properties of Trans-Lasso

In this subsection, we derive the theoretical guarantees for the Trans-Lasso algorithm. We first establish the model selection aggregation type of results for the Trans-Lasso estimator β^θ^.

Lemma 1 (Q-Aggregation for Trans-Lasso).

Assume that Condition 1 and Condition 2 hold true. Let θ^ be computed via (10) with λθ4σ02. With probability at least 1 − t, it holds that

1|Ic|XIc,.(0),β^θ^β22min0lL1|Ic|XIc,.(0),(β^(G^l)β)22+λθlog(L/t)n0. (17)

If Lc1n0 for some small enough constant c1, then

β^θ^β22min0lLβ^(G^l)β22+logLn0. (18)

Lemma 1 implies that the performance of β^θ^ only depends on the best candidate regardless of the performance of other candidates under mild conditions. As commented before, this result guarantees the robustness and efficiency of Trans-Lasso, which can be formally stated as follows. As the original Lasso is always in our dictionary, (17) and (18) imply that β^θ^ is not much worse than the Lasso in prediction and estimation. Formally, “not much worse” refers to the last term in (17), which can be viewed as the cost of “searching” for the best candidate model within the dictionary which is of order log L/n0. This term is almost negligible, say, when L = O(K), which corresponds to our constructed candidate estimators. This demonstrates the robustness of β^θ^ to adversarial auxiliary samples. Furthermore, if (12) holds, then the prediction and estimation errors of Trans-Lasso are comparable to the Oracle Trans-Lasso using the auxiliary samples in Ao.

The prediction error bound in (17) follows from Corollary 3.1 in Dai et al. (2012). However, the aggregation methods do not have theoretical guarantees in estimation errors in general. Indeed, an estimator with 2-error guarantee is crucial for more challenging tasks, such as out-of-sample prediction and inference. For our transfer learning task, we show in (18) that the estimation error is of the same order if the cardinality of the dictionary is Lcn0 for some small enough c. For our constructed dictionary, it suffices to require Kcn0. In many practical applications, K is relatively small compared to the sample sizes and hence this assumption is not very restrictive.

In the following, we provide sufficient conditions such that the desirable property (13) holds with R^(k) defined in (16) and hence (12) is satisfied. For each kAc, define a set

Hk={1jp:|Σj,.(k)w(k)Σj,.(0)β|>n*κ,κ<α/2}. (19)

Recall that α < 1 is defined such that t* = nα. In fact, Hk is the set of “strong” marginal statistics that can be consistently selected into T^k for each kAc. We see that Σj,.(k)w(k)Σj,.(0)β=Σj,.δ(k) if Σ(k) = Σ(0) for kAc. The definition of Hk in (19) allows for heterogeneous designs among non-informative auxiliary samples.

Condition 3.

  1. For each kAc, each row of X(k) is i.i.d. Gaussian with mean zero and covariance matrix Σ(k) and maxkAcΛmax((k)) is finite. For each kAc, the random noises ϵi(k) are i.i.d. Gaussian with mean zero and variance σk2 and E[(yi(k))2] is finite.

  2. It holds that logplogKc1n* for a small enough constant c1. Moreover,
    minkAcjHk|Σj,.(k)w(k)Σj,.(0)β|2c2logpn*1α (20)
    for some constant c2 > 0.

The Gaussian assumptions in Condition 3(a) guarantee the desirable properties of SURE screening for the non-informative auxiliary studies. In fact, the largest eigenvalue of Σ(k), kAc can grow as O(n*τ) for some τ ≥ 0 and τ + α < 1 following the proof in Fan and Lv (2008). The Gaussian assumption can be relaxed to be sub-Gaussian random variables according to some recent studies (Ahmed and Bajwa, 2019). For the conciseness of the proof, we consider Gaussian distributed random variables with bounded eigenvalues. Condition 3(b) puts a constraint on the relative dimensions. It is trivial in the regime that pKn*ξ for any finite ξ > 0. The expression (20) requires that for each Ac, there exists a subset of strong marginal statistics with not-so-small cardinality. This condition is mild by choosing α such that logpn*1α and α = 1/2 is an obvious choice revoking the first part of Condition 3(b). For instance, if minkAcE[Δ^(k)]c0>0, then (20) holds with any α ≤ 1/2. In words, a sufficient condition for (20) is that at least one marginal statistic in the k-th study is of constant order for kAc. We see that larger n* makes Condition 3 weaker. As mentioned before, it is helpful to remove the auxiliary samples with very small sample sizes from the analysis.

In the next theorem, we demonstrate the theoretical properties of R^(k) and provide a complete analysis of the Trans-Lasso algorithm. Let Ao be a subset of A such that

Ao={kA:Σ(0)δ(k)22c1minkAcjHk|Σj,.(k)w(k)Σj,.(0)β|2}

for some small constant c1 < 1 and Hk defined in (19). In general, one can see that the informative auxiliary samples with sparser δ(k) are more likely to be included into Ao. Specially, the fact that maxkA(0)δ(k)22Σ(0)22h2 implies Ao=A when h is sufficiently small. We will show (13) for such Ao with R^(k) defined in (16). Let nAo=kA°nk.

Theorem 3 (Convergence Rate of the Trans-Lasso).

Assume Conditions 1, 2, and 3. Then

(maxkA°R^(k)<minkAcR^(k))1. (21)

Let β^θ^ be computed using the Trans-Lasso algorithm with λθ4σ02. If slogp/(nAo+n0)+{h(logp/n0)1/2}(slogp/n0)=o(1) and Kcn0 for a sufficiently small constant c > 0, then

infBΘ1(s,h)(1|Ic|XIc,.(0)(β^θ^β)22β^θ^β22slogpnAo+n0+slogpn0ηh+logKn0)1 (22)

as (n0,nAo,p).

Remark 1.

Under the conditions of Theorem 3, if

Σ(0)22h2cminkAcjHk|Σj,.(k)w(k)Σj,.(0)β|2forsomec<1,

then (maxkAR^(k)<minkAcR^(k))1 and as (n0,nA,p),

infBΘ1(s,h)(1|Ic|XIc,.(0)(β^θ^β)22β^θ^β22slogpnA+n0+slogpn0ηh+logKn0)1.

Theorem 3 establishes the convergence rate of the Trans-Lasso when A is unknown. The result in (21) implies the estimated sparse indices in Ao and in Ac are separated with high probability. As illustrated before, a consequence of (21) is (12) for the candidate sets G^l defined in (14). Together with Theorem 1 and Lemma 1, we arrive at (22).

It is worth mentioning that Condition 3 is only employed to show the gain of Trans-Lasso. The robustness property of Trans-Lasso holds without any conditions on the non-informative samples (Lemma 1). In practice, missing a few informative auxiliary samples may not be a grave concern. One can see that when nAo is large enough such that the first term on the right-hand side of (22) no longer dominates, increasing the number of auxiliary samples will not improve the convergence rate. In contrast, it is more important to guarantee that the estimator is not affected by the adversarial auxiliary samples. The empirical performance of Trans-Lasso is carefully studied in Section 5.

4. Extensions to Heterogeneous Designs

In this section, we extend the algorithms and theoretical results developed in Sections 2 and 3 to the case where the covariates have different covariance structures in different studies.

The Oracle Trans-Lasso algorithm proposed in Section 2 can be directly applied to the setting where the design matrices are moderately heterogeneous. Formally, we first introduce a relaxed version of Condition 1 as follows. Define

CΣ=1+maxjpmaxkAej(Σ(k)Σ(0))(kAαkΣ(k))11,

which characterizes the differences between Σ(k) and Σ(0) for kA. Notice that CΣ is a constant if max1jpej(Σ(k)Σ(0))0C< for all kA, where examples include block diagonal Σ(k) with constant block sizes or banded Σ(k) with constant bandwidths for kA.

Condition 4.

For each kA{0}, each row of X(k) is i.i.d. Gaussian with mean zero and covariance matrix Σ(k). The smallest eigenvalue of Σ(k) are bounded away from zero for all kA{0}. The largest eigenvalue of Σ(0) is bounded away from infinity.

The following theorem characterizes the rate of convergence of the Oracle Trans-Lasso estimator in terms of CΣ. Let ηh,Σ=(CΣhlogp/n0)(CΣ2h2).

Theorem 4 (Oracle Trans-Lasso with heterogeneous designs).

Assume that Condition 2 and Condition 4 hold true. Suppose A is known with CΣhslogp/n0 and n0nA. We take λw and λδ as in Theorem 1. If slogp/nA+CΣh(logp/n0)1/2=o(1), then

infBΘ1(s,h)(1n0X(0)(β^β)22β^β22slogpnA+n0+slogpn0ηh,Σ)1exp(c1logp). (23)

The right-hand side of (9) is sharper than s log p/n0 if nAn0 and CΣhlogp/n0s. We see that small CΣ is favorable. This implies that the Oracle Trans-Lasso is guaranteed to perform well with sparse contrasts and similar covariance matrices to the primary one.

We now provide theoretical guarantees for the Trans-Lasso with heterogeneous designs when A is unknown. In this case, the sparsity index R(k) takes the format Σ(k)w(k)Σ(0)β22. It measures the sparsity of δ(k) but also the covariance heterogeneity. We consider A˜o, a subset of A such that

A˜o={kA:Σ(k)w(k)Σ(0)β22<c1minkAcjHk|Σj,.(k)w(k)Σj,.(0)β|2}

for some c1 < 1 and Hk defined in (19). This is a generalization of A° to the case of heterogeneous designs.

Corollary 1 (Trans-Lasso with heterogeneous designs).

Assume Conditions 2, 3, and 4. Let β^θ^ be computed via the Trans-Lasso algorithm with λθ4σ02. If slogp/(nA˜o+n0)+{CΣh(logp/n0)1/2}(slogp/n0)=o(1) and Kcn0 for a small enough constant c, then

infBΘ1(s,h)(1|Ic|XIc,.,(0)(β^θ^β)22β^θ^β22slogpnA˜o+n0+slogpn0ηh,Σ+logKn0)1

as (n0,nA˜o,p).

Corollary 1 provides an upper bound for the Trans-Lasso with heterogeneous designs. The numerical experiments for this setting are studied in Section 5.

5. Simulation Studies

In this section, we evaluate the empirical performance of the proposed methods and some other comparable methods in various numerical experiments. Specifically, we evaluate the performance of five methods, including Lasso, Oracle Trans-Lasso proposed in Section 2.1, Trans-Lasso proposed in Section 3.1, and two other ad hoc transfer learning methods related to ours. The first one implements Trans-Lasso except that the bias-correction step (Step 2) of the Oracle Trans-Lasso is omitted. We call this method the “aggregated Lasso” (Agg-Lasso), as it implements our proposed adaptive aggregation step and applies Lasso to each candidate set. The purpose is to understand the necessity of the bias-correction step in Oracle Trans-Lasso. The second one follows the steps of Trans-Lasso but uses a different aggregation step. Specifically, we consider R^(k)=β^Lw^(k)1, k = 1, . . . ,K, where β^L and w^(k) are the Lasso estimators based on each of the corresponding studies. Moreover, the Q-aggregation step is replaced with the cross-validation, where we select the set G^l that minimizes the out-of-sample prediction errors. We call this algorithm “Ad hoc ℓ1-transfer”. The purpose of including this method is to understand the performance of our proposed R^(k) based on SURE screening and Q-aggregation. In the Supplementary Materials, we report the performance of the estimated sparse indices R^(k) based on Trans-Lasso and Ad hoc 1-transfer. The R code for all the methods are available at https://github.com/saili0103/TransLasso.

5.1. Identity Covariance Matrix for the Designs

We consider p = 500, n0 = 150, and n1, . . . , nK = 100 for K = 20. The covariates xi(k) are i.i.d. Gaussian with mean zero and identity covariance matrix for all 0 ≤ kK and ϵi(k) are i.i.d. Gaussian with mean zero and variance one for all 0 ≤ kK. For the target parameter β, we set s = 16, βj = 0.3 for j ∈ {1, . . . , s}, and βj = 0 otherwise. For the regression coefficients in auxiliary samples, we consider two configurations.

  1. For a given A, if kA, let
    wj(k)=βj0.31(jHk),
    where Hk is a random subset of [p] with |Hk| = h ∈ {2, 6, 12}. If kA, we set Hk to be a random subset of [p] with |Hk| = 2s and wj(k)=βj0.51(jHk). We set w1(k)=0.3 for k = 1, . . . ,K.
  2. For a given A, if kA, let Hk = {1, . . . , 100} and
    wj(k)=βj+ξj1(kHk),whereξji.i.d.N(0,h/100),
    where h ∈ {2, 6, 12} and N(a, b) is the normal with mean a and standard deviation b. If kA, we set Hk = {1, . . . , 100} and
    wj(k)=βj+ξj1(jHk),whereξji.i.d.N(0,2s/100).
    We set w1(k)=0.3 for k = 1, . . . ,K. The setting (i) can be treated as either 0- or 1-sparse contrasts. In practice, the true parameters are unknown and we use A to denote the set of auxiliary samples without distinguishing 0- or 1-sparsity. We consider |A|{0,4,8,,20}.

In Figure 1, we report sum of squared estimation errors (SSE) for each estimator |A|. Each point is summarized from 200 independent simulations. As expected, the performance of the Lasso does not change as |A| increases. On the other hand, all four other transfer learning-based algorithms have estimation errors decreasing as A increases. As h increases, the problem gets harder and the estimation errors of all four methods increase. In settings (i) and (ii), the Oracle Trans-Lasso has the smallest estimation errors in most settings. The proposed Trans-Lasso, which is agnostic to A, is always the second-best. The gap between the Oracle Trans-Lasso and Trans-Lasso is a result of the uncertainty of aggregation and sample splitting for constructing the initial estimators. We also observe that when A=, the Trans-Lasso can have smaller errors than the oracle Trans-Lasso where the latter one does not use auxiliary information. This implies that some auxiliary information can still be borrowed. Due to the randomness of the parameter generation, our definition of A may not always be the best subset of auxiliary samples that give the smallest estimation errors.

Figure 1.

Figure 1.

Estimation errors of the Ad hoc 1-transfer, Agg-Lasso, Lasso, Oracle Trans-Lasso, and Trans-Lasso with identity covariance matrices of the predictors. The two rows correspond to configurations (i) and (ii), respectively. The y-axis corresponds to bβ22 for some estimator b.

Among the two variants, Ad hoc 1-transfer is also adaptive but has slightly larger estimation errors than Trans-Lasso when h is large. This demonstrates the advantage of Q-aggregation with our proposed sparsity index over the cross-validation type of aggregation with 1-distance based sparsity index. The Agg-Lasso method has larger estimation errors than Trans-Lasso and Ad hoc 1-transfer, even when h is small. This demonstrates the necessity of the bias-correction step in the Oracle Trans-Lasso.

5.2. Homogeneous Designs among A{0}

We now consider xi(k) as i.i.d. Gaussian with mean zero and a equi-correlated covariance matrix, where Σj,j = 1 and Σj,k = 0.8 if jk for kA{0}. For kA{0}, xi(k) are i.i.d. Gaussian with mean zero and a Toeplitz covariance matrix whose first row is

Σ1,.(k)=(1,1/(k+1),,1/(k+1)2k1,0p2k). (24)

Other true parameters and the dimensions of the samples are set to be the same as in Section 5.1. From the results presented in Figure 2, we see that the Trans-Lasso and Oracle Trans-Lasso have reliable performance in the current setting. The average estimation errors are larger in Figure 2 than those in Section 5.1 as the covariates are highly correlated in the current setting. When h is relatively large, we see that Agg-Lasso and Ad hoc 1-transfer have significantly larger estimation errors than Trans-Lasso. This again demonstrates the advantage of Trans-Lasso over some ad hoc methods.

Figure 2.

Figure 2.

Estimation errors of the Ad hoc 1-transfer, Agg-Lasso, Lasso, Oracle Trans-Lasso, and Trans-Lasso with homogeneous covariance matrices. The two rows correspond to configurations (i) and (ii), respectively. The y-axis corresponds to bβ22 for some estimator b.

5.3. Heterogeneous Designs

We next consider a setting where Σ(k) are distinct for k = 0, . . . ,K. Specifically, for k = 1, . . . ,K, let xi(k) as i.i.d. Gaussian with mean zero and a Toeplitz covariance matrix whose first row is (24). Moreover, Σ(0)=Ip. Other parameters and the dimensions of the samples are set to be the same as in Section 5.1. Figure 3 shows that the general patterns observed under homogeneous designs still hold. Trans-Lasso still gives the best estimation performance under the heterogeneous designs as compared with alternative methods.

Figure 3.

Figure 3.

Estimation errors of the Ad hoc 1-transfer, Agg-Lasso, Lasso, Oracle Trans-Lasso, and Trans-Lasso with heterogeneous covariance matrices. The two rows correspond to configurations (i) and (ii), respectively. The y-axis corresponds to bβ22 for some estimator b.

6. Application to Genotype-Tissue Expression Data

In this section, we demonstrate the performance of our proposed transfer learning algorithm in analyzing the Genotype-Tissue Expression (GTEx) data (https://gtexportal.org/). Overall, the data sets measure gene expression levels from 49 tissues of 838 human donors, in total comprising 1,207,976 observations of 38,187 genes. In our analysis, we focus on genes that are related to the central nervous system (CNS), which were assembled as MODULE 137 (https://www.gsea-msigdb.org/gsea/msigdb/cards/MODULE_137.html). This module includes a total of 545 genes and additional 1,632 genes that are significantly enriched in the same experiments as the genes of the module. A complete list of genes can be found at http://robotics.stanford.edu/~erans/cancer/modules/module_137.

6.1. Data Analysis Method

It is of biological interest to understand the CNS gene regulations in different tissues/cell types. Statistically, we consider predicting the expression levels of a target gene using other CNS genes in multiple tissues. Such an analysis provides insights on how other genes regulate the expression of a target gene. To demonstrate the replicability of our proposal, we consider multiple target genes and multiple target tissues and estimate their corresponding models one by one.

For an illustration of the computation process, we consider gene JAM2 (Junctional adhesion molecule B), as the response variable. JAM2 is a protein coding gene on chromosome 21 interacting with a variety of immune cell types and may play a role in lymphocyte homing to secondary lymphoid organs (Johnson-Léger et al., 2002). Mutations in JAM2 has been found to cause primary familial brain calcification (Cen et al., 2020; Schottlaender et al., 2020). We consider the association between JAM2 and other CNS genes in a brain tissue as the target models and the association between JAM2 and other CNS genes in other tissues as the auxiliary models. As there are multiple brain tissues in the dataset, we treat each of them as the target at each time. The list of target tissues can be found in Figure 4. The min, average, and max of primary sample sizes in these target tissues are 126, 177, and 237, respectively. More information on the target tissues is given in the Supplementary Materials. JAM2 expresses in 49 tissues in our dataset and we use 47 tissues with more than 120 measurements on JAM2. The average number of auxiliary samples for each target model is 14,837 over all the non-target tissues. The covariates in use are the genes that are in the enriched MODULE_137 and do not have missing values in all of the 47 tissues. The final covariates include a total of 1,079 genes. The data is standardized before analysis.

Figure 4.

Figure 4.

Prediction errors of Agg-Lasso, Naive Trans-Lasso, Trans-Lasso, and Ad hoc 1-transfer relative to the Lasso evaluated via 5-fold cross validation for gene JAM2 in multiple tissues.

We compare the prediction performance of Trans-Lasso with Lasso, Agg-Lasso, Ad hoc ℓ1-transfer, and Naive Trans-Lasso. Implementation of the first four methods is the same as in Section 5. The Naive Trans-Lasso implements the Oracle Trans-Lasso algorithm assuming all the auxiliary studies are informative. Evaluating this method can help us understand the overall informative level of the auxiliary samples. We split the target sample into five folds and use four folds to train the algorithms and use the remaining fold to test their prediction performance. We repeat this process five times each with a different fold of test samples. We mention that one individual can provide expression measurements on multiple tissues and these measurements are hard to be independent. While the dependence of the samples can reduce the efficiency of the estimation algorithms, using auxiliary samples may still be beneficial. However, one need to choose proper tuning parameters. The tuning parameter for the Lasso and λw are chosen by 8-fold cross-validation. The tuning parameter λδ is set to be λwkAnk/n0. Other tuning parameters and configurations are the same as for the simulations.

6.2. Prediction Performance of the Trans-Lasso for JAM2 Expression

Figure 4 demonstrates the prediction errors of different methods for predicting gene expression JAM2 using other genes. We see that all the transfer learning methods in consideration make improvements over the Lasso in most experiments. The performance of Naive Trans-Lasso implies that there is heterogeneity among tissues and some auxiliary studies can be non-informative. Hence, adaptation to unknown A is important. Among the adaptive transfer learning methods, Trans-Lasso achieves the smallest prediction errors in almost all the experiments. Its average gain is 22% comparing to the Lasso. This shows that our characterization of the similarity between a target model and a given auxiliary model is suitable for the current problem. Agg-Lasso gives similar prediction errors as Trans-Lasso in most of the tissues but has significantly worse performance for Cortex, Hippocampus, and Pituitary tissues. The average proportion of explained variance given by the Lasso and that given by the Trans-Lasso are 0.75 and 0.80, respectively, indicating improved fit from transfer learning.

6.3. Prediction Performance of Other 25 Genes on Chromosome 21

To demonstrate the replicability of our proposal, we also consider other genes on chromosome 21 which are in Module_137 as our target genes. We report the overall prediction performance of these 25 genes in Figure 5. A complete list of these genes and some summary information can be found in the Supplementary Materials. Generally speaking, we see that the Trans-Lasso has the best overall performance among all the target tissues when compared to the other two related methods, Agg-Lasso and Ad hoc 1-transfer. The deteriorating performance of the naive Trans-Lasso implies that adaptation to the unknown informative set is crucial for successful knowledge transfer.

Figure 5.

Figure 5.

Prediction errors of Ad hoc 1-transfer, Agg-Lasso, Naive Trans-Lasso*, and Trans-Lasso relative to the Lasso for the 25 genes on chromosome 21 and in Module 137, in multiple target tissues. The Naive Trans-Lasso has two outliers for the tissue Cerebellum not showing in the figure with values 1.61 and 1.95.

7. Discussion

This paper studies high-dimensional linear regression in the presence of auxiliary samples. The similarity of the target model and a given auxiliary model is characterized by the sparsity of their contrast vectors. Transfer learning algorithms for estimation and prediction are developed that are adaptive to the unknown informative set. Numerical experiments and GTEx data analysis support the theoretical findings and demonstrate its effectiveness in applications.

In the machine learning literature, transfer learning methods have been proposed for different purposes, but few have statistical guarantees. There are several interesting problems related to the present paper for further research. First, transfer learning in nonlinear models can be studied. Using our similarity characterization of the auxiliary studies, transfer learning in high-dimensional generalized linear models (GLMs) can be formulated. GLMs include logistic and Poisson models that are widely used for classification. The main challenge is that the moment equation above (7) is nonlinear and the resulting δA is not necessarily sparse. Hence, transfer learning beyond linear models remain open problems and can be studied under different characterizations for the similarity structure. Second, it is interesting to study statistical inference, such as constructing confidence intervals and hypothesis testing with auxiliary samples. Given the results derived in this paper, one may expect weaker sample size conditions in the transfer learning setting than those in the single-task setting. It is interesting to provide a precise characterization and to develop a minimax optimal confidence interval in the transfer learning setting.

Supplementary Material

supinfo

Acknowledgments

This research was supported by NIH grants GM129781 and GM123056 and NSF Grant DMS-1712735.

Contributor Information

Sai Li, Department of Biostatistics, Epidemiology and Informatics, Perelman School of Medicine, University of Pennvania, Philadelphia, PA 19104.

T. Tony Cai, Department of Statistics, The Wharton School, University of Pennsylvania, Philadelphia, PA 19104.

Hongzhe Li, Department of Biostatistics, Epidemiology and Informatics, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104.

References

  1. Agarwal A, Negahban S, and Wainwright MJ (2012). Noisy matrix decomposition via convex relaxation: Optimal rates in high dimensions. The Annals of Statistics 40(2), 1171–1197. [Google Scholar]
  2. Ahmed T and Bajwa WU (2019). Exsis: Extended sure independence screening for ultrahigh-dimensional linear models. Signal Processing 159, 33–48. [Google Scholar]
  3. Ando RK and Zhang T (2005). A framework for learning predictive structures from multiple tasks and unlabeled data. Journal of Machine Learning Research 6, 1817–1853. [Google Scholar]
  4. Bastani H (2020). Predicting with proxies: Transfer learning in high dimension. Management Science 67(5), 2657–3320. [Google Scholar]
  5. Bühlmann P and van de Geer S (2015). High-dimensional inference in misspecified linear models. Electronic Journal of Statistics 9(1), 1449–1473. [Google Scholar]
  6. Cai TT and Wei H (2021). Transfer learning for nonparametric classification: Minimax rate and adaptive classifier. The Annals of Statistics 49(1), 100–128. [Google Scholar]
  7. Candes E and Tao T (2007). The dantzig selector: Statistical estimation when p is much larger than n. The annals of Statistics 35(6), 2313–2351. [Google Scholar]
  8. Cen Z, Chen Y, Chen S, et al. (2020). Biallelic loss-of-function mutations in jam2 cause primary familial brain calcification. Brain 143(2), 491–502. [DOI] [PubMed] [Google Scholar]
  9. Chen X, Kim S, Lin Q, Carbonell JG, and Xing EP (2010). Graph-structured multi-task regression and an efficient optimization method for general fused lasso. arXiv preprint arXiv:1005.3579. [Google Scholar]
  10. Cross-Disorder Group of the Psychiatric Genomics Consortium (2019). Genomic relationships, novel loci, and pleiotropic mechanisms across eight psychiatric disorders. Cell 179(7), 1469–1482. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Dai D, Han L, Yang T, and Zhang T (2018). Bayesian Model Averaging with Exponentiated Least Squares Loss. IEEE Transactions on Information Theory 64(5), 3331–3345. [Google Scholar]
  12. Dai D, Rigollet P, and Zhang T (2012). Deviation optimal learning using greedy q-aggregation. The Annals of Statistics 40(3), 1878–1905. [Google Scholar]
  13. Danaher P, Wang P, and Witten DM (2014). The joint graphical lasso for inverse covariance estimation across multiple classes. Journal of the Royal Statistical Society. Series B (Statistical methodology) 76(2), 373–397. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Daumé III H (2007). Frustratingly easy domain adaptation. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, pp. 256–263. [Google Scholar]
  15. Dondelinger F, Mukherjee S, and Initiative ADN (2020). The joint lasso: high-dimensional regression for group structured data. Biostatistics 21(2), 219–235. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Fagny M, Paulson JN, Kuijjer ML, et al. (2017). Exploring regulation in tissues with eqtl networks. Proceedings of the National Academy of Sciences 114(37), E7841–E7850. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Fan J and Li R (2001). Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American statistical Association 96(456), 1348–1360. [Google Scholar]
  18. Fan J and Lv J (2008). Sure independence screening for ultrahigh dimensional feature space. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 70(5), 849–911. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Hu Y, Li M, Lu Q, et al. (2019). A statistical framework for cross-tissue transcriptome-wide association analysis. Nature genetics 51(3), 568–576. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Johnson-Léger CA, Aurrand-Lions M, Beltraminelli N, et al. (2002). Junctional adhesion molecule-2 (jam-2) promotes lymphocyte transendothelial migration. Blood, The Journal of the American Society of Hematology 100(7), 2479–2486. [DOI] [PubMed] [Google Scholar]
  21. Lee SH, Ripke S, Neale BM, et al. (2013). Genetic relationship between five psychiatric disorders estimated from genome-wide snps. Nature genetics 45, 984–994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Li S, Cai TT, and Li H (2020). Supplements to “Transfer learning for high-dimensional linear regression: Prediction, estimation, and minimax optimality”. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Li YR, Li J, Zhao SD, and et al. (2015). Meta-analysis of shared genetic architecture across ten pediatric autoimmune diseases. Nature Medicine 21, 1018–1027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Liu Y and Kozubowski TJ (2015). A folded laplace distribution. Journal of Statistical Distributions and Applications 2(1), 1–17. [Google Scholar]
  25. Lounici K, Pontil M, and Tsybakov AB (2009). Taking advantage of sparsity in multi-task learning. arXiv:0903.1468. [Google Scholar]
  26. Mak TSH, Porsch RM, Choi SW, Zhou X, and Sham PC (2017). Polygenic scores via penalized regression on summary statistics. Genetic Epidemiology 41(6), 469–480. [DOI] [PubMed] [Google Scholar]
  27. Mei S, Fei W, and Zhou S (2011). Gene ontology based transfer learning for protein subcellular localization. BMC bioinformatics 12, 44. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Pan W and Yang Q (2013). Transfer learning in heterogeneous collaborative filtering domains. Artificial intelligence 197, 39–55. [Google Scholar]
  29. Pierson E, Koller D, Battle A, et al. (2015). Sharing and specificity of co-expression networks across 35 human tissues. PLoS Comput Biol 11(5), e1004220. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Raskutti G, Wainwright MJ, and Yu B (2011). Minimax rates of estimation for high-dimensional linear regression over q-balls. IEEE transactions on information theory 57(10), 6976–6994. [Google Scholar]
  31. Rigollet P and Tsybakov A (2011). Exponential screening and optimal rates of sparse estimation. The Annals of Statistics 39(2), 731–771. [Google Scholar]
  32. Schottlaender LV, Abeti R, Jaunmuktane Z, et al. (2020). Bi-allelic jam2 variants lead to early-onset recessive primary familial brain calcification. The American Journal of Human Genetics 106(3), 412–421. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Shin H-C, Roth HR, Gao M, et al. (2016). Deep convolutional neural networks for computer-aided detection: Cnn architectures, dataset characteristics and transfer learning. IEEE transactions on medical imaging 35(5), 1285–1298. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Sun T and Zhang C-H (2012). Scaled sparse linear regression. Biometrika 99(4), 879–898. [Google Scholar]
  35. Sun YV and Hu Y-J (2016). Integrative analysis of multi-omics data for discovery and functional studies of complex human diseases. In Advances in genetics, Volume 93, pp. 147–190. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Tibshirani R (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological) 58(1), 267–288. [Google Scholar]
  37. Torrey L and Shavlik J (2010). Transfer learning. In Handbook of research on machine learning applications and trends: algorithms, methods, and techniques, pp. 242–264. IGI Global. [Google Scholar]
  38. Tsybakov AB (2014). Aggregation and minimax optimality in high-dimensional estimation. In Proceedings of the International Congress of Mathematicians, Volume 3, pp. 225–246. [Google Scholar]
  39. Turki T, Wei Z, and Wang JT (2017). Transfer learning approaches to improve drug sensitivity prediction in multiple myeloma patients. IEEE Access 5, 7381–7393. [Google Scholar]
  40. Verzelen N (2012). Minimax risks for sparse regressions: Ultra-high dimensional phenomenons. Electronic Journal of Statistics 6, 38–90. [Google Scholar]
  41. Wang S, Shi X, Wu M, and Ma S (2019). Horizontal and vertical integrative analysis methods for mental disorders omics data. Scientific Reports, 1–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Weiss K, Khoshgoftaar TM, and Wang D (2016). A survey of transfer learning. Journal of Big Data 3, 9. [Google Scholar]
  43. Zhang C-H (2010). Nearly unbiased variable selection under minimax concave penalty. The Annals of statistics 38(2), 894–942. [Google Scholar]
  44. Zhernakova A, Van Diemen CC, and Wijmenga C (2009). Detecting shared pathogenesis from the shared genetics of immune-related diseases. Nature Reviews Genetics 10(1), 43–55. [DOI] [PubMed] [Google Scholar]
  45. Zou H (2006). The adaptive lasso and its oracle properties. Journal of the American statistical association 101(476), 1418–1429. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

supinfo

RESOURCES