Skip to main content
Bioinformatics logoLink to Bioinformatics
. 2023 Nov 10;39(12):btad680. doi: 10.1093/bioinformatics/btad680

Penalized regression with multiple sources of prior effects

Armin Rauschenberger 1,, Zied Landoulsi 2, Mark A van de Wiel 3,2, Enrico Glaab 4,2
Editor: Jonathan Wren
PMCID: PMC10699841  PMID: 37951587

Abstract

Motivation

In many high-dimensional prediction or classification tasks, complementary data on the features are available, e.g. prior biological knowledge on (epi)genetic markers. Here we consider tasks with numerical prior information that provide an insight into the importance (weight) and the direction (sign) of the feature effects, e.g. regression coefficients from previous studies.

Results

We propose an approach for integrating multiple sources of such prior information into penalized regression. If suitable co-data are available, this improves the predictive performance, as shown by simulation and application.

Availability and implementation

The proposed method is implemented in the R package transreg (https://github.com/lcsb-bds/transreg, https://cran.r-project.org/package=transreg).

1 Background

For many biomedical prediction or classification studies, there is a previous study with a similar target and a similar high-dimensional feature space, e.g. hundreds of microRNAs (miRNAs), thousands of genes, or millions of single-nucleotide polymorphisms (SNPs). Given a trained model from a previous study, we could use it to obtain predicted values or predicted probabilities for the study of interest, but these predictions are only reliable if the two studies have the same target, the same features, and the same population. However, we expect the feature-target effects from two studies to be strongly correlated in more situations: slightly different targets (e.g. disease status versus disease stage), slightly different features (e.g. imperfectly overlapping feature space, different measurement technique), slightly different populations (e.g. hospitalized versus non-hospitalized patients), or even different modelling approaches (e.g. simple regression versus multiple regression). As it is challenging to estimate feature-target effects in high-dimensional settings, it might be advantageous to use results from previous studies as prior information for the study of interest.

Consider two prediction or classification problems, each one with a target vector and a feature matrix (samples in the rows, features in the columns). Suppose that both feature matrices cover the same features (each column in the first matrix corresponds to a column in the second matrix). In two special cases, the two problems reduce to a single problem: (i) if both problems have the same target and concern samples from the same population, they are in essence one ‘single-target’ problem (combine target vectors and feature matrices by rows, respectively), potentially with batch effects; and (ii) if both problems concern the same samples, they are in essence one ‘multi-target’ problem (combine target vectors by columns, feature matrices are the same). In other cases, however, we might be in a transfer learning setting (Table 1).

Table 1.

Abstract representation of the dataset of interest (without asterisk, black) and an additional dataset (with asterisk, grey).a

Single-target learningMulti-target learningTransfer learning(y1yny1*ym*)(x11x1pxn1xnpx11*x1p*xm1*xmp*)(y1y1*ynyn*)(x11x1pxn1xnp)(y1yn)(x11x1pxn1xnp)            (y1*ym*)(x11*x1p*xm1*xmp*)
a

Single-target learning (left): same targets, same features*, different samples (from one population). Multi-target learning (centre): different targets, same features*, same samples. Transfer learning (right): same or different targets, matched features*, different samples (from one or two populations). *Partially overlapping feature spaces are also possible. See, for example, Rauschenberger and Glaab (2021) for multi-target learning or Section 2.8 “Extension” for transfer learning.

In such settings—two or more regression problems with related targets and matched features—it might be possible to transfer information from one problem to another. Transfer learning, in contrast to multi-target regression and seemingly unrelated regressions, addresses related regression problems not for the same but for different samples. If the regression problems are sufficiently related to each other, we expect their regression coefficients to be correlated (positively or negatively). When fitting the regression model of interest, we could therefore account for the estimated regression coefficients from the other model. Transferring information on the importance and the direction of the feature effects, we could potentially increase the predictive performance.

Jiang et al. (2016) proposed the prior lasso to account for prior information in high-dimensional predictive modelling. Their method involves a preprocessing step and a weighting step. In the preprocessing step, the prior information is used to predict the target from the features. They present a solution for one set of prior effects from a closely related study (multiplying the feature matrix by the prior effects), but extensions to multiple sets of prior effects or loosely related studies may be feasible. Let y represent the target and let y^prior represent the fitted values based on the prior information. In the weighting step, they minimize the penalized combined likelihood L(x,y;β)+ηL(x,y^prior;β)ρ(λ;β) with respect to the coefficients β, where η0 (balance) and λ0 (regularization). If the balancing hyperparameter η is larger than zero, the prior predictions y^prior influence the estimation of the parameters β.

Dhruba (2021) proposed a transfer learning method based on distribution mapping. Even if features or targets follow different distributions in two datasets, it is possible to build a predictive model using the first dataset and make predictions for the second dataset. Requiring matched features and targets in the source dataset and unmatched features and targets in the target dataset, their method transfers (i) features from the target to the source domain and (ii) predictions from the source to the target domain. In contrast, we consider transfer learning settings with matched features and targets in the target dataset.

Tian and Feng (2022) proposed and implemented transfer learning for ridge and lasso regression. Their transfer learning algorithm involves two steps: (i) estimating common coefficients for the target dataset and the transferable source datasets (ω^) and (ii) estimating the deviations from the common coefficients to the target coefficients (δ^). Both steps together lead to the estimated target coefficients (β^=ω^+δ^). Before applying their transfer learning algorithm, Tian and Feng (2022) apply a transferable source detection algorithm to exclude source datasets that are too different from the target dataset. This avoids that non-transferable sources render the common coefficients misleading for the target dataset (‘negative transfer’). In the case of lasso regularization in the two steps, there is sparsity in the common estimates as well as in the deviations from the common estimates to the target estimates (and thereby also in the target estimates).

The method from Tian and Feng (2022) requires not only the target dataset but also the source dataset(s). However, data protection regulations or restrictive data sharing policies might prevent researchers from accessing a source dataset, or the available storage or processing capacity might be insufficient for analysing massive source datasets. Although federated transfer learning (Liu et al. 2020) enables the joint analysis of related datasets without sharing them, it is often not practically feasible for biomedical researchers to set up corresponding collaborative analyses with multiple data holders. There is therefore a need for transfer learning methods that require neither direct nor indirect access to the source data but only to the complementary data (co-data) derived from the source data. Such methods allow us to exploit publicly available summary statistics from external studies, e.g. P-values and effect sizes from a genome-wide association study (GWAS), to increase the predictive performance in the study of interest.

We propose a two-step transfer learning method, modelling with and without co-data in the first step and combining different models in the second step. Unless the source and target datasets are very similar, the coefficients from the source dataset(s) will not fit well to the target dataset. We therefore propose to calibrate these coefficients—preserving their signs and their order—so that they can be transferred from the source dataset(s) to the target dataset. Additionally, we also estimate the coefficients directly from the target dataset, ignoring the co-data. The calibrated coefficients from the source dataset(s) as well as the estimated coefficients from the target dataset allow us to predict the outcome from the features. Finally, we combine the linear predictors from the models with and without co-data and calculate either predicted values (linear regression) or predicted probabilities (logistic regression).

Kawaguchi et al. (2022) proposed a related transfer learning method. This method penalizes (i) the differences between the coefficients and weighted sums of the prior effects (one difference for each feature) and (ii) the weights for these weighted sums (one weight for each set of prior effects), possibly with two different penalties (e.g. ridge and lasso). It not only allows for high-dimensional data (i.e. more features than samples) but also for high-dimensional co-data (i.e. more sets of prior effects than samples). When the unknown feature effects are non-linearly related to an important source of prior effects, however, this method might leave room for improvement.

In a related transfer learning setting, prior information is only available on the importance but not on the direction of the feature effects, i.e. with complementary data consisting of prior weights rather than prior effects. In the generalized linear model framework, the weighted lasso (Bergersen et al. 2011), the feature-weighted elastic net (Tay et al. 2023), and penalized regression with differential shrinkage (Zeng et al. 2021) account for prior weights in the penalty function, through feature-specific penalty factors or feature-specific regularization parameters. Adaptive group-regularized ridge regression (van de Wiel et al. 2016) is not only applicable to categorical co-data but also to numerical co-data (prior weights), by the means of creating groups of features from numerical co-data and forcing the group-penalties to be monotonically decreasing. An extension from van Nee et al. (2021) makes this approach even more suitable for numerical co-data. For single sources of co-data, it might be possible to extend these methods to prior information on the importance as well as the direction of feature effects by imposing sign constraints on the coefficients. Prior weights have not only been exploited in regression analysis, e.g. co-data moderated random forests (te Beest et al. 2017) adapt the sampling probabilities of the features to the prior weights.

2 Materials and methods

2.1 Model

Suppose one target and p features are available for n samples, with many more features than samples (pn). We index the samples by i in {1,,n} and the features by j in {1,,p}. Our aim is to estimate the generalized linear model:

E[yi]=h1(β0+j=1pβjxij) .

For any sample i, the model expresses the expected value of its target (yi) as a function of its features (xi1,,xip). The link function h(·) depends on the family of distributions for the target (Gaussian: identity, binomial: logit, Poisson: log). In the linear predictor, β0 represents the unknown intercept, and βj represents the unknown slope of feature j (i.e. the effect of the feature on the linear predictor of the target). Given the estimated intercept β^0 and the estimated slopes {β^1,,β^p}, we could predict the target of previously unseen samples:

y^i=h1(β^0+j=1pβ^jxij) .

2.2 Co-data

Suppose m sources of co-data are available, indexed by k in {1,,m}, with many fewer sources than samples (mn). Let zjk indicate the prior effect from source k for feature j. Our method is designed for quantitative co-data that provide an insight into the importance (absolute value) and the direction (sign) of the feature effects. Each set of prior effects (1z°k1) is assumed to be positively correlated with the true coefficients (cor(z°k,β)>0). For any source of co-data, the prior effects may be re-scaled (not re-centred), for example to the interval from −1 to +1. In other words, the proposed method is invariant under multiplication of the prior effects by a positive scalar (zc×z where c >0). We explain in the next section why this is important.

It might seem trivial to also allow for co-data that only provide an insight into the importance but not the direction of the feature effects (i.e. prior weights instead of prior effects). Each set of prior weights (0z°k1) is assumed to be positively correlated with the true absolute coefficients (cor(z°k,|β|)>0). To obtain prior effects, one might want to assign the signs of the Spearman correlation coefficients between the target and the features to the prior weights. However, marginal effects and conditional effects can have opposite signs. If we wanted to extend our approach to prior weights, we would have to discover the signs inside the calibration procedure (see below), which would be related to high-dimensional regression with binary coefficients (Gamarnik and Zadik 2017).

2.3 Base-learners with co-data

Suppose we are in a transfer learning setting with two prediction or classification problems. For simplicity, we assume that the features do not differ in scale between the two problems. For illustration, we consider two artificial situations (where we would not use transfer learning in practice): (i) if both problems concern the same target on the same scale and the samples come from the same population, we could use the estimated regression coefficients from one problem to make predictions for the other problem; and (ii) if the two problems concern the same target on different scales, we could also recycle the estimated regression coefficients, but we would have to adjust for the different scales.

When transferring estimated regression coefficients from one problem to another problem, it might not only be necessary to change their scale but it might also be beneficial to change their shape. For example, it might be that for one problem weak and strong effects matter, while for the other problem only strong effects matter. We should therefore also be able to make differences between small coefficients more important or less important than those between large coefficients. We propose two calibration methods, namely exponential and isotonic calibration, to adapt the prior information to the data. For each source of co-data k, both calibration methods estimate the model:

E[yi]=h1(αk+j=1pγjkxij) ,

where the calibrated prior effects {γ^1k,,γ^pk} depend on the initial prior effects {z1k,,zpk}. The difference between exponential and isotonic calibration is how the former depend on the latter.

  • Exponential calibration: Let γjk=θksign(zjk)|zjk|τk, for j in {1,,p}, where the factor θk and the exponent τk are non-negative real numbers (θk0 and τk0). We first fit one simple non-negative regression for different values of τk (i.e. estimate αk and θk given τk), and then optimize τk. Once θk and τk have been estimated, the initial prior effects zjk determine the final prior effects γ^jk=θ^ksign(zjk)|zjk|τ^k, for all j in {1,,p}. The estimated factor θ^k and the estimated exponent τ^k allow the model to change the scale and the shape of the prior effects. For example, θ^k=0 sets them to zero, |θ^k|<1 makes them smaller, |θ^k|>1 makes them larger, τ^k=0 sets them to the same value, τ^k<1 makes (absolutely) large ones more similar, and τ^k>1 makes (absolutely) small ones more similar. If one or more sets of prior effects might be negatively associated with the true coefficients, we could remove the non-negativity constraints from the simple regressions (allowing θ^k<0 to invert the signs of the prior effects).

  • Isotonic calibration: We estimate {γ1k,,γpk} under the constraint that the signs of the initial prior effects zjk determine the signs of the final prior effects γ^jk (i.e. γ^jk=0|zjk=0, γ^jk0|zjk>0,γ^jk0|zjk<0) and under the constraint that the order of the initial prior effects determines the order of the final prior effects (i.e. γ^jkγ^lk|zjkzlk,γ^jkγ^lk|zjkzlk), for all j and l in {1,,p}. If one or more sets of prior effects might be negatively associated with the true coefficients, we could fit each model with these constraints and the inverted constraints, and then select the better fit.

  • To make optimization more efficient, we rewrite the sign- and order-constrained problem as a sign-constrained problem (see Table 2). For each source of co-data, we order the columns of the feature matrix by increasing values of the prior effects. Suppose the first q columns correspond to negative prior effects and the last p–q columns correspond to non-negative prior effects. We take the cumulative sum of the feature columns from left to right for the former (columns 1 to q) and from right to left for the latter (columns p to q +1). We then estimate the coefficients on the left under non-positivity constraints, and those on the right under non-negativity constraints. Formally, the model equals
    E[yi]=h1(αk+j=1pδjkwij) ,

Table 2.

Isotonic calibration.a

(i)x°,1,z1,kx°,2,z2,kx°,q1,zq1,kx°,q,zq,k(ii)x°,(1)x°,(2)x°,(q1)x°,(q)(iii)w°,1=x°,(1)w°,2=x°,(1)+x°,(2)w°,q1=x°,(1)++x°,(q1)w°,q=x°,(1)++x°,(q)(iii)*δ^1,kδ^2,kδ^q1,kδ^q,k(ii)*γ^(1),k=δ^1,k++δ^q,kγ^(2),k=δ^2,k++δ^q,kγ^(q1),k=δ^q1,k+δ^q,kγ^(q),k=δ^q,k(i)*γ^1,kγ^2,kγ^q1,kγ^q,k

(i)x°,q+1,zq+1,kx°,q+2,zq+2,kx°,p1,zp1,kx°,p,zp,k(ii)x°,(q+1)x°,(q+2)x°,(p1)x°,(p)(iii)w°,q+1=x°,(q+1)++x°,(p)w°,q+2=x°,(q+2)++x°,(p)w°,p1=x°,(p1)+x°,(p)w°,p=x°,(p)(iii)*δ^q+1,kδ^q+2,kδ^p1,kδ^p,k(ii)*γ^(q+1),k=δ^q+1,kγ^(q+2),k=δ^q+1,k+δ^q+2,kγ^(p1),k=δ^q+1,k++δ^p1,kγ^(p),k=δ^q+1,k++δ^p,k(i)*γ^q+1,kγ^q+2,kγ^p1,kγ^p,k
a

The aim is to (i) estimate the effects of the features under sign and order constraints determined by q negative (top) and p–q non-negative (bottom) prior effects, i.e. estimate γ1k,,γpk for x1,,xp under γ^jk=0|zjk=0, γ^jk0|zjk>0,γ^jk0|zjk<0,γ^jkγ^lk|zjkzlk, and γ^jkγ^lk|zjkzlk. This can be solved by (ii) estimating the effects of the features ordered by the co-data under sign and order constraints, i.e. estimate γ(1),k,,γ(p),k for x(1),,x(p) under γ^(j),k0|jp,γ^(j),k0|j>p, and γ^(1),kγ^(p),k. This in turn can be solved by (iii) estimating the effects of the combined features under sign constraints, i.e. estimate δ1,,δp for w1,,wp under δ^j0|jp and δ^j0|j>p. Our algorithm receives the original features and the prior effects (i), orders the features by the prior effects (ii), combines the features (iii), estimates the effects of the combined features (iii)*, calculates the estimated effects of the ordered features (ii)*, and returns the estimated effects of the original features (i)*. *A row with an asterisk contains the estimates for the features in the row with the same number but without an asterisk.

where wij=l=1jxi(l) and δjk0 for j in {1,,q}, and wij=l=jpxi(l) and δjk0 for j in {q+1,,p}, with the subscript within brackets indicating the order of the prior effects. The linear predictor of the sign-constrained model, i.e. αk+j=1pδjkwij, is equivalent to the linear predictor of the order-constrained model, i.e. αk+j=1pγ(j)kxi(j), because

j=1qδjkwij=j=1qδjk(l=1jxi(l))=j=1q(l=jqδlk)xi(j)=j=1qγ(j)kxi(j) ,j=q+1pδjkwij=j=q+1pδjk(l=jpxi(l))=j=q+1p(l=q+1jδlk)xi(j)=j=q+1pγ(j)kxi(j) .

After estimating the coefficients of the sign-constrained model by maximum likelihood, we therefore estimate those of the order-constrained model by γ^(j)k=l=jqδ^lk for j in {1,,q} and γ^(j)k=l=q+1jδ^lk for j in {q+1,,p}.

While exponential calibration involves three unknown parameters, namely the intercept αk, the factor θk, and the exponent τk, isotonic calibration involves 1+p unknown parameters, namely the intercept αk and the slopes γk={γ1k,,γpk}, for each set of co-data. Figure 1 shows the difference between exponential and isotonic calibration in several empirically assessed scenarios.

Figure 1.

Figure 1.

Final prior effects γ (y-axis) against initial prior effects z (x-axis), under exponential calibration (red points on continuous curve) and isotonic calibration (blue points on discontinuous curve). The thin black line corresponds to perfect calibration (γ=β). We simulated the feature matrix X from a standard Gaussian distribution (n =200, p =500) and the initial prior effects z from a trimmed standard Gaussian distribution (trimmed below the 1% and above the 99% quantile). We set the true coefficients to (1) β=z, (2) β=sign(z)|z|, (3) β=sign(z)z2, (4) β=I[z>0]z, (5) β=I[z>1], or (6) β=I[z0]|z|+I[z>0]z2. And we simulated the response vector y from Gaussian distributions with the means η and the variance Var(η), where η=. While exponential calibration performs slightly better in the first three scenarios (top), isotonic calibration performs much better in the last three scenarios (bottom).

After calibration, we pre-assess the utility of each set of co-data. To do this, we calculate the residuals (depending on the family of distributions) between the fitted and the observed targets. We suggest to retain a set of co-data only if the residuals are significantly smaller than those of the intercept-only model (one-sided Wilcoxon signed-rank test) at the nominal 5% level (P-value .05).

2.4 Base-learners without co-data

We also fit the model without any co-data. We estimate the coefficients by maximizing the penalized likelihood:

β^=argmaxβ{L(x;β)ρ(λ;β)} ,

where L(x,β) is the likelihood and ρ(λ;β) is the penalty. The likelihood depends on the family of distributions (Gaussian, binomial, Poisson), and the penalty can be the ridge (L2) or the lasso (L1) penalty. The penalty shrinks the squared (ridge) or absolute (lasso) slopes {β1,,βp} towards zero (without penalizing the intercept β0). We denote the estimated intercept by β^0 and the estimated slopes by {β^1,,β^p}.

2.5 Cross-validation

We split the samples into 10 folds to perform 10-fold internal cross-validation. In each iteration, we fit the models to nine included folds and predict the target for the excluded fold.

Let the n × m matrix H^(0,cv) represent the feature-dependent part of the cross-validated linear predictors from the models with co-data. Specifically, the entry in row i (sample) and column k (source of co-data) equals

ηik(0,cv)=0×α^kκ(i)+j=1pγ^jkκ(i)xij ,

where the superscript κ(i) indicates that the (ignored) intercept αk and the slopes γjk for j in {1,,p} are estimated without using the fold of sample i, as in Rauschenberger and Glaab (2021).

The models without any co-data do not only have 1+p unknown parameters, namely the intercept β0 and the slopes {β1,,βp}, but also the unknown hyperparameter λ. In each iteration, we fit this model for a decreasing sequence of 100 values for the regularization parameter λ, indexed by l in {1,,100}, using the computationally efficient approach from Friedman et al. (2010, glmnet).

Accordingly, let the n×100 matrix H^(1,cv) represent the cross-validated linear predictors from the model without co-data. Specifically, the entry in row i (sample) and column l (regularization parameter) equals

ηil(1,cv)=β^0κ(i),l+j=1pβ^jκ(i),lxij ,

where the superscripts κ(i) and l indicate that the intercept β0 and the slopes {β1,,βp} are estimated without using the fold of sample i and given the regularization parameter λl.

To optimize the predictive performance of the co-data independent model, we would select the λ that minimizes the cross-validated loss (λmin). As we base our predictions not only on the co-data independent model but also on the co-data dependent model(s), λmin might be too small. The reason is that the co-data might be informative to the extent that the co-data independent model requires more penalization. We could let the meta-learner select the optimal λ from the whole sequence, but this might render the inclusion and exclusion of co-data dependent models unstable. Our ad hoc solution is to include the optimal regularization parameter for the co-data independent model (λmin) and a slightly larger one (λ1se). The latter is given by the one-standard-error rule, which increases λ until the cross-validated loss equals its minimum plus one standard error.

We concatenate H^(0,cv) with the columns of H^(1,cv) that correspond to λmin and λ1se to obtain the n×(m+2) matrix H^(cv). The first m columns correspond to the models with co-data, and the last two columns correspond to the model without co-data.

2.6 Meta-learner

We combine the base-learners with and without co-data by stacked generalization (Wolpert 1992), on the level of the linear predictors (Rauschenberger et al. 2021). In the meta-layer, we regress the target on the cross-validated linear predictors from the base-layer:

E[yi]=h1(ω0+k=1m+2ωkH^ik(cv)) .

Leaving the intercept unrestricted (<ω0<+) but imposing the lower bound zero on the slopes (ω10,,ωm+20), we estimate these coefficients under lasso regularization. Due to the feature selection property of the lasso, a source of co-data can be excluded (ω^k=0) or included (ω^k>0), where k is in {1,,m}. Similarly, the models without co-data can be excluded (ω^k=0) or included (ω^k>0), where k=m+1 for the model with λmin and k=m+2 for the model with λ1se. The estimated slopes function as weights for the co-data dependent models (ω^1,,ω^m) and for the co-data independent models (ω^m+1,ω^m+2). Thus, we do not only select sources but also weight them according to their relevance.

2.7 Interpretation

The coefficients β^min and β^1se give insight into the feature-target effects estimated without co-data, with β^min,j and β^1se,j representing the effect of feature j, where j is in {1,,p}. The coefficients ω^ give insight into the importance of the sources of co-data, with ω^k representing the importance of source k, where k is in {1,,m}. For a previously unseen sample i, the predicted value is:

y^i=h1(ω^0+k=1m+2ω^kH^ik)=h1(β^0+j=1pβ^jxij) ,whereβ^0=ω^0+ω^m+1β^min,0+ω^m+2β^1se,0andβ^j=(k=1mω^kγ^jk)+ω^m+1β^min,j+ω^m+2β^1se,j . (1)

Thus, the estimated effect for a feature (β^j) is a weighted sum of estimated coefficients with co-data (γ^j1,,γ^jm) and the estimated coefficients without co-data (β^min,j,β^1se,j).

Sparse models (few non-zero coefficients) are often considered to be more interpretable than dense models (many non-zero coefficients). While the original coefficients are dense (j=1pI[β^j0]=p) or sparse (j=1pI[β^j0]p) depending on the choice between ridge and lasso regularization, the weights may contain some zeros due to significance filtering or lasso regularization (k=1m+2I[ω^k0]m+2). As soon as one set of dense prior effects is selected, however, the combined coefficients also become dense (j=1pI[β^j0]p). This means that the feature selection property of the lasso is not maintained. We should therefore choose between ridge and lasso regularization (i) to make the model without co-data more predictive or interpretable (ii) or to make the model with co-data more predictive (iii) but not to make the model with co-data more interpretable.

2.8 Extension

In some applications, prior information might be available and reliable for some features but missing or unreliable for other features. Note that a missing prior effect can be interpreted as an unreliable prior effect of zero. Although the base-learners with co-data might still be predictive, the meta-learner (weighted average of the base-learners with and without co-data) might be not more predictive than the base-learner without co-data. The reason is that the meta-learner assigns the same weight to all prior effects, rather than more weight to available and reliable prior effects and less weight to missing or unreliable prior effects. We therefore propose an alternative approach for applications with partially informative sources of co-data.

In the following, we use the term ‘meta-features’ for the cross-validated linear predictors from the base learners with co-data. Each meta-feature—one column of the n × m matrix H^(0,cv)—corresponds to one source of co-data. In the meta-layer, we regress the target on the meta-features and the base-features:

E[yi]=h1(β0+k=1mωkH^ik(0,cv)+j=1pβjxij) ,

with non-negativity constraints for the weights for the meta-features (ω10,,ωm0) but without constraints for the intercept (β0) and the slopes for the base-features (β1,,βp).

We estimate the weights for the meta-features and the slopes for the base-features using penalized maximum likelihood:

{ω^,β^}=argmax{ω,β}{L(x;ω,β)ρ(λ;β)} ,

where L(x;ω,β) is the likelihood and ρ(λ;β) is the penalty. We do not penalize the weights for the meta-features (mn) but only the slopes for the base-features (pn). The more sources of co-data are available, the more it becomes necessary to penalize their weights. But then the weights ω and the slopes β might need differential penalization, for example a lasso penalty for the meta-features (selection of sources) and a ridge penalty for the base-features (many small effects). To make this computationally efficient, we would need a fast cross-validation procedure for multiple penalties (cf. van de Wiel et al. 2021) with non-negativity constraints (meta-features) and mixed lasso and ridge penalization (meta-features versus base-features). This extension is therefore only applicable in settings with few sources of co-data.

The predicted value for a previously unseen sample i is

y^i=h1(β^0+k=1mω^kH^ik(0)+j=1pβ^jxij)=h1(β^0+j=1pβ^jxij) ,whereβ^0=β^0andβ^j=(k=1mω^kγ^jk)+β^j . (2)

As the coefficients β are shrunk towards zero but the coefficients ω are not penalized, the combined coefficients β are shrunk towards a weighted sum of the calibrated prior effects (γ^ω^). When the regularization parameter tends to infinity (λ), the estimated deviations from this weighted sum approach zero (β^0) and the combined estimates approach this weighted sum (β^γ^ω^). Lasso regularization ensures sparsity in the deviations from the calibrated prior effects (j=1pI[β^j0]p)—in contrast to ridge regularization—but not in the combined coefficients (j=1pI[βj0]  p). As the combined coefficients may deviate more from unreliable than from reliable calibrated prior effects, this extension is suitable for partially informative co-data. As opposed to ‘standard stacking’, we refer to this extension as ‘simultaneous stacking’. Algorithm 1 includes the pseudo-code for both approaches.

Algorithm 1.

Pseudo-code for the proposed transfer learning method with standard stacking (left) and simultaneous stacking (right). Sub-procedures (italicized) are explained below.

Algorithm 1.

  • Exp/Iso-Calibrate : Performs exponential or isotonic calibration (Section). Requires target vector, feature matrix, and prior effects. Ensures calibrated prior effects.

  • Ridge/Lasso-CV : Tunes the hyperparameter of ridge or lasso regression by k-fold cross-validation. Requires target vector, feature matrix, and fold identifiers. Ensures optimal regularization parameter.

  • Ridge/Lasso-Fit : Estimates the parameters of ridge or lasso regression. Requires target vector, feature matrix, and regularization parameter. Ensures estimated coefficients.

  • Sta/Sim-Combine: Combines estimated parameters from standard stacking (Equation 1) or simultaneous stacking (Equation 2). Requires estimated parameters from base-learners and meta-learner. Ensures combined estimates.

Note: The procedures Exp/Iso-Calibrate, Ridge/Lasso-CV, and Ridge/Lasso-Fit depend on the family of distributions (Gaussian versus binomial), but they also require choices (exponential versus isotonic, ridge versus lasso).

3 Simulation

We performed two simulation studies to compare the predictive performance between our transfer learning method and the one from Tian and Feng (2022). In contrast to the method from Tian and Feng (2022), which requires the feature-target effects in the target and the source dataset(s) to be positively correlated and on the same scale (i.e. βtargetβsource), our method also allows for negatively correlated effects and for effects on different scales (i.e. βtargetc×βsource). Although it is possible to overcome this restriction by inverting the target (Gaussian: ysourceysource, binomial: ysource1ysource), by re-scaling the target (Gaussian: ysource1/c×ysource), or by inverting or re-scaling the features (Xsourcec×Xsource), we believe it is more user-friendly to directly allow for negative correlations and different scales. To ensure a fair comparison between the two methods, we simulate positively correlated effects on the same scale. Furthermore, although the method from Tian and Feng (2022) is in theory also suitable for mixed response types, the current version of the related R package glmtrans requires the source dataset(s) and the target dataset to have the same response type (Gaussian, binomial, Poisson). We therefore always simulate the same response type in the source and target domains. In addition, we also compared our method with the one from Kawaguchi et al. (2022, xrnet).

3.1 External simulation

We use the simulation approach from Tian and Feng (2022). In each iteration, we call the function glmtrans::models with the arguments (1) family of distributions: family= ”gaussian” (default) or family=”binomial”, (2) source or target datasets: type=”all” (default), (3) difference between source and target coefficients: h=5 (default) or h=250, (4) number of source datasets: K=5 (default), (5) sample size for target dataset: n.target=100 (default), (6) sample size for each source dataset: n.source=150 (default), (7) number of non-zero coefficients: s=15 (default) or s=50, (8) number of features: p=1000 (default), number of transferable source datasets: Ka=1, Ka=3 or Ka=K=5 (default).

The simulation from Tian and Feng (2022) involves the following steps:

  • Features: The correlation between features i and j is set to Σij=0.5|ij|, where i and j are in {1,,p}. Let Σ represent the correlation matrix and let Σ=RR represent its Cholesky decomposition, where R is an upper triangular matrix. For the target dataset (n0=100) and each source dataset (n1==n5=150), the n0×p matrix X0=E0R and the nk×p matrices Xk=EkR for k in {1,,5} represent the features, where the n0×p matrix E0 and the nk×p matrices Ek contain Gaussian noise.

  • Coefficients: Let βj represent the effect of feature j, for j in {1,,p}, and denote the p-dimensional coefficient vectors by β0 for the target dataset and {β1,,β5} for the source datasets. For the target dataset, the first s elements are set to βj=0.5 (causal) and the last p–s elements are set to βj=0 (non-causal). For transferable source datasets, the first s elements are set to βj=0.5+(1)zjh/p and the last p–s elements are set to βj=(1)zjh/p, where zj is a realization of zjBernoulli(0.5). For non-transferable source datasets, the first s elements are set to βj=(1)zj2h/p (non-causal), the next s elements are set to βj=0.5+(1)zj2h/p (causal), and the last p2s elements are a random sample of s causal and p3s non-causal elements generated in the same way. To obtain a non-transferable source, it would be sufficient to randomly select causal elements rather than inverting causal and non-causal elements (indices 1 to 2s).

  • Targets: In the Gaussian case, the target vector is the n-dimensional vector y0=X0β0+ϵ0 for the target dataset, and the n-dimensional vector yk=0.5+Xkβk+ϵk for source dataset k, where the n-dimensional vectors {ϵ0,,ϵ5} contain Gaussian noise. In the binomial case, let p0=1/(1+exp(X0β0)) for the target dataset and pk=1/(1+exp(0.5Xkβk)) for the source datasets. The n-dimensional vectors {y0,,y5} are the target vectors, with each element following a Bernoulli distribution with the probability given by {p0,,p5}.

3.2 Internal simulation

The simulation from Tian and Feng (2022) uses the same effect size for all causal features and a decreasing correlation structure with a fixed base. We therefore designed our own data-generating mechanism (i) to simulate different effect sizes for different causal features (βjR instead of βj{0,0.5}) and (ii) to vary the strength of correlation between features (σij=ρx|ij| instead of σij=0.5|ij|).

Our simulation involves the following steps:

  • Features: Setting the mean of feature i to μi=0, the variance of feature i to σii=1, and the covariance between features i and j to σij=ρx|ij|, for all i and j in {1,,p}, we simulate multiple feature matrices from the multivariate Gaussian distribution with mean vector μ and covariance matrix Σ, namely the n0×p feature matrix X0 for the target dataset, and the n1/2/3×p feature matrices {X1,X2,X3} for the source datasets (n0=100,n1=n2=n3=150,p=500).

  • Coefficients: Setting the mean and the variance for dataset k to μk=0 and σkk=1, for all k in {0,1,2,3}, and the covariance between datasets k and l to σkl=0 if either k or l equals 1, or to σkl=ρβ if both k and l are in {0, 2, 3}, we simulate two p×4 matrices from the multivariate Gaussian distribution with mean vector μ and covariance matrix Σ, namely B1 and B2. We define the coefficients as B=B1I[B2>ϕ1(1π)], where ϕ is the Gaussian cumulative distribution function and π equals 0.2 (dense) or 0.05 (sparse), and denote the p-dimensional coefficient vectors by β0 for the target dataset and by {β1,β2,β3} for the source datasets. While one set of coefficients is non-transferable (β1), we transform the transferable sets of coefficients with β2sign(β2)|β2|2 and β3sign(β3)|β3|.

  • Targets: For the target dataset, we compute z0=X0β0 and standardize z0 to obtain z0*. For the source datasets, we proceed similarly to obtain {z1*,z2*,z3*}. The simulated targets equal yk=h1(wzk*+1wϵk), where h(·) is a link function and ϵk follows a standard Gaussian distribution, for k in {0,,3}. Given 0w1, we have Var(wzk*+1wϵk)=wVar(zk*)+(1w)Var(ϵk)=1. While h(·) is the identity link in the Gaussian case, it is the logit link in the binomial case, where the simulated probabilities are rounded to simulated classes. We set the signal-to-noise ratio to 4:1 (w = 0.8).

3.3 Simulation results

In addition to the target dataset, the method from Tian and Feng (2022, glmtrans) requires the source datasets, while our method (transreg) and the method from Kawaguchi et al. (2022, xrnet) require the prior effects derived from the source datasets. Since these methods have different requirements, we first simulate the source datasets (for glmtrans) and then derive the prior effects from the simulated source datasets (for xrnet and transreg). As prior effects, we use the estimated coefficients from penalized regression on the source datasets. We choose the type of regularization for all methods subject to the simulation setting, namely ridge regularization for dense settings (glmtrans: α = 0; xrnet and transreg: αsource=αtarget=0) and lasso or lasso-like elastic net regularization for sparse settings (glmtrans: α = 1; xrnet and transreg: αsource=0.95,αtarget=1). The idea of the lasso-like elastic net regularization is to render the prior information more stable.

In each simulation setting, we simulate 100 training samples and 10000 testing samples (hold-out) for the target dataset. Tables 3 and 4 show the predictive performance for the testing data in the external and internal simulation, under exponential and isotonic calibration. We observe that transfer learning (with glmtrans, xrnet, or transreg) often leads to a significant improvement with respect to standard penalized regression (with glmnet). Concerning the proposed method, this holds for the two different calibration approaches and the two different stacking approaches. These simulation studies do not show that the proposed transfer learning method outperforms other transfer learning methods. The advantage of the proposed method as compared to the method from Tian and Feng (2022, glmtrans) is that it does not require the source data but only the prior effects derived from the source data, and the advantage as compared to the method from Kawaguchi et al. (2022, xrnet) is that it allows for non-linear relationships between the prior effects and the true effects. While the method from Kawaguchi et al. (2022, xrnet) shrinks the coefficients towards a linear function of the prior effects, the proposed method adapts the prior effects to the true effects through exponential or isotonic calibration.

Table 3.

Predictive performance in external simulation.a

transreg
K a h α Family ρ¯x max(ρ^β) glmnet glmtrans xrnet exp.sta exp.sim iso.sta iso.sim
1 5 0 Gaussian 0.01 1.00 73.2 ± 3.0 43.9 ± 9.2* 47.1 ± 1.8* 32.2 ± 3.2* 30.9 ± 2.8* 24.5 ± 3.6* 23.4 ± 3.3*
3 5 0 Gaussian 0.01 1.00 73.2 ± 3.0 33.5 ± 6.5* 29.3 ± 2.4* 18.1 ± 2.5* 16.7 ± 2.0* 13.5 ± 2.0* 12.7 ± 1.6*
5 5 0 Gaussian 0.01 1.00 73.2 ± 3.0 24.0 ± 3.7* 22.6 ± 1.5* 14.2 ± 1.7* 13.2 ± 1.7* 10.6 ± 1.0* 10.0 ± 0.9*
1 250 0 Gaussian 0.01 0.40 73.2 ± 3.0 31.8 ± 9.4* 63.8 ± 3.6* 54.8 ± 6.0* 57.8 ± 6.5* 49.5 ± 6.9* 50.0 ± 7.3*
3 250 0 Gaussian 0.01 0.42 73.2 ± 3.0 33.3 ± 8.2* 51.4 ± 4.7* 46.5 ± 7.9* 47.0 ± 9.2* 39.4 ± 6.5* 37.6 ± 7.5*
5 250 0 Gaussian 0.01 0.43 73.2 ± 3.0 33.3 ± 5.0* 43.4 ± 3.4* 40.0 ± 6.3* 39.2 ± 7.3* 32.7 ± 4.4* 30.0 ± 4.7*
1 5 1 Gaussian 0.01 1.00 17.3 ± 3.8 12.5 ± 1.4* 14.6 ± 1.7 12.2 ± 1.9* 14.7 ± 1.9* 11.0 ± 1.6* 11.6 ± 1.6*
3 5 1 Gaussian 0.01 1.00 17.3 ± 3.8 10.5 ± 0.7* 11.5 ± 0.6* 10.8 ± 0.6* 11.7 ± 1.6* 9.8 ± 0.5* 9.9 ± 0.6*
5 5 1 Gaussian 0.01 1.00 17.3 ± 3.8 10.0 ± 0.4* 11.0 ± 0.6* 10.5 ± 0.5* 11.1 ± 1.3* 9.6 ± 0.3* 9.8 ± 0.4*
1 250 1 Gaussian 0.01 0.24 17.3 ± 3.8 22.2 ± 13.4 17.6 ± 4.1 14.7 ± 3.4* 17.8 ± 4.8 14.7 ± 3.4* 18.2 ± 5.0
3 250 1 Gaussian 0.01 0.27 17.3 ± 3.8 19.8 ± 6.7 17.7 ± 3.6 14.7 ± 3.4* 18.4 ± 4.8 14.8 ± 3.4* 18.7 ± 5.5
5 250 1 Gaussian 0.01 0.27 17.3 ± 3.8 19.8 ± 6.7 17.5 ± 3.6 14.8 ± 3.3* 19.0 ± 4.6 14.7 ± 3.5* 21.3 ± 6.6
1 5 0 Binomial 0.01 1.00 91.5 ± 1.5 91.0 ± 4.2 84.5 ± 1.0* 82.4 ± 5.6* 82.7 ± 5.4* 77.8 ± 3.7* 78.5 ± 4.7*
3 5 0 Binomial 0.01 1.00 91.5 ± 1.5 85.9 ± 3.7* 75.9 ± 2.9* 69.8 ± 3.6* 70.0 ± 3.8* 66.7 ± 3.7* 66.6 ± 4.5*
5 5 0 Binomial 0.01 1.00 91.5 ± 1.5 79.9 ± 2.4* 70.8 ± 2.9* 63.1 ± 3.8* 63.2 ± 3.9* 61.2 ± 3.0* 62.2 ± 5.7*
1 250 0 Binomial 0.01 0.41 91.5 ± 1.5 92.0 ± 5.0 91.1 ± 2.0 89.4 ± 2.9* 90.2 ± 3.0 88.3 ± 3.2* 88.7 ± 3.0*
3 250 0 Binomial 0.01 0.42 91.5 ± 1.5 89.7 ± 4.3 86.5 ± 4.2* 86.3 ± 4.6* 87.8 ± 4.5* 84.0 ± 7.5* 84.2 ± 5.6*
5 250 0 Binomial 0.01 0.43 91.5 ± 1.5 85.3 ± 3.5* 83.0 ± 3.2* 84.3 ± 4.8* 85.9 ± 5.5* 79.4 ± 4.4* 79.0 ± 5.0*
1 5 1 Binomial 0.01 1.00 80.4 ± 8.2 77.8 ± 12.2 77.1 ± 8.4 71.4 ± 5.8* 73.0 ± 4.0* 70.3 ± 5.2* 72.4 ± 5.0*
3 5 1 Binomial 0.01 1.00 80.4 ± 8.2 70.5 ± 12.5* 69.5 ± 11.8* 67.9 ± 5.7* 70.8 ± 10.3* 64.7 ± 4.8* 64.4 ± 5.1*
5 5 1 Binomial 0.01 1.00 80.4 ± 8.2 59.5 ± 2.5* 64.4 ± 4.2* 65.7 ± 6.4* 65.5 ± 9.6* 62.7 ± 4.5* 61.8 ± 4.2*
1 250 1 Binomial 0.01 0.25 80.4 ± 8.2 82.1 ± 8.7 82.4 ± 9.9 81.5 ± 9.2 81.5 ± 9.1 81.2 ± 7.2 82.4 ± 10.1
3 250 1 Binomial 0.01 0.26 80.4 ± 8.2 80.5 ± 9.3 82.0 ± 8.8 80.7 ± 8.8 81.3 ± 9.5 79.2 ± 8.3 82.7 ± 11.2
5 250 1 Binomial 0.01 0.27 80.4 ± 8.2 78.9 ± 9.4* 84.6 ± 15.5 80.3 ± 8.3 83.4 ± 8.8 79.4 ± 10.7 86.3 ± 16.8
a

In each setting (row), we simulate 10 datasets, calculate the performance metric (mean-squared error for numerical prediction, logistic deviance for binary classification) for the test sets, express these metrics as percentages of those from prediction by the mean, and show the mean and standard deviation of these percentages. Settings: number of transferable source datasets (Ka), differences between source and target coefficients (h), dense setting with ridge regularization (s =50, α = 0) or sparse setting with lasso regularization (s =15, α = 1), family of distribution (‘gaussian’ or ‘binomial’). These parameters determine (i) the mean Pearson correlation among the features in the target dataset (ρ¯x) and (ii) the maximum Pearson correlation between the coefficients in the target dataset and the coefficients in the source datasets (max(ρ^β)). Methods: regularized regression (glmnet), competing transfer learning methods (glmtrans, xrnet), proposed transfer learning method (transreg) with exponential/isotonic calibration and standard/simultaneous stacking. In each setting, the colour black (grey) highlights methods that are more (less) predictive than regularized regression without transfer learning (glmnet), asterisks (daggers) indicate methods that are significantly more (less) predictive at the 5% level (one-sided Wilcoxon signed-rank test), and an underline highlights the most predictive method.

Table 4.

Predictive performance in internal simulation.a

transreg
ρx ρβ α Family ρ¯x max(ρ^β) glmnet glmtrans xrnet exp.sta exp.sim iso.sta iso.sim
0.95 0.60 0 Gaussian 0.08 0.32 46.1 ± 7.3 48.5 ± 6.3 45.2 ± 7.0* 45.2 ± 7.6* 45.4 ± 7.1 45.1 ± 7.5* 44.5 ± 6.8*
0.99 0.60 0 Gaussian 0.32 0.32 33.2 ± 4.8 33.2 ± 6.4 32.6 ± 4.4* 32.2 ± 4.6* 32.5 ± 4.4* 32.1 ± 4.7* 32.0 ± 4.1*
0.95 0.80 0 Gaussian 0.08 0.54 45.1 ± 6.7 47.9 ± 11.1 42.0 ± 6.4* 43.7 ± 7.1* 43.3 ± 6.9* 43.8 ± 7.1* 41.9 ± 6.5*
0.99 0.80 0 Gaussian 0.32 0.54 32.5 ± 4.8 31.8 ± 4.2 31.1 ± 4.7* 32.0 ± 4.8* 31.6 ± 4.7* 31.6 ± 4.7* 30.4 ± 3.9*
0.95 0.99 0 Gaussian 0.08 0.89 44.1 ± 5.1 42.2 ± 5.6 37.3 ± 4.4* 40.9 ± 5.3* 37.6 ± 4.5* 40.1 ± 5.7* 38.8 ± 5.7*
0.99 0.99 0 Gaussian 0.32 0.89 31.0 ± 5.1 29.4 ± 2.1 27.6 ± 3.0* 30.4 ± 5.3 28.4 ± 3.8* 29.8 ± 4.8* 29.1 ± 3.9*
0.95 0.60 1 Gaussian 0.08 0.27 37.7 ± 6.8 39.1 ± 7.2 37.9 ± 6.6 37.3 ± 7.2 38.0 ± 7.3 37.4 ± 7.0 41.5 ± 6.4
0.99 0.60 1 Gaussian 0.32 0.27 28.5 ± 2.4 29.9 ± 3.5 29.1 ± 3.4 28.8 ± 2.5 29.0 ± 2.7 28.7 ± 2.7 28.7 ± 2.5
0.95 0.80 1 Gaussian 0.08 0.45 37.5 ± 4.8 37.4 ± 5.0 36.9 ± 5.2 35.7 ± 4.2* 36.8 ± 4.7 35.9 ± 4.7* 39.5 ± 5.4
0.99 0.80 1 Gaussian 0.32 0.45 29.9 ± 2.4 29.9 ± 2.8 29.5 ± 3.4 29.1 ± 2.5 29.4 ± 3.3 29.1 ± 2.8 29.6 ± 3.8
0.95 0.99 1 Gaussian 0.08 0.87 38.0 ± 6.4 33.0 ± 5.7* 34.5 ± 7.9* 34.8 ± 8.2* 34.1 ± 6.8* 34.0 ± 7.6* 35.2 ± 8.0
0.99 0.99 1 Gaussian 0.32 0.87 30.2 ± 4.4 29.5 ± 4.5 29.2 ± 4.6 29.4 ± 4.8 29.1 ± 4.2* 28.6 ± 3.5* 29.4 ± 4.4
0.95 0.60 0 Binomial 0.08 0.32 77.1 ± 4.5 81.2 ± 6.1 76.7 ± 4.4 76.8 ± 4.8 76.1 ± 5.0 77.7 ± 4.8 77.7 ± 5.2
0.99 0.60 0 Binomial 0.32 0.32 65.5 ± 4.7 67.0 ± 5.9 65.0 ± 4.7 63.7 ± 4.8* 65.3 ± 4.4 63.8 ± 4.8* 64.0 ± 4.9
0.95 0.80 0 Binomial 0.08 0.54 74.9 ± 4.4 81.2 ± 10.7 73.5 ± 5.1* 75.0 ± 5.7 73.9 ± 5.6 74.7 ± 4.7 73.5 ± 5.4*
0.99 0.80 0 Binomial 0.32 0.54 64.3 ± 3.6 64.7 ± 4.6 61.8 ± 4.7* 63.2 ± 4.5* 63.7 ± 5.0 63.0 ± 4.7* 63.6 ± 5.1
0.95 0.99 0 Binomial 0.08 0.89 75.4 ± 5.0 74.8 ± 4.7* 69.5 ± 4.2* 72.4 ± 6.5* 71.5 ± 6.3* 71.7 ± 4.9* 71.4 ± 4.8*
0.99 0.99 0 Binomial 0.32 0.89 62.4 ± 5.0 61.8 ± 5.0 58.1 ± 4.6* 61.0 ± 6.7 58.9 ± 6.7* 60.4 ± 6.2* 59.6 ± 5.0*
0.95 0.60 1 Binomial 0.08 0.27 76.5 ± 6.0 75.8 ± 3.9 75.7 ± 4.1 75.8 ± 5.8 76.5 ± 4.9 75.9 ± 5.4 75.9 ± 2.3
0.99 0.60 1 Binomial 0.32 0.27 61.2 ± 4.4 61.5 ± 4.9 63.4 ± 6.8 63.1 ± 5.6 62.2 ± 5.7 62.7 ± 5.4 61.5 ± 4.6
0.95 0.80 1 Binomial 0.08 0.45 78.2 ± 10.5 76.6 ± 10.2 76.3 ± 9.8 75.0 ± 6.0 77.1 ± 9.0 75.2 ± 6.1 77.9 ± 6.6
0.99 0.80 1 Binomial 0.32 0.45 64.9 ± 4.9 64.8 ± 5.6 65.5 ± 2.9 66.3 ± 6.7 64.4 ± 5.2 65.2 ± 6.0 65.5 ± 4.3
0.95 0.99 1 Binomial 0.08 0.87 80.1 ± 6.0 73.3 ± 5.7* 70.7 ± 6.3* 69.2 ± 4.5* 68.8 ± 6.8* 70.1 ± 5.2* 69.3 ± 5.3*
0.99 0.99 1 Binomial 0.32 0.87 63.2 ± 5.1 62.9 ± 4.7 61.2 ± 5.6* 62.7 ± 6.1 61.1 ± 5.7* 61.8 ± 5.8 61.2 ± 6.2*
a

In each setting (row), we simulate 10 datasets, calculate the performance metric (mean-squared error for numerical prediction, logistic deviance for binary classification) for the test sets, express these metrics as percentages of those from prediction by the mean, and show the mean and standard deviation of these percentages. Settings: correlation parameter for features (ρx), correlation parameter for coefficients (ρβ), dense setting with ridge regularization (π=30%, α = 0) or sparse setting with lasso regularization (π=5%, α = 1), family of distribution (‘gaussian’ or ‘binomial’). These parameters determine (i) the mean Pearson correlation among the features in the target dataset (ρ¯x) and (ii) the maximum Pearson correlation between the coefficients in the target dataset and the coefficients in the source datasets (max(ρ^β)). Methods: regularized regression (glmnet), competing transfer learning methods (glmtrans, xrnet), proposed transfer learning method (transreg) with exponential/isotonic calibration and standard/simultaneous stacking. In each setting, the colour black (grey) highlights methods that are more (less) predictive than regularized regression without transfer learning (glmnet), asterisks (daggers) indicate methods that are significantly more (less) predictive at the 5% level (one-sided Wilcoxon signed-rank test), and an underline highlights the most predictive method.

4 Applications

4.1 External applications

First, we consider an adapted version of the application on cervical cancer from van de Wiel et al. (2016). The aim is to transfer information from a methylation study with biopsy samples to another methylation study with self-collected samples in order to better discriminate between low-grade and high-grade precursor lesions. Specifically, we transfer the signs of the effect sizes and the P-values from the source dataset to the target dataset (n =44 samples, p=9491 features). We then examine whether this prior information increases the predictive performance of ridge regression, which is more predictive than lasso regression in this application. Next to our transfer learning method (transreg) and the one from Kawaguchi et al. (2022, xrnet), we consider the co-data learning methods from Tay et al. (2023, fwelnet) and van Nee et al. (2021, ecpc). While these transfer learning methods exploit information on the importance and direction of the effects (co-data: sign(coef)log10(P-value)), these co-data learning methods only exploit information on their importance (co-data: log10(P-value)). After 10 repetitions of 10-fold cross-validation, we observe that the proposed method (not with exponential but with isotonic calibration) often increases the predictive performance of ridge regression (transreg.exp.sta: 0/10, transreg.exp.sim: 4/10, transreg.iso.sta: 7/10, transreg.iso.sim: 10/10, fwelnet: 7/10, ecpc: 5/10, xrnet: 3/10). We also observe that exploiting information on the importance as well as the direction of the effects (transreg) can be more beneficial than exploiting information on the importance of the effects only (fwelnet, ecpc), as can be seen in the mean change in cross-validated logistic deviance (transreg.exp.sta: +6.56%, transreg.exp.sim: +0.43%, transreg.iso.sta: −2.50%, transreg.iso.sim: −9.25%, fwelnet: −0.12%, ecpc: −1.52%, xrnet: +2.59%). The relatively poor performance of the competing transfer learning method (xrnet) might be caused by a non-linear relationship between the prior effects (transformed P-values) and the true effects. Here, isotonic calibration outperforms exponential calibration, and simultaneous stacking outperforms standard stacking. A potential explanation for the large difference in performance between exponential and isotonic calibration is that positive effects might be more important than negative effects in this application, for a biological reason (methylation increases the probability of cancer) and a statistical reason (effects of overexpression are easier to detect than those of underexpression). While exponential calibration behaves symmetrically for negative and positive prior effects, isotonic calibration can shrink negative prior effects towards zero.

Second, we consider an adapted version of the application on pre-eclampsia from Tay et al. (2023). Measurements of p=1125 plasma proteins are available for n =166 patients at multiple time points (48×2+8×3+20×4+74×5+16×6=666). The aim is to transfer information from late time points (gestational age > 20 weeks) to early time points (gestational age 20 weeks). We repeatedly split the patients into one source dataset and one target dataset. Patients with only late time points are always in the source dataset, and other patients are randomly allocated to the source and the target dataset. (Note that this application is somewhat artificial, as it might be better to drop transfer learning in favour of using all earliest time points in the regression of interest.) Using the source dataset, we estimate two logistic regression models under ridge regularization, once using the early time points and once using all time points. For each patient, all time points are assigned to the same cross-validation fold, and the weight is split evenly among the time points. We then use the two sets of estimated regression coefficients as co-data for the target dataset. In the regression for the target dataset, we only include the earliest time point of each patient. Using 10-fold cross-validation, we estimate the predictive performance of ridge regression with and without transfer learning. After repeating source-target splitting and cross-validation 10 times, we observe that transfer learning tends to decrease the cross-validated logistic deviance (transreg.exp.sta: 8/10, transreg.exp.sim: 7/10, transreg.iso.sta: 7/10, transreg.iso.sim: 9/10, fwelnet: 5/10, ecpc: 6/10, xrnet: 10/10). It is more beneficial to share information not only on the importance but also the direction of the effects, according to the mean change in cross-validated logistic deviance (transreg.exp.sta: −2.61%, transreg.exp.sim: −4.29%, transreg.iso.sta: −3.67%, transreg.iso.sim: −8.33%, fwelnet: +0.04%, ecpc: −6.87%, xrnet: −12.12%). Simultaneous stacking again outperforms standard stacking, but exponential and isotonic calibration show a similar performance. In this application, where the prior effects are probably close to the true effects, the competing transfer learning method (xrnet) outperforms the proposed one.

For both applications, Fig. 2 shows the change in predictive performance from modelling without to modelling with prior information.

Figure 2.

Figure 2.

Change in predictive performance in external applications due to prior information. Percentage change in cross-validated logistic deviance (y-axis) from ridge regression (with glmnet) to other methods (x-axis), for 10 repetitions of 10-fold cross-validation. A negative value (blue) indicates an improvement, and a positive value (red) indicates a deterioration, based on the metric ’logistic deviance’. Top: application on cervical cancer. Bottom: application on pre-eclampsia. From left to right: co-data learning methods (fwelnet, ecpc), competing transfer learning method (xrnet), proposed transfer learning method (transreg) with exponential/isotonic calibration and standard/simultaneous stacking.

4.2 Internal application

In this application, we transfer information from a meta-analysis of genome-wide association studies on Parkinson’s disease (PD-GWAS, Nalls et al. 2019) to the Luxembourg Parkinson’s study (LuxPARK, Hipp et al. 2018). The aim is to classify samples into Parkinson’s disease (PD) patients and healthy controls based on SNPs.

At the time of our study, the LuxPARK dataset included genotyping and clinical data of 790 PD cases and 766 healthy controls. DNA samples were genotyped using the NeuroChip array (Blauwendraat et al. 2017). Quality control steps of genotyping data were conducted according to the standard procedures reported previously (Pavelka et al. 2022). Missing genotyping data were imputed using the reference panel from the Haplotype Reference Consortium (release 1.1) on the Michigan Imputation Server (Das et al. 2016) (RRID: ID_017579), with a filter for imputation quality (r2 > 0.3).

As common SNPs exhibit weak effects on PD, the sample size is likely insufficient to train a highly predictive model. However, publicly available summary statistics from the largest-to-date PD-GWAS (with around 38000 cases and 1400000 controls from European ancestry) (Nalls et al. 2019) might serve as prior information on the SNP effects. For each SNP, these summary statistics are the combined results from simple logistic regression of the PD status on the SNP, namely the estimated slope (logarithmic odds ratio), its standard error, and the associated P-value. Importantly, the LuxPARK cohort was not part of the PD-GWAS, meaning that the prior information comes from independent data. As the LuxPARK cohort and the PD-GWAS cohorts have a similar ethnic background, the prior information might allow us to increase the predictive performance.

The two lists of SNPs—from the LuxPARK genotyping data (target dataset) and the PD-GWAS summary statistics (source dataset)—are partially overlapping. SNP data are high-dimensional and strongly correlated. From each block of SNPs in the target dataset (250 kb window), we retain the most significant one and those that are in weak pairwise linkage disequilibrium with it (r2<0.1). Next, we only retain the SNPs appearing also in the source dataset. These two filtering steps together reduce the dimensionality in the target dataset from around 18 million SNPs to 196018 SNPs. We code the SNP data for dominant effects, with 0 meaning no alternate allele (0/0) and 1 meaning one or two alternate alleles (0/1 or 1/1).

It seems that the results from the source dataset are informative, because 5.80% of the P-values are nominally significant at the 0.05 level (11377 out of 196018), 77 are significant at a false discovery rate of 5% (Benjamini–Hochberg), and 35 are significant at a family-wise error rate of 5% (Holm–Bonferroni). As SNPs with a low minor allele frequency might have large effect sizes but insignificant P-values, we base the prior effects not on the estimated coefficients (β^) but on the signed logarithmic P-values (sign(β^)log10(P)). For each SNP, we compared the reference and the alternate alleles between the two datasets: (i) if both datasets have the same reference allele and the same alternate allele, the signed logarithmic P-value from the source dataset becomes the prior effect for the target dataset; (ii) if the reference allele of each dataset is the alternate allele of the other dataset (swapped alleles), we invert the sign of the signed logarithmic P-value; and (iii) if the two datasets have two different sets of alleles (multiallelic SNP), we set the prior effect to zero.

Rather than using the 196018 SNPs for predictive modelling in the target dataset, we also filter them based on their significance in the source dataset (which is already a type of transfer learning). For each cut-off in {5×102,5×103,,5×1010}, we exclude all SNPs above and include all SNPs below. This means that for the target dataset, we retain a specific number of the most significant SNPs from the source dataset. For each significance cut-off, we compare three modelling approaches:

  • Uninformed approach: We use logistic regression with ridge or lasso penalization to model the PD status based on the included SNPs. All included SNPs are treated equally, irrespective of their estimated effect in the source dataset.

  • Naïve transfer learning: After calculating for each sample the sum across the signed logarithmic P-values from the source dataset multiplied by the SNPs from the target dataset, we fit a simple logistic regression of the PD status on this sum.

  • Transfer learning: The proposed transfer learning approach uses the signed logarithmic P-values from the source dataset as prior effects for the target dataset.

Figure 3 shows the predictive performance of modelling with estimated effects (uninformed approach), with prior effects (naïve transfer learning), or with both (transfer learning). We obtained the results with repeated nested cross-validation (10 repetitions, 10 external folds, 10 internal folds), using the same folds for all methods. If the significance cut-off is very strict, leading to a small number of significant SNPs, transfer learning does not improve the predictive performance of ridge and lasso regression. In these low-dimensional settings with many fewer SNPs than samples, prior information on the SNPs is not helpful. But otherwise transfer learning does improve the predictive performance of ridge and lasso regression. This holds for all four flavours of the proposed transfer learning method (exponential versus isotonic calibration, standard versus simultaneous stacking), but isotonic calibration works considerably better than exponential calibration and simultaneous stacking works marginally better than standard stacking.

Figure 3.

Figure 3.

Predictive performance in internal application. Mean cross-validated area under the ROC curve (AUC) from 10 times 10-fold cross-validation (y-axis) against P-value cutoff (x-axis) for regression without (solid line) and with (dashed lines) transfer learning (bright blue: standard stacking, dark blue: simultaneous stacking), under either ridge (left) or lasso (right) regularization and either exponential (top) or isotonic (bottom) calibration. The numbers within brackets indicate the dimensionality, and the dotted line is for naïve transfer learning. While AUC=0.5 represents random classification, AUC=1.0 represents perfect classification. Given 766 controls and 790 cases, a random classifier achieves an AUC above 0.524 in 5% of the cases.

Depending on the significance cut-off determining the number of significant SNPs, the performance of naïve transfer learning can be as high as the one of transfer learning with isotonic calibration. In these cases, the prior effects are predictive to the extent that it is not even necessary to estimate any effects. An explanation for the high performance of naïve transfer learning might be (i) the large sample size in the source dataset for testing the marginal effects of the SNPs together with (ii) the linkage disequilibrium clumping leading to a selection of relatively independent SNPs.

5 Discussion

We proposed a two-step transfer learning method for exploiting estimated coefficients from related studies to improve the predictive performance in the study of interest. First, we adapt the prior effects from the source datasets to the target dataset, either with exponential or isotonic calibration. While exponential calibration is more robust to outliers (only three free parameters), isotonic calibration is more flexible (only maintains order of prior effects). We expect the former to be superior if the prior effects are close to the true effects, and the latter to be superior if there is no exponential relationship. Second, we combine the calibrated prior effects with information from the observed data, based on two variants of stacked generalization. While the first variant (standard stacking) is more suitable if there are many sources of co-data (‘averaging calibrated prior effects and estimated effects’), the second variant (simultaneous stacking) is more suitable if there is one source of co-data with partially unreliable or partially missing prior effects (‘shrinking combined effects towards calibrated prior effects’).

The proposed transfer learning method allows for multiple sources of prior information. It does not require the source dataset(s) but only the prior effects derived from the source dataset(s). The proposed sequential transfer learning method has a competitive predictive performance with the less flexible parallel transfer learning method from Tian and Feng (2022, glmtrans). In the case of closely related tasks, accounting for prior effects with transfer learning seems to be more beneficial than accounting for prior weights with the co-data methods from Tay et al. (2023, fwelnet) and van Nee et al. (2021, ecpc). And in the case of non-linearly related prior effects, the proposed method seems to outperform the one from Kawaguchi et al. (2022, xrnet). We therefore believe that the proposed method could tackle many biomedical predictions problems with one or more sets of prior effects.

In some applications, only one type of prior information derived from the source datasets is available. In other applications, multiple types of prior information are available (or the source datasets themselves). Then we can choose from multiple types of prior information. If the source and target datasets have the same feature space, estimated coefficients from penalized regression might be a reasonable choice. If the feature spaces are different, however, it is problematic that (i) lasso regression erratically selects among correlated features and (ii) ridge regression distributes weight among correlated features. This means that the presence or absence of additional correlated features in the source datasets might change the prior information on the features of interest. The same problem arises under contamination of a subset of features (van de Wiel et al. 2016). We therefore expect that signed logarithmic P-values from pairwise testing (sign(β^)log10(P)) will often be more informative than estimated coefficients from multiple regression (β^).

Acknowledgements

We are grateful to Quentin Klopfenstein for helpful discussions on differential penalization, to the responsible and reproducible research (R3) initiative for the pre-publication check, and to the anonymous reviewers for constructive criticism. Data used in the preparation of this manuscript were obtained from NCER-PD. All participants provided written informed consent, and the study has been approved by the National Ethics Board (CNER Ref: 201407/13). We would like to thank all participants of the Luxembourg Parkinson’s study for their important support of our research. Furthermore, we acknowledge the joint effort of the NCER-PD consortium members from the partner institutions Luxembourg Centre for Systems Biomedicine, Luxembourg Institute of Health, Centre Hospitalier de Luxembourg, and Laboratoire National de Santé generally contributing to the Luxembourg Parkinson’s Study as listed below: Alexander Hundt2, Alexandre Bisdorff5, Amir Sharify2, Anne Grünewald1, Anne-Marie Hanff2, Armin Rauschenberger1, Beatrice Nicolaï3, Brit Mollenhauer12, Camille Bellora2, Carlos Vega Moreno1, Chouaib Mediouni2, Christophe Trefois1, Claire Pauly1,3, Clare Mackay10, Clarissa Gomes1, Daniela Berg11, Daniela Esteves2, Deborah McIntyre2, Dheeraj Reddy Bobbili1, Eduardo Rosales Jubal2, Ekaterina Soboleva1, Elisa Gómez de Lope1, Elodie Thiry3, Enrico Glaab1, Estelle Henry2, Estelle Sandt2, Evi Wollscheid-Lengeling1, Françoise Meisch1, Friedrich Mühlschlegel4, Gaël Hammot2, Geeta Acharya2, Gelani Zelimkhanov3, Gessica Contesotto2, Giuseppe Arena1, Gloria Aguayo2, Guilherme Marques2, Guy Berchem3, Guy Fagherazzi2, Hermann Thien2, Ibrahim Boussaad1, Inga Liepelt11, Isabel Rosety1, Jacek Jaroslaw Lebioda1, Jean-Edouard Schweitzer1, Jean-Paul Nicolay19, Jean-Yves Ferrand2, Jens Schwamborn1, Jérôme Graas2, Jessica Calmes2, Jochen Klucken1,2,3, Johanna Trouet2, Kate Sokolowska2, Kathrin Brockmann11, Katrin Marcus13, Katy Beaumont2, Kirsten Rump1, Laura Longhino3, Laure Pauly1, Liliana Vilas Boas3, Linda Hansen1,3, Lorieza Castillo2, Lukas Pavelka1,3, Magali Perquin2, Maharshi Vyas1, Manon Gantenbein2, Marek Ostaszewski1, Margaux Schmitt2, Mariella Graziano17, Marijus Giraitis2,3, Maura Minelli2, Maxime Hansen1,3, Mesele Valenti2, Michael Heneka1, Michael Heymann2, Michel Mittelbronn1,4, Michel Vaillant2, Michele Bassis1, Michele Hu8, Muhammad Ali1, Myriam Alexandre2, Myriam Menster2, Nadine Jacoby18, Nico Diederich3, Olena Tsurkalenko2, Olivier Terwindt1,3, Patricia Martins Conde1, Patrick May1, Paul Wilmes1, Paula Cristina Lupu2, Pauline Lambert2, Piotr Gawron1, Quentin Klopfenstein1, Rajesh Rawal1, Rebecca Ting Jiin Loo1, Regina Becker1, Reinhard Schneider1, Rejko Krüger1,2,3, Rene Dondelinger5, Richard Wade-Martins9, Robert Liszka14, Romain Nati3, Rosalina Ramos Lima2, Roseline Lentz7, Rudi Balling1, Sabine Schmitz1, Sarah Nickels1, Sascha Herzinger1, Sinthuja Pachchek1, Soumyabrata Ghosh1, Stefano Sapienza1, Sylvia Herbrink6, Tainá Marques1, Thomas Gasser11, Ulf Nehrbass2, Valentin Grouès1, Venkata Satagopam1, Victoria Lorentz2, Walter Maetzler15, Wei Gu1, Wim Ammerlann2, Yohan Jaroz1, Zied Landoulsi1. 1Luxembourg Centre for Systems Biomedicine, University of Luxembourg, Esch-sur-Alzette, Luxembourg, 2Luxembourg Institute of Health, Strassen, Luxembourg, 3Centre Hospitalier de Luxembourg, Strassen, Luxembourg, 4Laboratoire National de Santé, Dudelange, Luxembourg, 5Centre Hospitalier Emile Mayrisch, Esch-sur-Alzette, Luxembourg, 6Centre Hospitalier du Nord, Ettelbrück, Luxembourg, 7Parkinson Luxembourg Association, Leudelange, Luxembourg, 8Oxford Parkinson’s Disease Centre, Nuffield Department of Clinical Neurosciences, University of Oxford, Oxford, UK, 9Oxford Parkinson’s Disease Centre, Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, UK, 10Oxford Centre for Human Brain Activity, Wellcome Centre for Integrative Neuroimaging, Department of Psychiatry, University of Oxford, Oxford, UK, 11Center of Neurology and Hertie Institute for Clinical Brain Research, Department of Neurodegenerative Diseases, University Hospital Tübingen, Tübingen, Germany, 12Paracelsus-Elena-Klinik, Kassel, Germany, 13Ruhr-University of Bochum, Bochum, Germany, 14Westpfalz-Klinikum GmbH, Kaiserslautern, Germany, 15Department of Neurology, University Medical Center Schleswig-Holstein, Kiel, Germany, 16Department of Neurology Philipps, University Marburg, Marburg, Germany, 17Association of Physiotherapists in Parkinson’s Disease Europe, Esch-sur-Alzette, Luxembourg, 18Private practice, Ettelbruck, Luxembourg, 19Private practice, Luxembourg, Luxembourg.

Contributor Information

Armin Rauschenberger, Luxembourg Centre for Systems Biomedicine (LCSB), University of Luxembourg, 4362 Esch-sur-Alzette, Luxembourg.

Zied Landoulsi, Luxembourg Centre for Systems Biomedicine (LCSB), University of Luxembourg, 4362 Esch-sur-Alzette, Luxembourg.

Mark A van de Wiel, Department of Epidemiology and Data Science (EDS), Amsterdam University Medical Centers (Amsterdam UMC), 1081 HV Amsterdam, The Netherlands.

Enrico Glaab, Luxembourg Centre for Systems Biomedicine (LCSB), University of Luxembourg, 4362 Esch-sur-Alzette, Luxembourg.

Author contributions

E.G. acquired funding. A.R. and M.A.v.d.W. developed the method. A.R. analysed the data and drafted the manuscript. Z.L. processed the SNP data and critically revised the internal application. M.A.v.d.W. and E.G. critically revised the manuscript. All authors read and approved the final manuscript.

Conflict of interest

None declared.

Funding

This work was supported by the Luxembourg National Research Fund (FNR) for the project ‘Clinnova—a trans-regional digital health effort’ [NCER/23/16695277]; and the project DIGIPD [INTER/ERAPerMed 20/14599012] as part of the European Union’s Horizon 2020 Research and Innovation Program. The National Centre of Excellence in Research on Parkinson’s Disease (NCER-PD) is funded by the FNR [FNR/NCER13/BM/11264123].

Data availability

The R package transreg is available on GitHub (https://github.com/lcsb-bds/transreg) and CRAN (https://cran.r-project.org/package=transreg), with the code for the simulations and the applications in a vignette (https://lcsb-bds.github.io/transreg/). We obtained our results using R 4.3.0 (RRID: ID_001905) on a physical machine (aarch64-apple-darwin20, macOS Ventura 13.5.2). Data for the application on cervical cancer are available from van de Wiel et al. (2016), in the R package GRridge in the dataset ‘dataVerlaat’ (source data: Farkas et al. 2013, target data: van de Wiel et al. 2016). Data for the application on pre-eclampsia are available from Erez et al. (2017), in the supporting file ‘pone.0181468.s001.csv’. For the application on Parkinson’s disease, the source data are available from Nalls et al. (2019), in the online file ‘nallsEtAl2019_excluding23andMe_allVariants.tab’, and the target data are available upon request (request.ncer-pd@uni.lu). Information on reproducibility is also available on a frozen page (doi: 10.17881/hczj-3297).

References

  1. Bergersen LC, Glad IK, Lyng H.. Weighted lasso with data integration. Stat Appl Genet Mol Biol 2011;10:39. 10.2202/1544-6115.1703. [DOI] [PubMed] [Google Scholar]
  2. Blauwendraat C, Faghri F, Pihlstrom L. et al. ; International Parkinson’s Disease Genomics Consortium (IPDGC), COURAGE-PD Consortium. NeuroChip, an updated version of the NeuroX genotyping platform to rapidly screen for variants associated with neurological diseases. Neurobiol Aging 2017;57:247.e9–13. 10.1016/j.neurobiolaging.2017.05.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Das S, Forer L, Schönherr S. et al. Next-generation genotype imputation service and methods. Nat Genet 2016;48:1284–7. 10.1038/ng.3656. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Dhruba SR. Application of advanced machine learning based approaches in cancer precision medicine. Texas Tech University (TTU), Ph.D. Thesis, 2021. (R Package DMTL).
  5. Erez O, Romero R, Maymon E. et al. The prediction of late-onset preeclampsia: results from a longitudinal proteomics study. PLoS One 2017;12:e0181468. 10.1371/journal.pone.0181468. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Farkas SA, Milutin-Gašperov N, Grce M. et al. Genome-wide DNA methylation assay reveals novel candidate biomarker genes in cervical cancer. Epigenetics 2013;8:1213–25. 10.4161/epi.26346. [DOI] [PubMed] [Google Scholar]
  7. Friedman JH, Hastie T, Tibshirani R.. Regularization paths for generalized linear models via coordinate descent. J Stat Soft 2010;33:1–22. 10.18637/jss.v033.i01. (R package glmnet). [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Gamarnik D, Zadik I. High-dimensional regression with binary coefficients. Estimating squared error and a phase transition. In: Conference on Learning Theory, volume 65 of Proceedings of Machine Learning Research, July 7–10, 2017, Amsterdam, The Netherlands. Maastricht, The Netherlands: ML Research Press, 2017, 948–53.
  9. Hipp G, Vaillant M, Diederich NJ. et al. The Luxembourg Parkinson’s study: a comprehensive approach for stratification and early diagnosis. Front Aging Neurosci 2018;10:326. 10.3389/fnagi.2018.00326. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Jiang Y, He Y, Zhang H.. Variable selection with prior information for generalized linear models via the prior LASSO method. J Am Stat Assoc 2016;111:355–76. 10.1080/01621459.2015.1008363. (R script pLASSO). [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Kawaguchi ES, Li S, Weaver GM. et al. Hierarchical ridge regression for incorporating prior information in genomic studies. J Data Sci 2022;20:34–50. 10.6339/21-JDS1030. (R package xrnet). [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Liu Y, Kang Y, Xing C. et al. A secure federated transfer learning framework. IEEE Intell Syst 2020;35:70–82. 10.1109/MIS.2020.2988525. [DOI] [Google Scholar]
  13. Nalls MA, Blauwendraat C, Vallerga CL. et al. ; International Parkinson’s Disease Genomics Consortium. Identification of novel risk loci, causal insights, and heritable risk for Parkinson’s disease: a meta-analysis of genome-wide association studies. Lancet Neurol 2019;18:1091–102. 10.1016/S1474-4422(19)30320-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Pavelka L, Rauschenberger A, Landoulsi Z. et al. ; NCER-PD Consortium. Age at onset as stratifier in idiopathic Parkinson’s disease—effect of ageing and polygenic risk score on clinical phenotypes. NPJ Parkinsons Dis 2022;8:102. 10.1038/s41531-022-00342-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Rauschenberger A, Glaab E.. Predicting correlated outcomes from molecular data. Bioinformatics 2021;37:3889–95. 10.1093/bioinformatics/btab576. (R package joinet). [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Rauschenberger A, Glaab E, van de Wiel MA.. Predictive and interpretable models via the stacked elastic net. Bioinformatics 2021;37:2012–6. 10.1093/bioinformatics/btaa535. (R package starnet). [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Tay JK, Aghaeepour N, Hastie T. et al. Feature-weighted elastic net: using “features of features” for better prediction. Stat Sin 2023;33:259–79. 10.5705/ss.202020.0226. (R package fwelnet). [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Te Beest DE, Mes SW, Wilting SM. et al. Improved high-dimensional prediction with random forests by the use of co-data. BMC Bioinformatics 2017;18:584. 10.1186/s12859-017-1993-1. (R package CoRF). [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Tian Y, Feng Y.. Transfer learning under high-dimensional generalized linear models. J Am Stat Assoc 2022. (in press). 10.1080/01621459.2022.2071278. (R package glmtrans). [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. van de Wiel MA, Lien TG, Verlaat W. et al. Better prediction by use of co-data: adaptive group-regularized ridge regression. Stat Med 2016;35:368–81. 10.1002/sim.6732. (R package GRridge). [DOI] [PubMed] [Google Scholar]
  21. van de Wiel MA, van Nee MM, Rauschenberger A.. Fast cross-validation for multi-penalty high-dimensional ridge regression. J Comput Graph Stat 2021;30:835–47. 10.1080/10618600.2021.1904962. (R package multiridge). [DOI] [Google Scholar]
  22. van Nee MM, Wessels LF, van de Wiel MA.. Flexible co-data learning for high-dimensional prediction. Stat Med 2021;40:5910–25. 10.1002/sim.9162. (R package ecpc). [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Wolpert DH. Stacked generalization. Neural Netw 1992;5:241–59. 10.1016/S0893-6080(05)80023-1. [DOI] [Google Scholar]
  24. Zeng C, Thomas DC, Lewinger JP.. Incorporating prior knowledge into regularized regression. Bioinformatics 2021;37:514–21. 10.1093/bioinformatics/btaa776. (R package xtune). [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The R package transreg is available on GitHub (https://github.com/lcsb-bds/transreg) and CRAN (https://cran.r-project.org/package=transreg), with the code for the simulations and the applications in a vignette (https://lcsb-bds.github.io/transreg/). We obtained our results using R 4.3.0 (RRID: ID_001905) on a physical machine (aarch64-apple-darwin20, macOS Ventura 13.5.2). Data for the application on cervical cancer are available from van de Wiel et al. (2016), in the R package GRridge in the dataset ‘dataVerlaat’ (source data: Farkas et al. 2013, target data: van de Wiel et al. 2016). Data for the application on pre-eclampsia are available from Erez et al. (2017), in the supporting file ‘pone.0181468.s001.csv’. For the application on Parkinson’s disease, the source data are available from Nalls et al. (2019), in the online file ‘nallsEtAl2019_excluding23andMe_allVariants.tab’, and the target data are available upon request (request.ncer-pd@uni.lu). Information on reproducibility is also available on a frozen page (doi: 10.17881/hczj-3297).


Articles from Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES