Abstract
Multi-task sparse feature learning aims to improve the generalization performance by exploiting the shared features among tasks. It has been successfully applied to many applications including computer vision and biomedical informatics. Most of the existing multi-task sparse feature learning algorithms are formulated as a convex sparse regularization problem, which is usually suboptimal, due to its looseness for approximating an -type regularizer. In this paper, we propose a non-convex formulation for multi-task sparse feature learning based on a novel regularizer. To solve the non-convex optimization problem, we propose a Multi-Stage Multi-Task Feature Learning (MSMTFL) algorithm. Moreover, we present a detailed theoretical analysis showing that MSMTFL achieves a better parameter estimation error bound than the convex formulation. Empirical studies on both synthetic and real-world data sets demonstrate the effectiveness of MSMTFL in comparison with the state of the art multi-task sparse feature learning algorithms.
1 Introduction
Multi-task learning (MTL) exploits the relationships among multiple related tasks to improve the generalization performance. It has been applied successfully to many applications such as speech classification [16], handwritten character recognition [14, 17] and medical diagnosis [2]. One common assumption in multi-task learning is that all tasks should share some common structures including the prior or parameters of Bayesian models [18, 21, 24], a similarity metric matrix [16], a classification weight vector [6], a low rank subspace [4, 13] and a common set of shared features [1, 8, 10, 11, 12, 14, 20].
In this paper, we focus on multi-task feature learning, in which we learn the features specific to each task as well as the common features shared among tasks. Although many multi-task feature learning algorithms have been proposed, most of them assume that the relevant features are shared by all tasks. This is too restrictive in real-world applications [9]. To overcome this limitation, Jalali et al. (2010) [9] proposed an regularized formulation, called dirty model, to leverage the common features shared among tasks. The dirty model allows a certain feature to be shared by some tasks but not all tasks. Jalali et al. (2010) also presented a theoretical analysis under the incoherence condition [5, 15] which is more restrictive than RIP [3, 27]. The regularizer is a convex relaxation for the -type one, which, however, is too loose to well approximate the -type regularizer and usually achieves suboptimal performance (requiring restrictive conditions or obtaining a suboptimal error bound) [23, 26, 27]. To remedy the shortcoming, we propose to use a non-convex regularizer for multi-task feature learning in this paper.
Contributions
We propose to employ a capped- regularized formulation (non-convex) to learn the features specific to each task as well as the common features shared among tasks. To solve the non-convex optimization problem, we propose a Multi-Stage Multi-Task Feature Learning (MSMTFL) algorithm, using the concave duality [26]. Although the MSMTFL algorithm may not obtain a globally optimal solution, we theoretically show that this solution achieves good performance. Specifically, we present a detailed theoretical analysis on the parameter estimation error bound for the MSMTFL algorithm. Our analysis shows that, under the sparse eigenvalue condition which is weaker than the incoherence condition in Jalali et al. (2010) [9], MSMTFL improves the error bound during the multi-stage iteration, i.e., the error bound at the current iteration improves the one at the last iteration. Empirical studies on both synthetic and real-world data sets demonstrate the effectiveness of the MSMTFL algorithm in comparison with the state of the art algorithms.
Notations
Scalars and vectors are denoted by lower case letters and bold face lower case letters, respectively. Matrices and sets are denoted by capital letters and calligraphic capital letters, respectively. The norm, Euclidean norm, norm and Frobenius norm are denoted by ∥ · ∥1, ∥ · ∥, ∥ · ∥∞ and ∥ · ∥F, respectively. | · | denotes the absolute value of a scalar or the number of elements in a set, depending on the context. We define the norm of a matrix X as . We define as {1, …, n} and N(μ, σ2) as a normal distribution with mean μ and variance σ2. For a d×m matrix W and sets , we let be a d × 1 vector with the j-th entry being wji, if , and 0, otherwise. We also let be a d × m matrix with the (j, i)-th entry being wji, if , and 0, otherwise.
2 The Proposed Formulation
Assume we are given m learning tasks associated with training data {(X1, y1), …, (Xm, ym)}, where is the data matrix of the i-th task with each row as a sample; is the response of the i-th task; d is the data dimensionality; ni is the number of samples for the i-th task. We consider learning a weight matrix consisting of the weight vectors for m linear predictive models: yi ≈ fi(Xi) = Xiwi, . In this paper, we propose a non-convex multi-task feature learning formulation to learn these m models simultaneously, based on the capped- regularization. Specifically, we first impose the penalty on each row of W, obtaining a column vector. Then, we impose the capped- penalty [26, 27] on that vector. Formally, we formulate our proposed model as follows:
| (1) |
where l(W) is an empirical loss function of (W; λ (> 0) is a parameter balancing the empirical loss and the regularization; θ (> 0) is a thresholding parameter; wj is the j-th row of the matrix W. In this paper, we focus on the quadratic loss function: .

Intuitively, due to the capped- penalty, the optimal solution of Eq. (1) denoted as W* has many zero rows. For a nonzero row (w*)k, some entries may be zero, due to the -norm imposed on each row of W. Thus, under the formulation in Eq. (1), a certain feature can be shared by some tasks but not all the tasks. Therefore, the proposed formulation can leverage the common features shared among tasks.
The formulation in Eq. (1) is non-convex and is difficult to solve. To this end, we propose a Multi-Stage Multi-Task Feature Learning (MSMTFL) algorithm (see Algorithm 1). Note that if we terminate the algorithm with , the MSMTFL algorithm is equivalent to the regularized multi-task feature learning algorithm (Lasso). Thus, the solution obtained by MSMTFL can be considered as a refinement of that of Lasso. Although Algorithm 1 may not find a globally optimal solution, the solution has good performance. Specifically, we will theoretically show that the solution obtained by Algorithm 1 improves the performance of the parameter estimation error bound during the multi-stage iteration. Moreover, empirical studies also demonstrate the effectiveness of our proposed MSMTFL algorithm. We provide more details about intuitive interpretations, convergence analysis and reproducibility discussions of the proposed algorithm in the full version [7].
3 Theoretical Analysis
In this section, we theoretically analyze the parameter estimation performance of the solution obtained by the MSMTFL algorithm. To simplify the notations in the theoretical analysis, we assume that the number of samples for all the tasks are the same. However, our theoretical analysis can be easily extended to the case where the tasks have different sample sizes.
We first present a sub-Gaussian noise assumption which is very common in the analysis of sparse regularization literature [23, 25, 26, 27].
Assumption 1 Let be the underlying sparse weight matrix and , , where is a random vector with all entries being independent sub-Gaussians: there exists σ > 0 such that .
Remark 1 We call the random variable satisfying the condition in Assumption 1 sub-Gaussian, since its moment generating function is upper bounded by that of the zero mean Gaussian random variable. That is, if a normal random variable x ~ N(0, σ2), then we have .
Remark 2 Based on the Hoeffding's Lemma, for any random variable x ∊ [a, b] and , we have . Therefore, both zero mean Gaussian and zero mean bounded random variables are sub-Guassians. Thus, the sub-Gaussian noise assumption is more general than the Gaussian noise assumption which is commonly used in the literature [9, 11].
We next introduce the following sparse eigenvalue concept which is also common in the analysis of sparse regularization literature [22, 23, 25, 26, 27].
Definition 1 Given 1 ≤ k ≤ d, we define
Remark 3 is in fact the maximum (minimum) eigenvalue of , where is a set satisfying and is a submatrix composed of the columns of Xi indexed by . In the MTL setting, we need to exploit the relations of among multiple tasks.
We present our parameter estimation error bound on MSMTFL in the following theorem:
Theorem 1 Let Assumption 1 hold. Define and . Denote as the number of nonzero rows of . We assume that
| (3) |
| (4) |
where s is some integer satisfying . If we choose λ and θ such that for some :
| (5) |
| (6) |
then the following parameter estimation error bound holds with probability larger than 1 − η:
| (7) |
where is a solution of Eq. (2).
Remark 4 Eq. (3) assumes that the -norm of each nonzero row of is away from zero. This requires the true nonzero coefficients should be large enough, in order to distinguish them from the noise. Eq. (4) is called the sparse eigenvalue condition [27], which requires the eigenvalue ratio to grow sub-linearly with respect to s. Such a condition is very common in the analysis of sparse regularization [22, 25] and it is slightly weaker than the RIP condition [3, 27].
Remark 5 When (corresponds to Lasso), the first term of the right-hand side of Eq. (7) dominates the error bound in the order of
| (8) |
since λ satisfies the condition in Eq. (5). Note that the first term of the right-hand side of Eq. (7) shrinks exponentially as ℓ increases. When ℓ is sufficiently large in the order of , this term tends to zero and we obtain the following parameter estimation error bound:
| (9) |
Jalali et al. (2010) [9] gave an -norm error bound as well as a sign consistency result between and . A direct comparison between these two bounds is difficult due to the use of different norms. On the other hand, the worst-case estimate of the -norm error bound of the algorithm in Jalali et al. (2010) [9] is in the same order with Eq. (8), that is: . When dm is large and the ground truth has a large number of sparse (i.e., is a small constant), the bound in Eq. (9) is significantly better than the ones for the Lasso and Dirty model.
Remark 6 Jalali et al. (2010) [9] presented an -norm parameter estimation error bound and hence a sign consistency result can be obtained. The results are derived under the incoherence condition which is more restrictive than the RIP condition and hence more restrictive than the sparse eigenvalue condition in Eq. (4). From the viewpoint of the parameter estimation error, our proposed algorithm can achieve a better bound under weaker conditions. Please refer to [19, 25, 27] for more details about the incoherence condition, the RIP condition, the sparse eigenvalue condition and their relationships.
Remark 7 The capped- regularized formulation in Zhang (2010) [26] is a special case of our formulation when m = 1. However, extending the analysis from the single task to the multi-task setting is nontrivial. Different from previous work on multi-stage sparse learning which focuses on a single task [26, 27], we study a more general multi-stage framework in the multi-task setting. We need to exploit the relationship among tasks, by using the relations of sparse eigenvalues and treating the -norm on each row of the weight matrix as a whole for consideration. Moreover, we simultaneously exploit the relations of each column and each row of the matrix.
4 Proof Sketch
We first provide several important lemmas (please refer to the full version [7] or supplementary materials for detailed proofs) and then complete the proof of Theorem 1 based on these lemmas.
Lemma 1 Let with . Define such that , provided there exists ( is a set consisting of the indices of all entries in the nonzero rows of ). Under the conditions of Assumption 1 and the notations of Theorem 1, the followings hold with probability larger than 1 − η:
| (10) |
| (11) |
Lemma 1 gives bounds on the residual correlation with respect to . We note that Eq. (10) and Eq. (11) are closely related to the assumption on λ in Eq. (5) and the second term of the right-hand side of Eq. (7) (error bound), respectively. This lemma provides a fundamental basis for the proof of Theorem 1.
Lemma 2 Use the notations of Lemma 1 and consider such that . Let be a solution of Eq. (2) and . Denote . Let , and , . If , then the following inequality holds at any stage ℓ ≥ 1:
Denote , and notice that . Lemma 2 sasy that is upper bounded in terms of , which indicates that the error of the estimated coefficients locating outside of should be small enough. This provides an intuitive explanation why the parameter estimation error of our algorithm can be small.
Lemma 3 Using the notations of Lemma 2, we denote with being defined as in Lemma 1 and . Let be the indices of the largest s coefficients (in absolute value) of , , and . Then, the following inequalities hold at any stage ℓ ≥ 1:
| (12) |
| (13) |
Lemma 3 is established based on Lemma 2, by considering the relationship between Eq. (5) and Eq. (10), and the specific definition of . Eq. (12) provides a parameter estimation error bound in terms of ℓ2,1-norm by and the regularization parameters (see the definition of in Lemma 2). This is the result directly used in the proof of Theorem 1. Eq. (13) states that the error bound is upper bounded in terms of λ, the right-hand side of which constitutes the shrinkage part of the error bound in Eq. (7).
Lemma 4 Let , with some . is defined in Lemma 1. Then under the condition of Eq. (3), we have:
Lemma 4 establishes an upper bound of by , which is critical for building the recursive relationship between and in the proof of Theorem 1. This recursive relation is crucial for the shrinkage part of the error bound in Eq. (7).
4.1 Proof of Theorem 1
Proof For notational simplicity, we denote the right-hand side of Eq. (11) as:
| (14) |
Based on , Lemma 1 and Eq. (5), the followings hold with probability larger than 1 − η:
| (15) |
where the last inequality follows from , . According to Eq. (12), we have:
In the above derivation, the first inequality is due to Eq. (12); the second inequality is due to the assumption in Theorem 1, Eq. (15) and Lemma 4; the third inequality is due to Eq. (6); the last inequality follows from Eq. (13) and 1 − 0.8ℓ ≤ 1 (ℓ ≥ 1). Thus, following the inequality , we obtain:
Substituting Eq. (14) into the above inequality, we verify Theorem 1.
5 Experiments
We compare our proposed MSMTFL algorithm with three competing multi-task feature learning algorithms: ℓ1-norm multi-task feature learning algorithm (Lasso), ℓ1,2-norm multi-task feature learning algorithm (L1,2) [14] and dirty model multi-task feature learning algorithm (DirtyMTL) [9]. In our experiments, we employ the quadratic loss function for all the compared algorithms.
5.1 Synthetic Data Experiments
We generate synthetic data by setting the number of tasks as m and each task has n samples which are of dimensionality d; each element of the data matrix for the i-th task is sampled i.i.d. from the Gaussian distribution N(0, 1) and we then normalize all columns to length 1; each entry of the underlying true weight is sampled i.i.d. from the uniform distribution in the interval [−10, 10]; we randomly set 90% rows of as zero vectors and 80% elements of the remaining nonzero entries as zeros; each entry of the noise is sampled i.i.d. from the Gaussian distribution N(0, σ2); the responses are computed as
We first report the averaged parameter estimation error vs. Stage (ℓ) plots for MSMTFL (Figure 1). We observe that the error decreases as ℓ increases, which shows the advantage of our proposed algorithm over Lasso. This is consistent with the theoretical result in Theorem 1. Moreover, the parameter estimation error decreases quickly and converges in a few stages.
Figure 1.

Averaged parameter estimation error vs. Stage plots for MSMTFL on the synthetic data set (averaged over 10 runs). Here we set . Note that corresponds to Lasso; the results show the stage-wise improvement over Lasso.
We then report the averaged parameter estimation error in comparison with four algorithms in different parameter settings (Figure 2). For a fair comparison, we compare the smallest estimation errors of the four algorithms in all the parameter settings [25, 26]. As expected, the parameter estimation error of the MSMTFL algorithm is the smallest among the four algorithms. This empirical result demonstrates the effectiveness of the MSMTFL algorithm. We also have the following observations: (a) When λ is large enough, all four algorithms tend to have the same parameter estimation error. This is reasonable, because the solutions 's obtained by the four algorithms are all zero matrices, when λ is very large. (b) The performance of the MSMTFL algorithm is similar for different θ's, when λ exceeds a certain value.
Figure 2.

Averaged parameter estimation error vs. λ plots on the synthetic data set (averaged over 10 runs). MSMTFL has the smallest parameter estimation error among the four algorithms. Both DirtyMTL and MSMTFL have two parameters; we set λs/λb = 1, 0.5, 0.2, 0.1 for DirtyMTL (1/m ≤ λs/λb ≤ 1 was adopted in Jalali et al. (2010) [9]) and θ/λ = 50m, 10m, 2m, 0.4m for MSMTFL.
5.2 Real-World Data Experiments
We conduct experiments on two real-world data sets: MRI and Isolet data sets. (1) The MRI data set is collected from the ANDI database, which contains 675 patients' MRI data preprocessed using FreeSurfer1. The MRI data include 306 features and the response (target) is the Mini Mental State Examination (MMSE) score coming from 6 different time points: M06, M12, M18, M24, M36, and M48. We remove the samples which fail the MRI quality controls and have missing entries. Thus, we have 6 tasks with each task corresponding to a time point and the sample sizes corresponding to 6 tasks are 648, 642, 293, 569, 389 and 87, respectively. (2) The Isolet data set2 is collected from 150 speakers who speak the name of each English letter of the alphabet twice. Thus, there are 52 samples from each speaker. The speakers are grouped into 5 subsets which respectively include 30 similar speakers, and the subsets are named Isolet1, Isolet2, Isolet3, Isolet4, and Isolet5. Thus, we naturally have 5 tasks with each task corresponding to a subset. The 5 tasks respectively have 1560, 1560, 1560, 1558, and 1559 samples (Three samples are historically missing), where each sample includes 617 features and the response is the English letter label (1–26).
In the experiments, we treat the MMSE and letter labels as the regression values for the MRI data set and the Isolet data set, respectively. For both data sets, we randomly extract the training samples from each task with different training ratios (15%, 20% and 25%) and use the rest of samples to form the test set. We evaluate the four multi-task feature learning algorithms in terms of normalized mean squared error (nMSE) and averaged means squared error (aMSE), which are commonly used in multi-task learning problems [28, 29]. For each training ratio, both nMSE and aMSE are averaged over 10 random splittings of training and test sets and the standard deviation is also shown. All parameters of the four algorithms are tuned via 3-fold cross validation.
Table 1 and Figure 3 show the experimental results in terms of averaged nMSE (aMSE) and the standard deviation. From these results, we observe that: (a) Our proposed MSMTFL algorithm out-performs all the competing feature learning algorithms on both data sets, with the smallest regression errors (nMSE and aMSE) as well as the smallest standard deviations. (b) On the MRI data set, the MSMTFL algorithm performs well even in the case of a small training ratio. The performance for the 15% training ratio is comparable to that for the 25% training ratio. (c) On the Isolet data set, when the training ratio increases from 15% to 25%, the performance of the MSMTFL algorithm increases and the superiority of the MSMTFL algorithm over the other three algorithms is more significant. Our results demonstrate the effectiveness of the proposed algorithm.
Table 1.
Comparison of four multi-task feature learning algorithms on the MRI data set in terms of averaged nMSE and aMSE (standard deviation), which are averaged over 10 random splittings.
| measure | traning ratio | Lasso | L1,2 | DirtyMTL | MSMTFL |
|---|---|---|---|---|---|
| nMSE | 0.15 | 0.6651(0.0280) | 0.6633(0.0470) | 0.6224(0.0265) | 0.5539(0.0154) |
| 0.20 | 0.6254(0.0212) | 0.6489(0.0275) | 0.6140(0.0185) | 0.5542(0.0139) | |
| 0.25 | 0.6105(0.0186) | 0.6577(0.0194) | 0.6136(0.0180) | 0.5507(0.0142) | |
|
| |||||
| aMSE | 0.15 | 0.0189(0.0008) | 0.0187(0.0010) | 0.0172(0.0006) | 0.0159(0.0004) |
| 0.20 | 0.0179(0.0006) | 0.0184(0.0005) | 0.0171(0.0005) | 0.0161(0.0004) | |
| 0.25 | 0.0172(0.0009) | 0.0183(0.0006) | 0.0167(0.0008) | 0.0157(0.0006) | |
Figure 3.

Averaged test error (nMSE and aMSE) vs. training ratio plots on the Isolet data set. The results are averaged over 10 random splittings.
6 Conclusions
In this paper, we propose a non-convex multi-task feature learning formulation based on the capped- regularization. The proposed formulation learns the specific features of each task as well as the common features shared among tasks. We propose to solve the non-convex optimization problem by employing a Multi-Stage Multi-Task Feature Learning (MSMTFL) algorithm, using concave duality. We also present a detailed theoretical analysis in terms of the parameter estimation error bound for the MSMTFL algorithm. The analysis shows that our MSMTFL algorithm achieves good performance under the sparse eigenvalue condition, which is weaker than the incoherence condition. Experimental results on both synthetic and real-world data sets demonstrate the effectiveness of our proposed MSMTFL algorithm in comparison with the state of the art multi-task feature learning algorithms. In our future work, we will focus on a general non-convex regularization framework for multi-task feature learning settings (involving different loss functions and non-convex regularization terms) and derive theoretical bounds.
Acknowledgements
This work is supported in part by 973 Program (2013CB329503), NSFC (Grant No. 91120301, 60835002 and 61075004), NIH (R01 LM010730) and NSF (IIS-0953662, CCF-1025177).
Footnotes
This work was completed when the first author visited Arizona State University.
References
- [1].Argyriou A, Evgeniou T, Pontil M. Convex multi-task feature learning. Machine Learning. 2008;73(3):243–272. [Google Scholar]
- [2].Bi J, Xiong T, Yu S, Dundar M, Rao R. An improved multi-task learning approach with applications in medical diagnosis. Machine Learning and Knowledge Discovery in Databases. 2008:117–132. [Google Scholar]
- [3].Candes E, Tao T. Decoding by linear programming. IEEE Transactions on Information Theory. 2005;51(12):4203–4215. [Google Scholar]
- [4].Chen J, Liu J, Ye J. Learning incoherent sparse and low-rank patterns from multiple tasks. SIGKDD. 2010:1179–1188. doi: 10.1145/2086737.2086742. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [5].Donoho D, Elad M, Temlyakov V. Stable recovery of sparse overcomplete representations in the presence of noise. IEEE Transactions on Information Theory. 2006;52(1):6–18. [Google Scholar]
- [6].Evgeniou T, Pontil M. Regularized multi–task learning. SIGKDD. 2004:109–117. [Google Scholar]
- [7].Gong P, Ye J, Zhang C. Multi-stage multi-task feature learning. arXiv:1210.5806. 2012 [Google Scholar]
- [8].Gong P, Ye J, Zhang C. Robust multi-task feature learning. SIGKDD. 2012:895–903. doi: 10.1145/2339530.2339672. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Jalali A, Ravikumar P, Sanghavi S, Ruan C. A dirty model for multi-task learning. NIPS. 2010:964–972. [Google Scholar]
- [10].Kim S, Xing E. Tree-guided group lasso for multi-task regression with structured sparsity. ICML. 2009:543–550. [Google Scholar]
- [11].Lounici K, Pontil M, Tsybakov A, Van De Geer S. Taking advantage of sparsity in multi-task learning. COLT. 2009:73–82. [Google Scholar]
- [12].Negahban S, Wainwright M. Joint support recovery under high-dimensional scaling: Benefits and perils of -regularization. NIPS. 2008:1161–1168. [Google Scholar]
- [13].Negahban S, Wainwright M. Estimation of (near) low-rank matrices with noise and high-dimensional scaling. The Annals of Statistics. 2011;39(2):1069–1097. [Google Scholar]
- [14].Obozinski G, Taskar B, Jordan M. Multi-task feature selection. Statistics Department, UC Berkeley, Tech. Rep. 2006 [Google Scholar]
- [15].Obozinski G, Wainwright M, Jordan M. Support union recovery in high-dimensional multivariate regression. Annals of statistics. 2011;39(1):1–47. [Google Scholar]
- [16].Parameswaran S, Weinberger K. Large margin multi-task metric learning. NIPS. 2010:1867–1875. [Google Scholar]
- [17].Quadrianto N, Smola A, Caetano T, Vishwanathan S, Petterson J. Multitask learning without label correspondences. NIPS. 2010:1957–1965. [Google Scholar]
- [18].Schwaighofer A, Tresp V, Yu K. Learning gaussian process kernels via hierarchical bayes. NIPS. 2005:1209–1216. [Google Scholar]
- [19].Van De Geer S, Bühlmann P. On the conditions used to prove oracle results for the lasso. Electronic Journal of Statistics. 2009;3:1360–1392. [Google Scholar]
- [20].Yang X, Kim S, Xing E. Heterogeneous multitask learning with joint sparsity constraints. NIPS. 2009:2151–2159. [Google Scholar]
- [21].Yu K, Tresp V, Schwaighofer A. Learning gaussian processes from multiple tasks. ICML. 2005:1012–1019. [Google Scholar]
- [22].Zhang C, Huang J. The sparsity and bias of the lasso selection in high-dimensional linear regression. The Annals of Statistics. 2008;36(4):1567–1594. [Google Scholar]
- [23].Zhang C, Zhang T. A general theory of concave regularization for high dimensional sparse estimation problems. Statistical Science. 2012 [Google Scholar]
- [24].Zhang J, Ghahramani Z, Yang Y. Learning multiple related tasks using latent independent component analysis. NIPS. 2006:1585–1592. [Google Scholar]
- [25].Zhang T. Some sharp performance bounds for least squares regression with regularization. The Annals of Statistics. 2009;37:2109–2144. [Google Scholar]
- [26].Zhang T. Analysis of multi-stage convex relaxation for sparse regularization. JMLR. 2010;11:1081–1107. [Google Scholar]
- [27].Zhang T. Multi-stage convex relaxation for feature selection. Bernoulli. 2012 [Google Scholar]
- [28].Zhang Y, Yeung D. Multi-task learning using generalized t process. AISTATS. 2010 [Google Scholar]
- [29].Zhou J, Chen J, Ye J. Clustered multi-task learning via alternating structure optimization. NIPS. 2011:702–710. [PMC free article] [PubMed] [Google Scholar]
