Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2014 Oct 8.
Published in final edited form as: Adv Neural Inf Process Syst. 2012;2012:1430–1438.

Q-MKL: Matrix-induced Regularization in Multi-Kernel Learning with Applications to Neuroimaging*

Chris Hinrichs †,§, Vikas Singh §,, Jiming Peng , Sterling C Johnson
PMCID: PMC4189130  NIHMSID: NIHMS418899  PMID: 25309107

Abstract

Multiple Kernel Learning (MKL) generalizes SVMs to the setting where one simultaneously trains a linear classifier and chooses an optimal combination of given base kernels. Model complexity is typically controlled using various norm regularizations on the base kernel mixing coefficients. Existing methods neither regularize nor exploit potentially useful information pertaining to how kernels in the input set ‘interact’; that is, higher order kernel-pair relationships that can be easily obtained via unsupervised (similarity, geodesics), supervised (correlation in errors), or domain knowledge driven mechanisms (which features were used to construct the kernel?). We show that by substituting the norm penalty with an arbitrary quadratic function Q Inline graphic 0, one can impose a desired covariance structure on mixing weights, and use this as an inductive bias when learning the concept. This formulation significantly generalizes the widely used 1- and 2-norm MKL objectives. We explore the model’s utility via experiments on a challenging Neuroimaging problem, where the goal is to predict a subject’s conversion to Alzheimer’s Disease (AD) by exploiting aggregate information from many distinct imaging modalities. Here, our new model outperforms the state of the art (p-values ⪡ 10−3). We briefly discuss ramifications in terms of learning bounds (Rademacher complexity).

1 Introduction

Kernel learning methods (such as Support Vector Machines) are conceptually simple, strongly rooted in statistical learning theory, and can often be formulated as a convex optimization problem. As a result, SVMs have come to dominate the landscape of supervised learning applications in bioinformatics, computer vision, neuroimaging, and many other domains. A standard SVM-based ‘learning system’ may be conveniently thought of as a composition of two modules [1, 2, 3, 4]: (1) Feature pre-processing, and (2) a core learning algorithm. The design of a kernel (feature pre-processing) may involve using different sets of extracted features, dimensionality reductions, or parameterizations of the kernel functions. Each of these alternatives produces a distinct kernel matrix. While much research has focused on efficient methods for the latter (i.e., support vector learning) step, specific choices of feature pre-processing are frequently a dominant factor in the system’s overall performance as well, and may involve significant user effort. Multi-kernel learning [5, 6, 7] transfers a part of this burden from the user to the algorithm. Rather than selecting a single kernel, MKL offers the flexibility of specifying a large set of kernels corresponding to the many options (i.e., kernels) available, and additively combining them to construct an optimized, data-driven Reproducing Kernel Hilbert Space (RKHS) – while simultaneously finding a max-margin classifier. MKL has turned out to be very successful in many applications: on several important Vision problems (such as image categorization), some of the best known results on community benchmarks come from MKL-type methods [8, 9]. In the context of our primary motivating application, the current state of the art in multi-modality neuroimaging-based Alzheimer’s Disease (AD) prediction [10] is achieved by multi-kernel methods [3, 4], where each imaging modality spawns a kernel, or set of kernels.

In allowing the user to specify an arbitrary number of base kernels for combination MKL provides more expressive power, but this comes with the responsibility to regularize the kernel mixing coefficients so that the classifier generalizes well. While the importance of this regularization cannot be overstated, it is also a fact that commonly used lp norm regularizers operate on kernels separately, without explicitly acknowledging dependencies and interactions among them. To see how such dependencies can arise in practice, consider our neuroimaging learning problem of interest: the task of learning to predict the onset of AD. A set of base kernels K1, … , KM are derived from several different medical imaging modalities (MRI; PET), image processing methods (morphometric; anatomical modelling), and kernel functions (linear; RBF). Some features may be shared between kernels, or kernel functions may use similar parameters. As a result we expect the kernels’ behaviors to exhibit some correlational, or other cluster structure according to how they were constructed. (See Fig. 2 (a) and related text, for a concrete discussion of these behaviors in our problem of interest.) We will denote this relationship as QRM×M.

Figure 2.

Figure 2

Covariance Q used in AD experiments (a); three least eigen-vectors of its graph Laplacian (b-d); outer product of optimized β (e). Note the block structure in (a) relating to the imaging modalities and kernel functions.

Ideally, the regularization process should reflect these dependencies encoded by Q, as they can significantly impact the learning characteristics of a linearly combined kernel. Some extensions work at the level of group membership (e.g., [11]), but do not explicitly quantify these interactions. Instead, rather than penalizing covariances or inducing sparsity among groups of kernels, it may be beneficial to reward such covariances, so as to better reflect a latent cluster structure between kernels. In this paper, we show that a rich class of regularization schemes are possible under a new MKL formulation which regularizes on Q directly – the model allows one to exploit domain knowledge (as above) and statistical measures of interaction between kernels, employ estimated error covariances in ways that are not possible with lp-norm regularization, or, encourage sparsity, group sparsity or non-sparsity as needed – all within a convex optimization framework. We call this form of multi-kernel learning, Q-norm MKL or “Q-MKL”. This paper makes the following contributions: (a) presents our new Q-MKL model which generalizes 1- (and 2-) norm MKL models, (b) provides a learning theoretic result showing that Q-MKL can improve MKL’s generalization error rate, (c) develops efficient optimization strategies (to be distributed as a part of the Shogun toolbox), and (d) provides empirical results demonstrating statistically significant gains in accuracy on the important AD prediction problem.

1.1 Background

The development of MKL methods began with [5], which showed that the problem of learning the right kernel for an input problem instance could be formulated as a Semi-Definite Program (SDP). Subsequent papers have focused on designing more efficient optimization methods, which have enabled its applications to large-scale problem domains. To this end, the model in [5] was shown to be solvable as a Second Order Cone Program [12], a Semi-Infinite Linear Program [6], and via gradient descent methods in the dual and primal [7, 13]. More recently, efforts have focused on generalizing MKL to arbitrary p-norm regularizers where p > 1 [13, 14] while maintaining overall efficiency. In [14], the authors briefly mentioned that more general norms may be possible, but this issue was not further examined. A nonlinear “hyperkernel” method was proposed [15] which implicitly maps the kernels themselves to an implicit RKHS, however this method is computationally very demanding, (it has 4th order interactions among training examples). The authors of [16] proposed to first select the sub-kernel weights by minimizing an objective function derived from Normalized Cuts, and subsequently train an SVM on the combined kernel. In [17, 2], a method was proposed for selecting an optimal finite combination from an infinite parameter space of kernels. Contemporary to these results, [18] showed that if a large number of kernels had a desirable shared structure (e.g., followed directed acyclic dependencies), extensions of MKL could still be applied. Recently in [8], a set of base classifiers were first trained using each kernel and were then boosted to produce a strong multi-class classifier. At this time, MKL methods [8, 9] provide some of the best known accuracy on image categorization datasets such as Caltech101/256 (see www.robots.ox.ac.uk/~vgg/software/MKL/). Next, we describe in detail the motivation and theoretical properties of Q-MKL .

2 From MKL to Q-MKL

MKL Models

Adding kernels corresponds to taking a direct sum of Reproducing Kernel Hilbert spaces (RKHS), and scaling a kernel by a constant c scales the axes of it’s RKHS by c. In the MKL setting, the SVM margin regularizer 12w2 becomes a weighted sum 12m=1MwmHm2βm over contributions from RKHS’s H1,,HM, where the vector of mixing coefficients β scales each respective RKHS [14]. A norm penalty on β ensures that the units in which the margin is measured are meaningful (provided the base kernels are normalized). The MKL primal problem is given as

minw,b,β0,ξ012mMwmHm2βm+Cinξi+βp2s.t.yi(mMwm,ϕm(xi)Hm+b)1ξi, (1)

where ϕm(x) is the (potentially unknown) transformation from the original data space to the mth RKHS Hm. As in SVMs, we turn to the dual problem to see the role of kernels:

max0αCαT112Gq,GRM;Gm=(αy)TKm(αy), (2)

where ○ denotes element-wise multiplication, and the dual q-norm follows the identity 1p+1q=1. Note that the primal norm penalty βp2 becomes a dual-norm on the vector G. At optimality, wm = βm(αy)T ϕm(X), and so the term Gm=(αy)TKm(αy)=wmHm2βm2 is the vector of scaled classifier norms. This shows that the dual norm is tied to how MKL measures the margin in each RKHS.

The Q-MKL model

The key characteristic of Q-MKL is that the standard (squared) lp-norm penalty on β, along with the corresponding dual-norm penalty in (2), is substituted with a more general class of quadratic penalty functions, expressed as βTQβ=βQ2.βQ=βTQβ is a Mahalanobis (matrix-induced) norm so long as Q Inline graphic 0. In this framework, the burden of choosing a kernel is deferred to a choice of Q-function. This approach gives the algorithm greater flexibility while controlling model complexity, as we will discuss shortly. The model we optimize is,

minw,b,β0,ξ012mMwmHm2βm+Cinξi+βTQβs.t.yi(mMwm,ϕm(xi)Hm+b)1ξi, (3)

where the last objective term provides a bias relative to βTQβ. The dual problem becomes maxα αT112GTQ1G. It is easy to see that if Q = 1M×M, we obtain the P = 1 from of (1), i.e., 1-norm MKL, as a special case because βT1M×Mβ=β12. On the other hand, setting QtoIM×M (identity), reduces to 2-norm MKL.

3 The case for Q-MKL

Extending the MKL regularizer to arbitrary quadratics Q Inline graphic 0 significantly expands the richness of the MKL framework; yet we can show that for reasonable choices of Q, this actually decreases MKL’s learning-theoretic complexity. Joachims et al. [19] derived a theoretical generalization error bound on kernel combinations which depends on the degree of redundancy between support vectors in SVMs trained on base kernels individually. Using this type of correlational structure, we can derive a Q function between kernels to automatically select a combination of kernels which will maximize this bound. This type of Q function can be shown to have lower Rademacher complexity, (see below,) while simultaneously decreasing the error bound from [19], which does not directly depend on Rademacher complexity.

3.1 Virtual Kernels, Rademacher Complexity and Renyi Entropy

If we decompose Q into its component eigen-vectors, we can see that each eigen-vector defines a linear combination of kernels. This observation allows us to analyze Q-MKL in terms of these objects, which we will refer to as Virtual Kernels. We first show that as Q−1’s eigen-values decay, so do the traces of the virtual kernels. Assuming Q−1 has a bounded, non-uniform spectrum, this property can then be used to analyze, (and bound), Q-MKL’s Rademacher complexity, which has been shown to depend on the traces of the base kernels. We then offer a few observations on how Q−1’s Renyi entropy [20] relates to these learning theoretic bounds.

Virtual Kernels

In the following, assume that Q Inline graphic 0, and has eigen-decomposition Q = V ΛV, with V = {v1, … , vM}. First, observe that because Q’s eigen-vectors provide an orthonormal basis of RM,βRM can be expressed as a linear combination in this basis with γ as its coefficients: β=iγivi=Vγ. Substituting in βT Qβ we have

βTQβ=(γTVT)VΛVT(Vγ)=γT(VTV)Λ(VTV)γ=γTΛγ=iγi2λi (4)

This simple observation offers an alternate view of what Q-MKL is actually optimizing. Each eigen-vector vi of Q can be used to define a linear combination of kernels, which we will refer to as virtual kernel K~i=mvi(m)Km. Note that if K~i¯0, then they each define an independent RKHS. This can be ensured by choosing Q in a specific way, if desired. This leads to the following result:

Lemma 1

If K~i¯0, then Q-MKL is equivalent to 2-norm MKL using virtual kernels instead of base kernels.

Proof. Let μi=γiλi. Then βTQβ=μ22, (eq. 4) and K=mβmKm=mMiMγivi(m)Km=iMμiλ12mMvi(m)Km=iMμiK~i, where K~i=λ12mMvi(m)Km is the ith virtual kernel. The learned kernel K* is a weighted combination of virtual kernels, and the coefficients are regularized under a squared 2-norm.

Rademacher Complexity in MKL

With this result in hand, we can now evaluate the Rademacher complexity of Q-MKL by using a recent result for p-norm MKL. We first state a theorem from [21], which relates the Rademacher complexity of MKL to the traces of its base kernels.

Theorem 1

([21]) The empirical Rademacher complexity on a sample set S of size n, with M base kernels is given as follows (with η0=2322),

RS(HMp)η0quqn (5)

where u = [Tr(K1), … , Tr(KM)]T and 1p+1q=1.

The bound in (5) shows that the Rademacher complexity RS() depends on ∥uq, which is a norm on the traces of the base kernels. Assuming the base kernels are normalized to have unit trace, the bound for p = q = 2-norm MKL is governed by u2=M. However, in Q-MKL the virtual kernels traces are not equal, and are in fact given by Tr(K~i)=1Tviλi. With this expression for the traces of the virtual kernels, we can now prove that the bound given in (5) is strictly decreased as long as the eigen-values Ψi of Q−1 are in the range (0; 1]. (Adding 1 to the diagonal of Q is sufficient to guarantee this.)

Theorem 2

If Q1IM×M and K~i¯0 then the bound on Rademacher complexity given in (5) is strictly lower for Q-MKL than for 2-norm MKL.

Proof. By Lemma 1, we have that the bound in (5) will decrease if ∥u2, the norm on the virtual kernel traces, decreases. As shown above, the virtual kernel traces are given as Tr(K~i)=ψi1Tvi, meaning that u22=iNψi(1Tvi)2=iNψi1TviviT1=1TQ11. Clearly, this sum is maximal for Ψi = 1, ∀i, which is true if and only if Q1=IM×M. This means that when QIM×M, then the bound in (5) is strictly decreased.

Note that requiring the virtual kernels to be p.s.d., while achievable (see supplements,) is somewhat restrictive. In practice, such a Q matrix may not differ substantially from the identity matrix. We therefore provide the following result which frees us from this restriction, and has more practical significance.

Theorem 3

Q-MKL is equivalent to the following model:

minw,b,μ,ξ012mMwmVm2μm+Cinξi+μ22s.t.yi(mMwm,ϕm(xi)Vm+b)1ξi,Q12μ0, (6)

where ϕm() is the feature transform mapping data space to the mth virtual kernel, denoted as Vm.

While the virtual kernels themselves may be indefinite, recall that μ=Q12β, and so the constraint Q12μ0 is equivalent to β ≥ 0, guaranteeing that the combined kernel will be p.s.d. This formulation is slightly different than the 2-norm MKL formulation, however it does not alter the theoretical guarantee of [21], providing a stronger result.

Renyi Entropy

Renyi entropy [20] significantly generalizes the usual notion of Shannon entropy [22, 23, 24], has applications in Statistics and many other fields, and has recently been proposed as an alternative to PCA [22]. Thm. 2 points to an intuitive explanation of where the benefit from a Q regularizer comes from as well, if we analyze the Renyi entropy of the distribution on kernels defined by Q−1, if we treat Q−1 as a kernel density estimator. The quadratic Renyi entropy of a probability measure is given as,

H(p)=logp2(x)dx.

Now, if we use a kernel function (i.e., Q−1), and a finite sample (i.e., base kernels), as a kernel density estimator, (cf. [15],) then with some normalization we can derive an estimate of the underlying probability p^, which is a distribution over base kernels. We can then interpret its Renyi entropy as a complexity measure on the space of combined kernels. Eq. (5.2) in [23] relates the virtual kernel traces to the Renyi entropy estimator of Q−1 as p^2(x)dx=1N21TQ11,1 which leads to a nice connection to Thm. 2. This view informs us that setting Q1=IM×M, (i.e., 2-norm MKL), has maximal Renyi entropy because it is maximally uninformative; adding structure to Q−1 concentrates p^, reducing both its Renyi entropy, and Rademacher complexity together.

This series of results suggests an entirely new approach to analyzing the Rademacher complexity of MKL methods. The proof of Thm. 2 relies on decreasing a norm on the virtual kernel traces, which we now see directly relates to the Renyi entropy of Q−1, as well as with decreasing the Rademacher complexity of the search space of combined kernels. It is even possible that by directly analyzing Renyi entropy in a multi-kernel setting, this conjecture may be useful in deriving analogous bounds in, e.g., Indefinite Kernel Learning [25], because the virtual kernels are indefinite in general.

3.2 Special Cases: Q-SVM and relative margin

Before describing our optimization strategy, we discuss several variations on the Q-MKL model.

Q-SVM

An interesting special case of Q-MKL is Q-SVM, which generalizes several recent, (but independently developed,) models in the literature [26, 27, 28]. If it were the case that the base kernels were rank-1, (i.e., singleton features,) then each coefficient βm effectively becomes a feature weight, and a 2-norm penalty on β is a 2-norm penalty on weights. Q-MKL therefore reduces to a form of SVM in which the margin regularizer ∥w∥2 becomes wT Qw. Thus, in such cases we can reduce the Q-MKL model to a simple QP, which we call Q-SVM . Please refer to the supplements for details, and some experimental results.

Relative Margin

Several interesting extensions to the SVM and MKL frameworks have been proposed which focus on the relative margin methods [29, 30] which maximize the margin relative to the spread of the data. In particular Q-MKL can be easily modified to incorporate the Relative Margin Machine (RMM) model [29] by replacing Module 1 as in (7) with the RMM objective. Our alternating optimization approach, (described next,) is not affected by this addition; however, the additional constraints would mean that SMO-based strategies would not be applicable.

4 Optimization

We now present the core engine to solve (3). Most MKL implementations make use of an alternating minimization strategy which first minimizes the objective in terms of the SVM parameters, and then with respect to the sub-kernel weights β. Since the MKL problem is convex, this method leads to global convergence [7, 14] and minor modifications to standard SVM implementations are sufficient. Q-MKL generalizes βp2 to arbitrary convex quadratic functions, while the feasible set is the same as for MKL. This directly gives,

Property 1

The Q-MKL model in (3) is convex.

We will broadly follow this strategy, but as will become clear shortly, interaction between sub-kernel weights makes the optimization of β more involved (than [6, 14]), and requires alternative solution mechanisms. We may consider this process as a composition of two modules: one which solves for SVM dual parameters (α) with fixed β, and the other for solving for β with fixed α:

graphic file with name nihms-418899-f0005.jpg

Using a result from [14] we can replace the βT Qβ objective term with a quadratic constraint, which gives the problem in (8). Notice that (8) has a sum of ratios with optimization variables in the denominator, while the constraint is quadratic – this means that standard convex optimization toolkits may not be able to solve this problem without significant reformulation from its canonical form in (8).

Our approach is to search for a stationary point by representing the gradient as a non-linear system. Writing the gradient in terms of the Lagrange multiplier δ, and setting it equal to 0 gives:

wmHm2βm2δ(Qβ)m=0,m{1,,M}. (9)

We now seek to eliminate δ so that the non-linear system will be limited to quadratic terms in β, allowing us to use a non-linear system solver. Let W=Diag(w1H12,,wMHM2), and β2=(β12,,βM2). We can then write Wβ−2 = δ(Qβ). Now, solving for β (on the right hand side) gives

β=1δQ1Wβ2 (10)

Because Q Inline graphic 0, and β ≥ 0, at optimality the constraint βT Qβ ≤ 1 must be active. So, we can plug in the above identity to solve for δ,

1=(1δQ1Wβ2)TQ(1δQ1Wβ2)δ=(Wβ2)TQ1(Wβ2)=Wβ2Q1, (11)

which shows that δ effectively normalizes Wβ−2 according to Q−1. We can now solve (10) in terms of β using a nonlinear root finder, such as the GNU Scientific Library; in practice this method is quite efficient, typically requiring 10 to 20 outer iterations. Putting these parts together, we can propose following algorithm for optimizing Q-MKL: graphic file with name nihms-418899-f0007.jpg

4.1 Convergence

We can show that our model can be solved optimally by noting that Module 2 can be precisely optimized at each step. If Module 2 cannot be solved precisely, then Algorithm 1 may not converge. The following result assures us that indeed Module 2 can be solved precisely by reducing it to a convex Semi-Definite Program (SDP).

Theorem 4

The solution to Problem (8) is the same as the solution to the following SDP:

minν0,β0,ZRM×MwTν (12)
subject to[νm11βm]¯0,m[1βTβZ]¯0,Tr(QZ)1. (13)

Proof. The first PSD constraint (13) requires that νm=βm1, meaning that objective (12) is the same as that of Problem (8). From the second we have Z = ββT , and so Tr(QZ) = βT Qβ; therefore the feasible sets are equivalent.

The last PSD constraint is only necessary to ensure that βT Qβ ≤ 1, and can be replaced with that quadratic constraint. Doing so yields a Second-Order Cone Program (SOCP) which is also amenable to standard solvers. Note that it is not necessary to solve for β as an SDP, though it is concievable that in some cases this will nevertheless be an effective solution mechanism, depending on the size and characteristics of the problem.

5 Experiments

We performed extensive experiments to validate the Q-MKL model, examine the effect our regularization scheme has on β, and to assess its advantages in the context of our motivating neuroimaging application. In these main experiments, we demonstrate how domain knowledge can be adapted to improve the algorithm’s performance. Our focus on a practical application is intended as a demonstration of how domain knowledge can be seamlessly incorporated into a learning model, giving significant gains in accuracy. We also performed experiments on the UCI repositories, which are described in detail in the supplements. Briefly, in these experiments Q-MKL performed as well as, or better than, 1- and 2-norm MKL on most datasets, showing that even in the absence of significant domain knowledge, Q-MKL can still perform about as well as existing MKL methods.

5.1 Image preprocessing and methodology

Our main numerical experiments evaluate Q-MKL in the context of the motivating scenario described above, in which we wish to train a discriminative disease model of Alzheimer’s Disease (AD) using brain imaging data. In these experiments, we used brain scans of Alzheimer’s Disease (AD) patients and Cognitively Normal healthy controls (CN) from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) [31] in a set of cross-validation experiments. ADNI is a landmark study sponsored by the NIH, major pharmaceuticals and others to determine the extent to which multi-modal brain imaging can help predict on-set, and monitor progression of, AD. To this end, MKL type methods have already defined the state of the art for this application [3, 4]. For our experiments, 48 AD subjects and 66 controls were chosen who had both T1-weighted MR scans and Fluoro-Deoxy-Glucose PET (FDG-PET) scans at two time-points two years apart. Standard diffeomorphic methods, known generally as Voxel-Based Morphometry (VBM), (see SPM, www.fil.ion.ucl.ac.uk/spm/) were used to register scans to a common template and calculate Gray Matter (GM) densities at each voxel in the MR scans. We also used Tensor-Based Morphometry (TBM) to calculate maps of longitudinal voxel-wise expansion or contraction over a two year period. Feature selection was performed separately in each set of images by sorting voxels by t-statistic (calculated using training data), and choosing the highest 2000, 5000, 10000,… ,250000 voxels in 8 stages. We used linear, quadratic, and Gaussian kernels: a total of 24 kernels per set, (GM density maps, TBM maps, baseline FDG-PET, FDG-PET at 2-year follow up) for a total of 96 kernels. For Q-matrix we used the Laplacian of covariance between single-kernel α parameters, (recall the motivation from [19] in Section 3,) plus a block-diagonal representing clusters of kernels derived from the same imaging modalities.

5.2 Spatial SVM

Before describing out main experiments, we first return to the Q-SVM model briefly mentioned in Section 3.2. To demonstrate that regularizers in the form of a Q matrix can indeed influence the learned classifier, we performed classification experiments with the Laplacian of the inverse distance between voxels as a Q matrix, and voxel-wise GM density (VBM) as features. Using 10-fold cross-validation with 10 realizations, Q-SVM ’s accuracy was 0.819, compared to the regular SVM’s accuracy of 0.792. These accuracies are significantly different at the α = 0.0005 level under a paired t-test. In Fig 1 we show a comparison of weights trained by a regular SVM (a–b), and those trained by a spatially regularized SVM, (c–d). Note the greater spatial smoothness, and lower incidence of isolated “pockets”.

Figure 1.

Figure 1

Comparison of spatial smoothness of the weights chosen by Q-SVM and SVM with computed gray matter (GM) density maps. Left (a-b): classifier weights given by a standard SVM; Right (c-d): classifier weights given by Q-SVM .

5.3 Multi-modality Alzheimer’s disease (AD) prediction

Next, we performed multi-modality AD prediction experiments using all available kernels. Several different types of imaging modalities are available, each of which highlights a different aspect of disease pathology; MR provides structural information, while FDG-PET assesses hypo-metabolism. Further, we may use several image processing pipelines. Due to the inherent similarities in how the various kernels are derived, there are clear cluster structures / behaviors among the kernels, which we would like to exploit using Q-MKL. We used 10-fold cross-validation with 30 realizations, for a total of 300 folds. Accuracy, sensitivity and specificity were averaged over all folds. For comparison we also examined 1-, 1.5-, and 2-norm MKL. As MKL methods have emerged as the state of the art in this domain [3, 4], and have performed favorably in extensive evaluations against various baselines such as single-kernel methods, and naïve combinations, we therefore focus our analysis on comparison with existing MKL methods. Results are shown in Table 1. Q-MKL had the highest performance overall, reducing the error rate from 12.5% to 11.2%. (Significant at the α = 0.001 level.) Note that the in vivo diagnostic error rate for AD is believed to be near 8–10%, meaning that this improvement is quite significant. The primary benefit of current sparse MKL methods is that they effectively filter out uninformative or noisy kernels, however, the kernels used in these experiments are all derived from clinically relevant neuroimaging data, and are thus highly reliable. Q-MKL’s performance gives some evidence that it is able to combine the kernels in a way which boosts the overall accuracy.

Table 1.

Comparison of Q-MKL & MKL. Bold numerals indicate methods which did not differ from the best at the 0.01 level using a paired t-test. Lap. = “Laplacian”; diag = “Block-diagonal”.

Regularizer Acc. Sens. Spec.
β1-MKL 0.864 0.771 0.931
β1.5-MKL 0.875 0.790 0.936
β2-MKL 0.875 0.789 0.938

Covα 0.884 0.780 0.942
Lap.(Covα) 0.884 0.785 0.955
Lap.(Covα) + diag 0.888 0.786 0.956

Virtual kernel analysis

We next turn to an analysis of the covariance structures found in the data empirically as a concrete demonstration of the type of patterns towards which the Q-MKL regularizer is biasing β. Recall that the eigen-vectors of a Q matrix can show which patterns are encouraged or deterred, in proportion to their eigen-values. In Fig. 2, we compare the Q matrix used in the ADNI experiments, based on the correlations of single-kernel α parameters (a), the 3 least eigenvectors of its graph Laplacian (b–d), and the β vector optimized by Q-MKL . In (a), we can see that while the VBM (first block of 24 kernels) and TBM (second block of kernels) are highly correlated, they appear to be fairly uncorrelated to one another, while the FDG-PET kernels (last 48 kernels) are much more strongly interrelated. Interestingly, the first eigen-vector is almost entirely devoted to two large blocks of kernels – those which come from MRI data, and those which come from FDG-PET data. The positive elements in the off-diagonal encourage sparsity within these two super-blocks of kernels. Somewhat to the contrary, the next two eigen-vecors have negative weights in the region between TBM and VBM kernels, encouraging non-sparsity between these two blocks. In (d) we see that the optimized β discards most TBM kernels, (but not all,) putting the strongest weight on a few VBM kernels, and keeps a wider distribution of the FDG-PET kernels.

6 Conclusion

MKL is an elegant method for aggregating multiple data views, and is being extensively adopted for a variety of problems in machine learning, computer vision, bioinformatics, and neuroimaging. Q-MKL extends this framework to account for and exploit higher order interactions between kernels – derived from supervised, unsupervised, or domain-knowledge driven. This flexibility can impart greater control over how the model utilizes cluster structure among kernels, and effectively encourage cancellation of errors wherever possible. We have presented a convex optimization model to efficiently solve the resultant model, and shown experiments on a challenging problem of identifying Alzheimer’s disease based on multi modal brain imaging data (obtaining statistically significant improvements). Our implementation will be made available within the Shogun toolbox (www.shogun-toolbox.org).

Supplementary Material

supplement

Footnotes

1

Note that this involves a Gaussian assumption, but [24] provides extensions to non-Gauss kernels.

*

Supported by NIH (R01AG040396), (R01AG021155); NSF (RI 1116584), (DMS 09-15240 ARRA), and (CMMI-1131690); Wisconsin Partnership Proposal; UW ADRC; UW ICTR (1UL1RR025011); AFOSR (FA9550-09-1-0098); and NLM (5T15LM007359). The authors would like to thank Maxwell Collins and Sangkyun Lee for many helpful discussions.

References

  • [1].Guyon I, Elisseeff A. An introduction to variable and feature selection. JMLR. 2003;3:1157–1182. [Google Scholar]
  • [2].Gehler PV, Nowozin S. Let the kernel figure it out; principled learning of pre-processing for kernel classifiers. CVPR. 2009 [Google Scholar]
  • [3].Hinrichs C, Singh V, Xu G, Johnson SC. Predictive markers for AD in a multi-modality framework: An analysis of MCI progression in the ADNI population. Neuroimage. 2011;55(2):574–589. doi: 10.1016/j.neuroimage.2010.10.081. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Zhang D, Wang Y, Zhou L, Yuan H, Shen D. Multimodal Classification of Alzheimer’s Disease and Mild Cognitive Impairment. NeuroImage. 2011;55(3):856–867. doi: 10.1016/j.neuroimage.2011.01.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [5].Lanckriet GRG, Cristianini N, Bartlett P, El Ghaoui L, Jordan M. Learning the kernel matrix with semidefinite programming. JMLR. 2004;5:27–72. [Google Scholar]
  • [6].Sonnenburg S, Rätsch G, Schäfer C, Schölkopf B. Large scale multiple kernel learning. JMLR. 2006;7:1531–1565. [Google Scholar]
  • [7].Rakotomamonjy A, Bach F, Canu S, Grandvalet Y. SimpleMKL. JMLR. 2008;9:2491–2521. [Google Scholar]
  • [8].Gehler PV, Nowozin S. On feature combination for multiclass object classification. ICCV. 2009 [Google Scholar]
  • [9].Yang J, Li Y, Tian Y, Duan L, Gao W. Group-sensitive multiple kernel learning for object categorization. ICCV. 2009 doi: 10.1109/TIP.2012.2183139. [DOI] [PubMed] [Google Scholar]
  • [10].Vemuri P, Gunter JL, Senjem ML, Whitwell JL, Kantarci K, Knopman DS, et al. Alzheimer’s disease diagnosis in individual subjects using structural MR images: validation studies. Neuroimage. 2008;39(3):1186–1197. doi: 10.1016/j.neuroimage.2007.09.073. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [11].Szafranski M, Grandvalet Y, Rakotomamonjy A. Composite kernel learning. Machine learning. 2010;79(1):73–103. [Google Scholar]
  • [12].Bach FR, Lanckriet G, Jordan MI. Multiple kernel learning, conic duality, and the SMO algorithm. ICML. 2004 [Google Scholar]
  • [13].Orabona F, Jie L, Caputo B. Online-Batch Strongly Convex Multi Kernel Learning. CVPR. 2010 [Google Scholar]
  • [14].Kloft M, Brefeld U, Sonnenburg S, Zien A. lp-Norm Multiple Kernel Learning. JMLR. 2011;12:953–997. [Google Scholar]
  • [15].Ong CS, Smola A, Williamson B. Learning the kernel with hyperkernels. JMLR. 2005;6:1045–1071. [Google Scholar]
  • [16].Mukherjee L, Singh V, Peng J, Hinrichs C. Learning Kernels for variants of Normalized Cuts: Convex Relaxations and Applications. CVPR. 2010 doi: 10.1109/CVPR.2010.5540076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Gehler PV, Nowozin S. Infinite kernel learning. Technical Report 178, Max-Planck Institute for Biological Cybernetics. 2008;10 [Google Scholar]
  • [18].Bach FR. Exploring large feature spaces with hierarchical multiple kernel learning. NIPS. 2008 [Google Scholar]
  • [19].Joachims T, Cristianini N, Shawe-Taylor J. Composite kernels for hypertext categorisation. ICML. 2001 [Google Scholar]
  • [20].Renyi A. On measures of entropy and information. Fourth Berkeley Symposium on Mathematical Statistics and Probability. 1961:547–561. [Google Scholar]
  • [21].Cortes C, Mohri M, Rostamizadeh A. Generalization bounds for learning kernels. ICML. 2010 [Google Scholar]
  • [22].Jenssen R. Kernel entropy component analysis. IEEE Trans. PAMI. 2009:847–860. doi: 10.1109/TPAMI.2009.100. [DOI] [PubMed] [Google Scholar]
  • [23].Girolami M. Orthogonal series density estimation and the kernel eigenvalue problem. Neural Computation. 2002;14(3):669–688. doi: 10.1162/089976602317250942. [DOI] [PubMed] [Google Scholar]
  • [24].Erdogmus D, Principe JC. Generalized information potential criterion for adaptive system training. IEEE Transaction on Neural Networks. 2002;13(5):1035–1044. doi: 10.1109/TNN.2002.1031936. [DOI] [PubMed] [Google Scholar]
  • [25].Kowalski M, Szafranski M, Ralaivola L. Multiple indefinite kernel learning with mixed norm regularization. ICML. ACM. 2009 [Google Scholar]
  • [26].Xiang Z, Xi Y, Hasson U, Ramadge P. Boosting with spatial regularization. NIPS. 2009 [Google Scholar]
  • [27].Bergsma S, Lin D, Schuurmans D. Improved Natural Language Learning via Variance-Regularization Support Vector Machines. Conference on Natural Language Learning. 2010 [Google Scholar]
  • [28].Cuingnet R, Chupin M, Benali H, Colliot O. Spatial and anatomical regularization of svm for brain image analysis. NIPS. 2010 [Google Scholar]
  • [29].Shivaswamy P, Jebara T. Maximum relative margin and data-dependent regularization. JMLR. 2010;11:747–788. [Google Scholar]
  • [30].Gai K, Chen G, Zhang C. Learning kernels with radiuses of minimum enclosing balls. NIPS. 2010 [Google Scholar]
  • [31].Mueller SG, Weiner MW, et al. Ways toward an early diagnosis in Alzheimers disease: The ADNI. J. of the Alzheimer’s Association. 2005;1(1):55–66. doi: 10.1016/j.jalz.2005.06.003. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

supplement

RESOURCES