Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2014 Oct 16.
Published in final edited form as: Int Workshop Pattern Recognit Neuroimaging. 2013;2013:13–16. doi: 10.1109/PRNI.2013.13

Deriving statistical significance maps for support vector regression using medical imaging data

Bilwaj Gaonkar 1, Aristeidis Sotiras 1, Christos Davatzikos 1
PMCID: PMC4199337  NIHMSID: NIHMS504671  PMID: 25328907

Abstract

Regression analysis involves predicting a continuous variable using imaging data. The Support Vector Regression (SVR) algorithm has previously been used in addressing regression analysis in neuroimaging. However, identifying the regions of the image that the SVR uses to model the dependence of a target variable remains an open problem. It is an important issue when one wants to biologically interpret the meaning of a pattern that predicts the variable(s) of interest, and therefore to understand normal or pathological process. One possible approach to the identification of these regions is the use of permutation testing. Permutation testing involves 1) generation of a large set of ‘null SVR models’ using randomly permuted sets of target variables, and 2) comparison of the SVR model trained using the original labels to the set of null models. These permutation tests often require prohibitively long computational time. Recent work in support vector classification shows that it is possible to analytically approximate the results of permutation testing in medical image analysis. We propose an analogous approach to approximate permutation testing based analysis for support vector regression with medical imaging data. In this paper we present 1) the theory behind our approximation, and 2) experimental results using two real datasets.

Keywords: Permutation testing, Support Vector Regression

I. Introduction

Regression analysis involves prediction of continuous clinical variables using medical images [1], [2], [3], [4], [5]. Multivariate pattern analysis (MVPA) techniques such as SVR directly address the image based regression paradigm. Most MVPA algorithms including SVR train a model by observing image data with known target variables. Target variables associated with a hitherto unseen test image can be estimated using the trained model.

The SVR algorithm offers predictions of continuous clinical variables from images. However, it provides no direct mechanism to assess which image regions are most significant in predicting the target variables. This question is relevant in clinical studies and is crucial to the clinicians who want to biologically understand imaging patterns and form new hypotheses. Traditionally, mass univariate Voxel Based Analysis (VBA) is used to find regions associated with continuous clinical variables. Such analysis associates a statistical significance test with every voxel in the image by regressing the voxel intensity directly with the target variable. While this provides ease of interpretability, such analysis (unlike MVPA) will miss multivariate associations in data. This motivates the need for a multivariate alternative to VBA that can interpret the model trained by an MVPA method such as a SVR. In the pattern classification paradigm, permutation tests using support vector classifiers (SVC) provide a multivariate alternative to VBA. We present an extension of this permutation testing procedure to the regression paradigm using SVRs.

A major problem with SVR/SVC based permutation testing applied to medical imaging data is the computational time and resources required for the actual implementation of these tests. However, recent work [6] showed that an analytical short cut exists for SVC based permutation testing that reduces the time and resource requirements by several orders of magnitude. This paper describes the theory behind such an analytical approximation that applies in case of SVR based permutation testing.

The remainder of the paper is organized as follows: in Section 2 the intuition behind permutation testing for regression analysis is presented. Following, we detail the theory behind the analytical approximation of permutation testing. Section 3 presents the experimental results on two brain imaging datasets. The paper concludes in Section 4 with a discussion.

II. Method

A. Support Vector Regression: Background

Let us first explain how the SVR [7] algorithm is used in the context of predicting continuous clinical variables from images.

1) Training

In order to train an SVR, we stack preprocessed training image data into a matrix XRm×p whose rows xi index individuals in the population, and columns index image voxels. A continuous target variable yiR is associated with every xi in the training dataset. Then, the -SVR solves the following optimization problem:

w,b=minw,b12w2s.t.wTxi+byi,yiwTxib,i{1,,m} (1)

The solution fits a tube of width to the data [8]. When the number of samples is higher than the number of samples (n > p), it is not always possible to find a tube of width that contains all the data. In the medical image analysis setting, the dimensionality is always much greater than the sample size (p > n). Hence, it is always possible to fit a p-dimensional -tube through all the datapoints.

2) Testing

The SVR model is encoded by the pair { w*, b*}. For a new test subject whose vectorized image is represented by xtest, the prediction ytest made by the SVR algorithm is ytest = w*Txtest + b*.

B. Permutation testing for support vector regression

The dimensionality of the model vector w*, trained by the SVR, is the number of voxels in the image. Thus, every component of the vector w* can be mapped to a voxel in the image domain. This mapping associates an image with the SVR model. Henceforth, we call this image a w-map. It would be desirable to directly use this image for making inferences about which regions are significantly involved in making predictions. However, these weights: 1) can be biased to be large by the simple scaling/translation operations on the underlying voxel intensities; 2) provide no measure of statistical significance of a specific feature/voxel in the image. Thus, a more rigorous method for interpreting the SVR model is required.

Permutation testing is one such method. The concept of permutation testing for SVRs in 2D space is illustrated in Fig. 1. In permutation testing, the target variables yi are permuted randomly. For each random permutation, an SVR is used to compute wrp. After many thousands of permutations, we can generate an approximation to the null distribution of every component of wjDnullj where j ∈ {1,…,p}. Finally, the original labels are used to train w*. Comparing the components wj with Dnullj gives us a p-value associated with every voxel. It is important to note that the null distribution at any voxel depends on the null distribution at all other voxels. This dependence is also true for the components of w themselves. Hence, each component-wise test is based on data from all image voxels and is not univariate in the VBA sense. Furthermore, this interdependence has the potential to alleviate multiple comparisons problems associated with VBA.

Figure 1.

Figure 1

Concept of permutation testing in support vector regression. Comparison of w* to the null distribution generated by {w(1)null, …, w(k)null} is used for inference.

C. The analytical approximation of permutation testing

The main problem with the procedure detailed above is that it requires multiple runs of the SVR algorithm to approximate the underlying null distribution. This results in high computational demands. Massively parallel cluster computing is often used to perform these tests. In comparison to this, a typical run of VBA finishes in a few seconds on a typical computer. To close the gap, we propose an analytical approximation to SVR based permutation testing which runs in time comparable to VBA analysis while producing results that are comparable to empirical permutation testing.

The fundamental assumption behind the analytical approximation is that in high dimension, low sample size data, for most random permutations, the vast majority of the samples lie at the edges of the tube and are thus Support Vectors. This assumption is motivated by a similar assumption made in [6] with respect to support vector classification. Observations with real data confirm this phenomenon (Fig. 2). This assumption does not typically hold for the model trained with the actual targets. This is because there is enough structure in the data to learn from it. However, since most permutations are random the only way the algorithm can find a tube compatible with the entire dataset is by storing all of the data and its labels as support vectors. Under this assumption, for most permutations, the solution to (1) can be approximated by the solution to:

w,b=minw,b12w2wTxi+byi=ORyiwTxib=,i{1,,m} (2)

Figure 2.

Figure 2

Most samples are support vectors for most permutations for SVRs. Human dataset (left) and mouse dataset (right).

Now note that one of the two constraints has to hold for every sample for every permutation. For a permutation, a sample can either adhere to one constraint or another. Thus, for a particular permutation the optimization given by (2) can be solved using the Lagrange multiplier theory to yield exactly as it was done in [6].

L(w,b)=w22+λT((Xw+Jb)(y+L))

where the constraint vector LRm has components Li = ± and and the vector JRm with Ji = +1. Note that the constraint vector for one permutation will differ from that of the next based on which exact components are positive or negative. Setting wL(w,b)=0 and λL(w,b)=0 and solving for w yields:

w=C(y+L), (3)

where C denotes the matrix:

CXT(XXT)1+XT(XXT)1J(JT(XXT)1J)1JT(XXT)1.

However, every permutation is associated with it’s own vector L. Over a large number of permutations, we can expect either constraint in (2) to hold with equal probability for each sample. Thus we may write: P(Li = +) = 1/2, P(Li = −) = 1/2. Note that (3) can also be written in its component form as:

wj=i=1mCij(yi+Li). (4)

Because every element Cij is fully determined by the data matrix X, we can treat them as constants. By taking expectations on both sides of (4), we obtain:

E(wj)i=1mCijE(yi+Li)=E(yi)i=1mCij.

Note that E(Li) = 0 and that E(yi) does not change with i allowing us to pull it outside the summation sign. To explicitly acknowledge this invariance, we henceforth denote E(yi) simply as E(y). Similarly, the variance of wj can be predicted by taking variances on both sides:

Var(wj)=i=1mCij2(Var(yi)+Var(Li))=(Var(yi)+2)i=1mCij2.

Note again that the term V ar(yi) + 2 is invariant with respect to i. Henceforth, we simply denote this term as V ar(y) + 2. Thus, we write:

E(wj)=E(y)i=1mCijVar(wj)=(Var(y)+2)i=1mCij2.

Regarding the distribution of wj, it can be shown to be normal using the Lyapunov Central Limit Theorem (CLT). To see this, define zij=Cij(yi+Li). The variable zij is linearly dependent on yi + Li. We can infer the expectation and variance of zij from yj as:

E(zij)=CijE(y)Var(zij)=Cij2(Var(y)+2).

Note that zij are independent but not identically distributed, and wj are linear combinations of zij. Then, according to the Lyapunov CLT, wj is distributed normally if:

limm1[i=1mVar(zij)]2+δk=1mE[zkjμk2+δ]=0,δ>0. (5)

For δ = 1, we have:

E[zkjμk2+δ]=E[CkjykCkjE(yk)2+δ]=Ckj3E[ykE(yk)3].

Again we note that E [|ykE(yk)|3] is independent of k and henceforth denote it simply as E [|yE(y)|3]. Then, we can write the limit in (5) as:

limmE[yE(y)3]k=1mCkj3[(Var(y)+2)i=1mCij2]3=Kk=1m(limmCkj2i=1mCij2)3=0, (6)

where K is a constant independent of the sample indices k and i, defined as: K=E[yE(y)3][(Var(y)+2)]3. Because (6) will tend to zero in the limit, we have normality of wj by the Lyapunov CLT.

III. Experiments and Results

In order to validate the theory proposed above, we performed two experiments using imaging data. In the following, we discuss these experiments.

Human brain data

For this experiment, we used a dataset of 132 T1 images corresponding to normal subjects of age between 10 and 20 years. The experiment was done using Grey matter (GM), white matter (WM) and ventricular (CSF) tissue density maps (TDMs) that were generated after preprocessing of the raw images. TDMs convey information about the quantity of tissue present at each brain location in a common template space.

TDMs corresponding to the ith subject were vectorized, and the vectors of all 3 tissue types were concatenated into the vector xiR1×3q. The vectors of various samples were then stacked together to form the matrix X. Permutation testing was performed using 1000 permutations of labels. The null distribution, obtained using permutation tests, was compared to the model that was trained with the original labels, to obtain an experimental p-map. Similarly, the analytical null distributions predicted using the theory presented in Section 2 were used to generate an analytical p-map. The two p-maps are compared in Fig. 3a.

(a).

(a)

Representative slices of analytic and experimental p-maps for grey matter TDMs (left) and scatter plot of corresponding analytic and experimental p-values (for all three tissue types) for human brain data.

It is easy to note by visual inspection that the significant map obtained with the analytical approximate methods agrees with the one obtained through permutation testing. The main difference is that the proposed method required significantly less computational time than permutation testing to produce a result of equivalent quality. To gain a quantitative view of the level of agreement of the two solutions, Fig. 3a also shows the scatter plot between the experimental and analytic p-maps.

Developing mouse brain data

In this experiment, we applied the proposed method to the problem of white matter maturation in mouse brains. We used ex vivo acquired Diffusion Tensor images of a population of 79 inbred mice of C57BL/6J strain. The imaged mouse correspond to different postnatal stages, ranging from day 2 to day 80 [9]. Early developmental stages were sampled more densely because development is more emphasized during that period.

The images were deformably registered to a template image chosen from the age group of day 10 using DROID [10]. DTI-Studio [11] was used to estimate tensors from which, the Fractional Anisotropy was calculated resulting in images with dimension 300 × 300 × 200.

Similarly to the previous experiment, we compare the experimental p-map with the analytic one. By visually comparing correspond slices from the two mpas, we note that the predicted values closely follow the actual ones Fig. 3b. We also observe distinctively low p-values in the cortex and the genu of corpus callossum. These areas have been previously reported exhibiting noteworthy maturation profiles [9]. The scatter plot suggests that analytic and experimental p-values agree.

(b).

(b)

Representative slices of analytic and experimental p-maps (left) and scatter plot of corresponding analytic and experimental p-values for mouse brain data.

IV. Discussion

In this paper, we have provided the theoretical framework for analytically approximating permutation tests using SVRs. We have also provided a limited validation of this framework using two real datasets.

References

  • [1].Stonnington CM, Chu C, Klppel S, J. CR, Jr., Ashburner J, Frackowiak RS. Predicting clinical scores from magnetic resonance scans in alzheimer’s disease. NeuroImage. 2010;51(no. 4):1405–1413. doi: 10.1016/j.neuroimage.2010.03.051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Formisano E, De Martino F, Valente G. Multivariate analysis of fMRI time series: classification and regression of brain responses using machine learning. Magnetic Resonance Imaging. 2008 Sep;26(no. 7):921–934. doi: 10.1016/j.mri.2008.01.052. [DOI] [PubMed] [Google Scholar]
  • [3].Ashburner J. A Fast Diffeomorphic Image Registration Algorithm. NeuroImage. 2007;38(no. 1):95–113. doi: 10.1016/j.neuroimage.2007.07.007. [DOI] [PubMed] [Google Scholar]
  • [4].Zhang D, Shen D, A. D. N. Initiative Predicting future clinical changes of mci patients using longitudinal and multimodal biomarkers. PLoS ONE. 2012;7(no. 3):e33182. doi: 10.1371/journal.pone.0033182. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [5].Wang Y, Fan Y, Bhatt P, Davatzikos C. High-dimensional pattern regression using machine learning: From medical images to continuous clinical variables. NeuroImage. 2010;50(no. 4):1519–1535. doi: 10.1016/j.neuroimage.2009.12.092. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Gaonkar B, Davatzikos C. Deriving statistical significance maps for svm based image classification and group comparisons. MICCAI. 2012:723–730. doi: 10.1007/978-3-642-33415-3_89. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [7].Vapnik VN. The nature of statistical learning theory. Springer-Verlag New York, Inc.; New York, NY, USA: 1995. [Google Scholar]
  • [8].Smola AJ, Schölkopf B. A tutorial on support vector regression. Statistics and Computing. 2004;14(no. 3):199–222. [Google Scholar]
  • [9].Verma R, Mori S, Shen D, Yarowsky P, Zhang J, Davatzikos C. Spatiotemporal maturation patterns of murine brain quantified by diffusion tensor MRI and deformation-based morphometry. PNAS. 2005;102(no. 19):6978–83. doi: 10.1073/pnas.0407828102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Ingalhalikar M, Yang J, Davatzikos C, Verma R. Dtidroid: Diffusion tensor imaging-deformable registration using orientation and intensity descriptors. International Journal of Imaging Systems and Technology. 2010;20(no. 2):99–107. [Google Scholar]
  • [11].Jiang H, van Zijl PC, Kim J, Pearlson GD, Mori S. Dtistudio: Resource program for diffusion tensor computation and fiber bundle tracking. Computer Methods and Programs in Biomedicine. 2006;81(no. 2):106–116. doi: 10.1016/j.cmpb.2005.08.004. [DOI] [PubMed] [Google Scholar]

RESOURCES