Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2022 Feb 3.
Published in final edited form as: Inf Process Med Imaging. 2019 May 22;11492:99–111. doi: 10.1007/978-3-030-20351-1_8

On Training Deep 3D CNN Models with Dependent Samples in Neuroimaging

Yunyang Xiong 1, Hyunwoo J Kim 2, Bhargav Tangirala 1, Ronak Mehta 1, Sterling C Johnson 1, Vikas Singh 1
PMCID: PMC8813050  NIHMSID: NIHMS1053000  PMID: 35125834

Abstract

There is much interest in developing algorithms based on 3D convolutional neural networks (CNNs) for performing regression and classification with brain imaging data and more generally, with biomedical imaging data. A standard assumption in learning is that the training samples are independently drawn from the underlying distribution. In computer vision, where we have millions of training examples, this assumption is violated but the empirical performance may remain satisfactory. But in many biomedical studies with just a few hundred training examples, one often has multiple samples per participant and/or data may be curated by pooling datasets from a few different institutions. Here, the violation of the independent samples assumption turns out to be more significant, especially in small-to-medium sized datasets. Motivated by this need, we show how 3D CNNs can be modified to deal with dependent samples. We show that even with standard 3D CNNs, there is value in augmenting the network to exploit information regarding dependent samples. We present empirical results for predicting cognitive trajectories (slope and intercept) from morphometric change images derived from multiple time points. With terms which encode dependency between samples in the model, we get consistent improvements over a strong baseline which ignores such knowledge.

1. Introduction

The use of machine learning methods to perform classification or regression tasks with brain imaging data, in the last decade or so, has demonstrated that the clinical manifestation of various diseases such as Alzheimer’s disease (AD) and Parkinson’s disease (PD) can be reliably identified and predictive features obtained from such models are consistent with scientific hypotheses based on post-mortem studies [8]. In the last five years, following the broad adoption of convolutional neural networks for learning tasks in vision, we have seen the deployment of such architectures for brain imaging data as well [7]. While the diagnosis of various diseases using brain images remains important within brain image analysis, many scientific questions such as understanding brain development, evaluating normal aging trends or assessing the course of disease progression requires moving beyond cross-sectional analysis [17]. In other words, instead of one scan per participant, answering a number of key questions involves studies that need multiple scans per person. While sometimes, the setting may be longitudinal (gap between scans is 1–2 years), in other cases, the temporal relation between the scans may not be important. What is relevant, however, is that when each subject in a study provides more than one scan (or sample), the samples are no longer i.i.d. – instead, scans from each person are dependent.

Consider a simple regression task using brain imaging data, motivated by applications in Alzheimer’s disease (AD) research. Let us assume that structural magnetic resonance (MR) images are the predictor variables X. In AD research, cognitive trends are an important outcome variable. So, the slope and intercept (calculated over multiple years) of an appropriate measure of cognition, say the Auditory Verbal Learning Test (AVLT) [14], can be the response variable Y. This is a simple regression model which seeks to estimate f:xRpyR2, where x denotes the brain image and y is the multi-valued response: AVLT cognition intercept together with its slope over multiple years. The task of estimating the parameters that map the independent variable xX to the dependent variables yY can be approached via deep methods. If one were to use recent deep learning advances, one choice could be convolutional neural networks.

Notice that most regression models in vision and machine learning, including standard CNNs, assume that the samples are independent and identically distributed (i.i.d.). This usually allows a relatively simple model derivation that does not need to take the dependency between the samples into account. We know how to analyze the behavior of such models and tractable inference schemes are also available. Under the i.i.d assumption, standard regression techniques have one model for a dataset which maps all the samples in the same way. These models are called fixed effects models [18]. For instance, a linear regression as well as non-linear (or deep) regression models involve a set of fixed coefficients for the full set of examples. However, in the applications with dependent measurements, the i.i.d. assumption does not hold and fixed effects models are often suboptimal. For instance, longitudinal data or any study with multiple samples for a subject involve data that are dependent and affected by certain subject-specific terms. This is called a random effect. In such situations, to accurately model global trends, the random effects should be properly utilized.

Dependent samples in regression are modeled using mixed effects models [18]. Mixed effects models handle the fixed effects and random effects by explicitly leveraging the “group” (dependent samples) information to handle the inherent non-independence in the data. Many brain imaging studies have repeated measurements and non-independence with a hierarchical structure of data. We know that CNNs have been shown to be highly effective in a variety of medical imaging problems [2]. But CNNs for non-i.i.d samples are far less explored. This paper studies how CNNs can be modified to incorporate terms that account for dependent samples, motivated by neuroimaging applications.

The main goal of our work is to use random effects within 3D convolutional neural networks, and use this 3D CNN model to predict cognitive performance on multiple brain image scans per subject with visual interpretation. Our main contributions are: 1) We use random effects within 3D convolutional neural networks (RE-CNN) built on inverted residual blocks. 2) We show that our proposed RE-CNN works well for brain imaging datasets with dependent samples and outperforms a strong baseline (which does not incorporate random effects) by significant margins. 3) We provide a module for visual interpretation of voxels that are relevant for the prediction task. Experiments suggest that regions identified are consistent with existing hypotheses (i.e., cognitive performance affected by gross brain atrophy).

Related work:

Various approaches have been proposed to analyze repeated measurements. A model with RE-EM tree was presented in [15] to combine mixed effects models for longitudinal and clustered data with the flexibility of tree-based estimation methods. In [6], mixed-effects random forest was developed to analyze clustered data consisting of individuals nested within groups. Riemannian nonlinear mixed effects models are developed in [9] for longitudinal brain data analysis. In [1], non-linear mixed effects statistical model was proposed to estimate alteration of the shape of the hippocampus during the course of Alzheimer’s disease. Separately, visual interpretation methods are developed to help understand and indicate the spatial attention of convolutional neural network when making predictions. [20] shows that the visualization genuinely corresponds to the image structure that stimulates that feature map by performing occlusion sensitivity. [16] proposed the visualization of deep convolutional neural network based on computing the gradient of the class score with respect to the input image. In [5], a perturbation scheme was developed to find where highly complicated neural networks look in an image for evidence. Class Activation Mapping (CAM) in [21] is developed to visualize the linear combination of a late layer’s activation and class-specific weights.

2. Preliminaries

In this section, we briefly review the basic concepts and properties of linear mixed-effects models and nonlinear mixed-effects model to setup the rest of our presentation for AVLT cognition prediction. For simplicity, we introduce models with a univariate response variable (also called labels or dependent variable).

Fixed effect models can be considered as a linear or non-linear global model. An example for fixed effect models is a standard linear regression model,

y=xTβ+ϵ,ϵ~N(0,Σ) (1)

where yR is the response, x=[1,x1,x2,,xp]TRp+1 are the covariates, β=[β0,,βp]TRp+1 are fixed effects coefficients shared over all subjects.

Mixed effect models explicitly incorporates subject information by introducing a set of parameters for each subject (or group) to yield subject-specific adjustments. Recall that AVLT cognition prediction for AD research is a function fitting problem, mapping a single input brain image to AVLT cognition. We observe from (1) that all subjects have the exactly same estimation function to map an input 3D brain image to AVLT cognition output and the noise permitted in the estimation function also comes from a distribution identical to every subject. However, with multiple measurements from each subject, these 3D brain image scans are not independent and each subject may have a slightly different estimation function. Therefore, our goal is to use fixed effects for a global model but also incorporate random effects for flexible specific-subject adjustments, yielding slightly different mapping function of AVLT cognition prediction. By adding random effects to (1) for each participant, the linear mixed effects (LME) model can be written as,

y=xTβ+zTbi+ϵi,bi~N(0,D)andϵi~N(0,Σϵi),i=1,2,,N (2)

where N is the number of subjects, β=[β0,,βp]Rp+1 are the fixed effects (population parameters) shared over the entire population and bi=[b0i,,bqi]Rq+1 are the subject-specific random effects (individual parameters) for ith subject for subject-specific adjustment, x and z are the associated fixed effects and random effects covariates, D and Σϵi,i=1,2,,N, are the variance components of linear mixed effects model, ϵi is the measurement error, drawn from a subject-specific unknown zero mean Gaussian distribution with covariance structure, N(0,Σϵi), for dealing with the non-i.i.d. nature. When covariance matrices, D and Σϵi, are unknown, no closed form solution for estimation of fixed and random effects is available and EM algorithms are proposed for estimation [18].

Nonlinear mixed effects models (NLME) [18] is a generalization of a LME model. Given a nonlinear link function η, the NLME model is,

y=η(x,β,z,bi)+ϵi,bi~N(0,D),ϵi~N(0,Σϵi),i=1,2,,N (3)

where β are fixed effects vector and bi are random effects vector for ith subject.

3. Random effects 3D Convolutional Neural Networks

Linear mixed effects model assumes a linear parametric form, and sometimes, this assumption might be too restrictive. Assuming a linear model may not be the best option when the parametric form is unknown. Further, when the number of covariates are large (e.g., 3D brain image), including all of the attributes may lead to overfitting and poor predictions. Therefore, we generalize linear mixed effects model to nonlinear mixed effects model using convolutional neural networks: we will learn the mapping from the input 3D brain image to AVLT cognition intercept and slope while being cognizant to individual level differences of the subjects, denoted as random effects 3D convolutional neural network (RE-CNN). The cognition slope and intercept may be provided for each sample, or may be the same for all samples for a participant. To specify nonlinear mixed effects models, we write the model as a sum of fixed and random components,

η=ffix(x,β)+frnd(z,bi) (4)

where ffix is a nonlinear link function for fixed effects and frandom is a nonlinear link function for random effects. Generalized from the random intercept and random slope model [4], individually varying subject-specific intercept and slope, we define our nonlinear mixed effects model as

η=h(x)β+h(x)bi (5)

where h is a nonlinear link function, h(x)β represents the fixed component and h(x)bi represents the random component. Here, (5) can be considered in the following way: we apply a nonlinear transformation h(·) first and then learn the fixed effects and random effects on top of the output features from h.

3.1. Architecture of RE-CNN

Our model is based on the efficient MobileNetV2 architecture and a sequence of inverted residual blocks, to represent nonlinear transformation h. To predict AVLT cognition prediction from 3D brain images, our model uses an adapted 3D MobileNetV2 architecture (see Fig 1) for h, followed by a linear mixed regression loss after average pooling (Avg Pool).

Fig. 1:

Fig. 1:

An overview of 3D convolutional neural network structure incorporating linear mixed effects model, RE-CNN. InvResBlock denotes adapted 3D inverted residual block from 2D MobileNetV2 and it can be seen in Fig 2. InvResBlock ×1/2/3/4 means the number of repeated layers within InvResBlock. The output of average pooling is denoted as h(x), which is the input for the mixed effects models, ffix(x) = h(x)β for fixed effects estimation and frandom = h(x)bi is for random effects estimation. The output of the network is a 2D vector, the sum of fixed effects estimation and random effects estimation.

RE-CNN can be viewed as a linear mixed effects regression for AVLT cognition prediction on the average pooling layer, h(·), output from 3D MobileNetV2. In Fig 1, we can see that the top network learns the nonlinear fixed effects component, h(·)β, and the bottom network learns the nonlinear random component, h(·)bi. Therefore, learning the nonlinear mixed effects model (see (4)) represented by RE-CNN (see Fig 1) can be considered as an optimization where we train the network with the loss,

minffix,frndxloss(y,ffix(x)+frnd(x))=xyffix(x)frnd(x)2, (6)

where ffix(x)=h(x)βandfrnd(x)=h(x)bi

3.2. Parameter estimation of RE-CNN

Given N subjects, each subject with ni(i=1Nni=n) observations, our goal is to estimate the parameters of our nonlinear mixed effects regression model.

yi=h(Xi)β+h(Xi)bi+ϵi,bi~N(0,D)andϵi~N(0,Σϵi),i=1,2,,N (7)

where yi=[yi1,,yini]T is a ni × 1 vector of observations of the ith subject, Xi=[xi1,,xini]T is a ni × p covariates matrix, h(Xi)=[h(xi1),,h(xini)]T, and ϵi=[ϵ(i1),,ϵ(ini)]T is a ni × 1 measurements error vector.

When the variance components of (7), D and Σϵi,i=1,2,,N, and the nonlinear transformation, h, are known, the estimates of fixed effects parameters β and random effects parameters bi, i = 1,2,...,N can be solved by minimizing the following generalized log-likelihood loss function [18],

loss(β,bi|y)=i=1N[(yih(Xi)βh(Xi)bi)TΣϵi1(yih(Xi)βh(Xi)bi)+biTD1bi+log|D|+log|Σϵi|] (8)

The minimization of the loss in (8) can be performed by solving the so-called mixed model equation [13] with known D, Σϵi and h,

[h(X)TΣϵ1h(X)h(X)TΣϵ1h(X)h(X)TΣϵ1h(X)h(X)TΣϵ1h(X)+Dc1][βb]=[h(X)TΣϵ1yh(X)TΣϵ1y]. (9)

where y=[y1T,,yNT]T, h(X)=[h(X1)T,,h(XN)T]T, b=[b1T,,bNT]T, Dc=diag(D,,D),Σϵ=diag(Σϵ1,,ΣϵN).

Since the covariance matrices, D and Σϵi,i=1,,N, are unknown, we need to estimate them for solving (9). Similar to [18], we maximize the following log-likelihood function for estimating these matrices,

logL(Dc,Σϵ|y)=12[nlog2πlog|V|(yh(X)β)TV1(yh(X)β)] (10)

where Vi=h(Xi)Dh(Xi)T+Σϵi,i=1,,NandV=diag(V1,,VN).

Based on (9) and (10), the last step in parameter estimation for our model (7) is to estimate the nonlinear transformation h, represented by a 3D MobileNetV2 (see Fig. 1). Stochastic gradient algorithm (SGD) is a popular scheme to optimize neural networks. We thus directly apply SGD for training h.

When the covariance matrices are replaced by the current available values, their updated values can be obtained by (9) and (10). Therefore, we use an EM algorithm for parameter estimation for our RE-CNN, denoted as RE-EM. Assume Σϵi=σ2Ini,i=1,,N. Let t index the number of iterations and β^{t},b^i{t},h{t},σ^2{t} and D^{t} be the estimates of β, bi, h σ2 and D at tth iteration. The major update equations in our RE-EM procedure are,

  1. Solving (9),
    b^i{t}=D^{t1}h^{t}(Xi)TV^i1{t1}(yih^{t}(Xi)β^{t})ϵ^i{t}=yih{t}(Xi)β^{t}h{t}(Xi)b^i{t} (11)
    where V^i{t1}=h{t}(Xi)D^{t1}h{t}(Xi)T+σ^2{t1}Ini.
  2. Solving (10),
    σ^2{t}=1ni=1N[ϵ^iT{t}ϵ^i{t}+σ^2{t1}(niσ^2{t1}trace(V^i{t1}))]D^{t}=1Ni=1N[b^i{t}b^iT{t}+(D^{t1}D^{t1}h^{t}(Xi)TV^i1{t1}h^{t}(Xi)D^{t1})] (12)

Then, the RE-EM algorithm for our regression task can be written formally in Algorithm 1. When the training of Algorithm 1 has been completed, we can fix h and learn a fully connected layer, g, based on the hidden representation of h to approximate f^rnd(x) only based on input 3D brain image without knowing the subject “id” for cognition measures estimate (AVLT slope or intercept).

4. Experiments

In this section, we discuss the task of predicting AVLT cognition prediction (intercept and slope) using brain images and evaluate the effectiveness of our RE-CNN model. First, we briefly review a representation that measures local morphometric changes of brains over time for each subject. Using this representation as input, we compare our model with various schemes and evaluate generalization power on unseen subjects. Further, we demonstrate how to interpret the learned model showing activation maps in the image space by a visual explanation approach [21], which allows identifying brain regions that are correlated with cognitive trajectories.

Algorithm 1.

Parameters estimation of RE-CNN for AVLT cognition prediction

0: procedure RE-CNN Training
1: #iter t = 0, Let b^i{0}=0,σ^2{0}=1,D^{0}=Ip, and denote Xi,i=1,,N, the input 3D brain samples for training RE-CNN.
3: Update b^i{t} and ϵ^i{t} using expressions 11.
4: Update variance σ^2{t} and D^{t} using expressions 12.
5: Repeating steps 2, 3, and 4 until convergence.
5: procedure RE-CNN AVLT cognition Prediction
6: For an input 3D brain scan, x, taking the sum of corresponding fixed part RE-CNN AVLT cognition prediction and predicted random part RE-CNN AVLT cognition prediction, h^(x)β^+g^(h^(x))

4.1. RE-CNN on longitudinal CDT images for cognition prediction

Our model uses longitudinal brain images and cognitive scores. In our experiments, morphometric change maps are used as input to the network architecture – these maps are obtained using multiple registration steps.

Morphometric change maps (given by determinant of Cauchy Deformation Tensor) brain images are subject-specific warps derived from a longitudinal neuroimaging study. In this paper, they are derived from a preclinical Alzheimer’s disease (AD) cohort at Wisconsin. The longitudinal warps (or deformations) are with-in subject registrations of T1-weighted images between two consecutive time points, Φi,j:Ii,jIi,j+1. At each voxel, a CDT Det is a determinant of Cauchy-Deformation Tensor derived from the spatial derivatives Φi,j(vox) of the deformation field, i.e., det(Φi,j(vox)TΦi,j(vox)). For downstream analysis, CDT calculation should be spatio-temporally unbiased and registered in a common coordinate system for voxel-wise analysis. Our pipeline is as follows. First, we estimate an unbiased global template space. Following [9], we estimate a temporally unbiased subject-specific average for each subject and then use the averages to generate the global template space. The averages are obtained by Advanced Normalization Tools (ANTS) [3]. Within-subject deformations (longitudinal changes) are transported to the global template space by the parallel transport of stationary velocity fields (SVFs) by Lorenzi and Pennec [11], Specifically, all brain images are registered in the global template space by a rigid transformation and then nonlinear symmetric diffeomorphic deformations are obtained by [10]. Thus brain images registered pairwise between two consecutive time points lead to a SVF, representing a longitudinal morphometric change between visits i + 1 and i. In our cohort, 88 subjects had at least three visits (CDT Dets brain image computed from scans at two time points, see Fig 3) and a total of 202 samples (CDT Det images) are generated. Some subjects had two visits whereas a few had four visits.

Fig. 3:

Fig. 3:

Examples of input brain images and warps in AVLT cognition prediction. (A, C, E) the brain images are collected at three time points for one participant. (B) Warped spatial grid to move (A) to (C), (D) Warped spatial grid to move (C) to (E). Our CDT Det features are derived from the Jacobian of the deformations (C, E).

RE-CNN on longitudinal CDT Dets

We now discuss how to use morphometric change maps (multiple images per subject) to predict cognition slope and intercept, i.e., subjects’ multiple morphometric maps → their cognition intercept and slope. Here xijRH×W×C represents the CDT determinant brain image between two consecutive visits of each subject and yijR2. Given the small sample size, we split the data with 8 subjects/split and perform 11-fold cross validation. Then, Alg. 1 is used for training RE-CNN. Table 1 summarizes the comparisons of our RE-CNN and other baseline methods. Our model achieved the lowest root mean squared error. It shows that even if the subject id (group information) is not available at test time, our method leads to a more accurate fixed effects model by controlling for random effects during training. Our model performs better against a standard 3D CNN architecture for cognition slope and intercept prediction.

Table 1:

Comparison of RE-CNN and other baseline methods for AVLT cognition slope and intercept prediction. Our approach offers the smallest Root Mean Squared Error (RMSE) and the highest Pearson correlation score between AVLT cognition intercept and the predicted cognition intercept > 0.6 and the highest Pearson correlation score between AVLT cognition slope and the predicted cognition slope > 0.6. It means our model predicts more accurately than other baselines. Surprisingly, without any group information at test time, RE-CNN achieves better results than standard 3D CNNs that are not trained with group information. Even only using fixed effects part of RE-CNN, RE-CNNfix performs better than standard 3D CNN. It empirically shows that modeling random effects during training using group information may help learn better fixed effects models and leads to better generalization on unseen subjects.

Method RE-CNN RE-CNNfix 3D CNN GPR[12] SVR
RMSE 2.7 2.9 3.1 4.8 5.2

4.2. Visual Interpretation of RE-CNN on Brain Image

In this subsection, we describe our methods for explaining the predictions of RE-CNN in detail and then visually check heatmaps generated by a visual explanation method. There are two major ways to explain the predictions of 2D deep convolutional neural network. One scheme applies perturbations to data and conducts sensitivity analysis. Another strategy is to use architectural properties of CNNs to heuristically track the attention of convolutional neural networks. We extend 2D visual explanation methods [5,21] to 3D visual explanation.

Sensitivity analysis methods suffer from the problem that each time we only get an output change for one voxel, which makes this extremely computationally intensive for a brain image of size 224 × 224 × 133. To avoid the issues of one-at-a-time sensitivity analysis, we focus on a method based on the architectural properties of our 3D CNN, visualizing the activation map of convolutional layers directly when predictions are made [21,19]. We extend 2D activation mapping to 3D using the idea that the last convolution layer of the CNN contains spatial information indicating discriminative regions for making prediction. Thus, the activation mapping creates a spatial heatmap out of the activations from the last convolutional layer for visualizing discriminative parts. Assume we are given a 3D CNN, e.g., RE-CNN, and one brain image V and denote fu(x,y,z) to be the activation of unit u in the last convolutional layer at location (x, y, z). The global average pooling for unit u is Gu=1Nx,y,zfu(x,y,z), where N is the number of voxels in the corresponding convolutional layer. Notice that our RE-CNN is trained for regressing on AVLT cognition intercept and slope and the global average pooling layer is directly connected to the output layer, we extend the AVLT cognition intercept score as,

Score(INT)=uwuINTGu=1Nx,y,zuwuINTfu(x,y,z) (13)

where uwuINTfu(x,y,z) is defined for every spatial location (x,y,z) and their sum is proportional to the AVLT cognition intercept score. With the intercept score, we define our regression activation mapping (RAM) for AVLT cognition intercept as sRAMx,y,z(INT)=|uwuINTfu(x,y,z)|, where || is the absolute value. The activation mapping for AVLT cognition intercept can be taken as a heatmap of weighted sum of activations in every location (x,y,z) and can be computed by one forward pass given one brain image. Similarly, we can define the regression activation mapping (RAM) for AVLT cognition slope as RAMx,y,z(SLOPE)=|uwuSLOPEfu(x,y,z)|. The regression activation mapping (RAM) for AVLT total cognition can be defined as,

RAMx,y,z(TOT)=|u(wuINT+wuSLOPE)fu(x,y,z)| (14)

Our goal is to obtain visual explanations for highlighting brain regions important for cognition prediction. In our experiments, we take the morphometric change brain images from the test set for visual explanation/interpretation analysis and generate the heatmap from the horizontal, sagittal, and coronal sections. We present the highlighted regions which indicate that overall brain atrophy is associated with cognitive slope and intercept. The regression activation mapping heatmaps of RE-CNN are presented in Figure 4.

Fig. 4:

Fig. 4:

Visual explanation heatmaps of brain images. RE-CNN takes CDT Dets images to associate the morphometric change of brain images with AVLT cognition scores. The regions which have relatively large morphometric change are highlighted in grean in Row 1. The red highlights in Row 2 represent the activation map learned RE-CNN computed by RAM visual approach. Its activation shows the relevant regions for predicting AVLT cognition slope and intercept for each subject.

5. Conclusion

This paper proposes a 3D convolutional neural network that is able to handle dependent brain imaging data (multiple samples per subject) for predicting cognitive measures. By respecting the dependencies, we present a 3D MobileNetV2 based architecture with random effects terms, RE-CNN, to exploit the representation power of 3D CNNs but systematically deal with dependencies in the samples. We develop a RE-EM algorithm for parameter estimation. Experimentally, our RE-CNN achieves promising results. We applied visual explanations to interpret the learned RE-CNN model by assessing regression activation maps, which allows identifying brain regions that are associated with cognition in this observational study focused on better understanding Alzheimer’s disease (AD).

Fig. 2:

Fig. 2:

The 3D inverted residual block is adapted from 2D inverted residual block of MobileNetV2. Similarly, if the output resolution of the block differs from the input resolution, the residual connection of the block is skipped.

Acknowledgments

Supported by UW CPCP AI117924, R01 EB022883 and R01 AG062336. Partial support also provided by R01 AG040396, R01 AG021155, UW ADRC (AG033514), UW ICTR (1UL1RR025011) and NSF CAREER award RI 1252725. We also thank Nagesh Adluru for his help during data processing.

References

  • 1.Bne A, Colliot O, Durrleman S: Learning distributions of shape trajectories from longitudinal datasets: A hierarchical model on a manifold of diffeomorphisms. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (June 2018) [Google Scholar]
  • 2.Çiçek Ö, Abdulkadir A, et al. : 3d u-net: learning dense volumetric segmentation from sparse annotation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. pp. 424–432. Springer; (2016) [Google Scholar]
  • 3.Elhamifar E, Vidal R: Sparse manifold clustering and embedding. In: Advances in neural information processing systems. pp. 55–63 (2011) [Google Scholar]
  • 4.Fahrmeir L, Tutz G: Multivariate statistical modelling based on generalized linear models. Springer Science & Business Media; (2013) [Google Scholar]
  • 5.Fong RC, Vedaldi A: Interpretable explanations of black boxes by meaningful perturbation [Google Scholar]
  • 6.Hajjem A, Bellavance F, Larocque D: Mixed-effects random forest for clustered data. Journal of Statistical Computation and Simulation 84(6), 1313–1328 (2014) [Google Scholar]
  • 7.Ithapu VK, Singh V, Okonkwo O, Johnson SC: Randomized denoising autoencoders for smaller and efficient imaging based ad clinical trials. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. pp. 470–478. Springer; (2014) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Kehagia AA, Barker RA, Robbins TW: Neuropsychological and clinical heterogeneity of cognitive impairment and dementia in patients with parkinson’s disease. The Lancet Neurology 9(12), 1200–1213 (2010) [DOI] [PubMed] [Google Scholar]
  • 9.Kim HJ, Adluru N, et al. : Riemannian nonlinear mixed effects models: Analyzing longitudinal deformations in neuroimaging [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Lorenzi M, Ayache N, et al. : Lcc-demons: a robust and accurate symmetric diffeomorphic registration algorithm. NeuroImage 81, 470–483 (2013) [DOI] [PubMed] [Google Scholar]
  • 11.Lorenzi M, Pennec X: Efficient parallel transport of deformations in time series of images: from schilds to pole ladder. JMIG 50(1–2), 5–17 (2014) [DOI] [PubMed] [Google Scholar]
  • 12.Quiñonero-Candela J, Rasmussen CE: A unifying view of sparse approximate gaussian process regression. JMLR 6(Dec), 1939–1959 (2005) [Google Scholar]
  • 13.Robinson GK: That blup is a good thing: the estimation of random effects. Statistical science pp. 15–32 (1991) [Google Scholar]
  • 14.Schmidt M, et al. : Rey auditory verbal learning test: A handbook. Western Psychological Services; Los Angeles, CA: (1996) [Google Scholar]
  • 15.Sela RJ, Simonoff JS: Re-em trees: a data mining approach for longitudinal and clustered data. Machine learning 86(2), 169–207 (2012) [Google Scholar]
  • 16.Simonyan K, Zisserman A: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014) [Google Scholar]
  • 17.Smyser CD, et al. : Longitudinal analysis of neural network development in preterm infants. Cerebral cortex (2010) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Wu H, Zhang JT: Nonparametric regression methods for longitudinal data analysis: mixed-effects modeling approaches, vol. 515. John Wiley & Sons; (2006) [Google Scholar]
  • 19.Yang C, et al. : Visual explanations from deep 3D CNNs for Alzheimer’s disease classification. arXiv:1803.02544 (2018) [PMC free article] [PubMed] [Google Scholar]
  • 20.Zeiler MD, Fergus R: Visualizing and understanding convolutional networks. In: European conference on computer vision. pp. 818–833. Springer; (2014) [Google Scholar]
  • 21.Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A: Learning deep features for discriminative localization. In: CVPR; (2016) [Google Scholar]

RESOURCES