Abstract
Deep learning is a promising tool that uses nonlinear transformations to extract features from high-dimensional data. Deep learning is challenging in genome-wide association studies (GWAS) with high-dimensional genomic data. Here we propose a novel three-step approach (SWAT-CNN) for identification of genetic variants using deep learning to identify phenotype-related single nucleotide polymorphisms (SNPs) that can be applied to develop accurate disease classification models. In the first step, we divided the whole genome into nonoverlapping fragments of an optimal size and then ran convolutional neural network (CNN) on each fragment to select phenotype-associated fragments. In the second step, using a Sliding Window Association Test (SWAT), we ran CNN on the selected fragments to calculate phenotype influence scores (PIS) and identify phenotype-associated SNPs based on PIS. In the third step, we ran CNN on all identified SNPs to develop a classification model. We tested our approach using GWAS data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) including (N = 981; cognitively normal older adults (CN) = 650 and AD = 331). Our approach identified the well-known APOE region as the most significant genetic locus for AD. Our classification model achieved an area under the curve (AUC) of 0.82, which was compatible with traditional machine learning approaches, random forest and XGBoost. SWAT-CNN, a novel deep learning–based genome-wide approach, identified AD-associated SNPs and a classification model for AD and may hold promise for a range of biomedical applications.
Keywords: deep learning, genetic variants, Alzheimer’s disease, genome-wide association studies, phenotype influence scores
Introduction
Deep learning is a representative machine learning algorithm that enables nonlinear transformations to extract features of high-dimensional data [1], unlike traditional machine learning models that predict a linear combination of weights by assuming a linear relationship between input features and a phenotype of interest. Deep learning has been used to predict disease outputs by handling original high-dimensional medical imaging data without feature selection procedures [2, 3]. In genetic research, deep learning frameworks have been used to investigate molecular phenotypes that predict the effects of noncoding variants [4–10], differential gene expression [11] and potential transcription factor binding sites [12]. These tools use CHIP-Seq or DNase-Seq data as training data to predict chromatin features such as transcription factor binding or DNase hypersensitivity from DNA sequences. More recently, deep learning has been employed in the capture of mutations and the analysis of gene regulations, demonstrating its potential for furthering our understanding of epigenetic regulation [13]. Furthermore, deep learning is being used in gene therapy to design CRISPR guide RNAs using deep learning–based gene features [14–19].
Genome-wide association studies (GWAS) use a statistical approach by considering one single nucleotide polymorphism (SNP) at a time across the whole genome to identify population-based genetic risk variation for human diseases and traits [20, 21]. However, deep learning has not yet been used to perform GWAS, as it is challenging due to the so-called high-dimension low-sample-size (HDLSS) problem [22], which is known to impact phenotype prediction using genetic variation. Feature reduction approaches have been commonly used [23–25] to resolve this problem, but feature reduction using high-dimensional genomic data is also challenging due to an NP-hard problem [26, 27]. Therefore, it is necessary to develop a deep learning framework to identify genetic variants using whole genome data.
Here we proposed a novel three-step deep learning–based approach to select informative SNPs and develop classification models for a phenotype of interest. In the first step, we divided the whole genome into nonoverlapping fragments of an optimal size and then used deep learning algorithms to select phenotype-associated fragments containing phenotype-related SNPs. Different sized fragments and several deep learning algorithms were tested to select the optimal size for fragments and the optimal algorithm. In the second step, we ran the optimal deep learning algorithm using an overlapping Sliding Window Association Test (SWAT) within selected fragments to calculate phenotype influence scores (PIS) using SNPs and the phenotype of interest to identify informative SNPs. In the third step, we ran the optimal algorithm again on all identified informative SNPs to develop a classification model.
Alzheimer’s disease (AD), the most common form of dementia, is a neurodegenerative disorder that causes progressive deterioration of memory and cognitive function. The pathological hallmarks of the disease are toxic amyloid-β plaques and neurofibrillary tau tangles in the brain [28, 29], with the strongest genetic risk factor being the ε4 allele of apolipoprotein E (APOE). APOE ε4 allele carriers are more prone to amyloid deposition and have a 3- to 4-fold increased risk of AD [30–32]. In addition to amyloid, tau and APOE, many other aging- and neurodegeneration-associated biological pathways are being actively investigated for their role in AD pathogenesis and for their potential as targets for therapeutic development. Examples include inflammation, cellular senescence, telomere shortening, altered neurogenesis, dysregulated lipid metabolism, altered mitochondrial function and brain energetics, and other age-associated factors [33–39]. In addition, processes related to clearance of misfolded proteins are important including autophagy, the primary mechanism that removes protein aggregates [40] [41]. Relatedly, mitophagy plays an essential role in maintaining mitochondrial homeostasis and when impaired may contribute to AD-related pathophysiology [42, 43]. As amyloid-β is clearly linked to the initiation and progression of AD, it has been targeted for drug treatment. Despite longstanding global efforts and numerous failed trials, the FDA recently granted accelerated approval of aducanumab, the first potentially disease-modifying antiamyloid treatment [44]. In addition, it is of fundamental importance to identify biomarkers for the detection of AD at presymptomatic stages to slow or prevent disease progression [45–47]. In the past few years, artificial intelligence (AI) approaches have been used to identify AD biomarkers through brain image analysis [2, 3, 48], cerebrospinal fluid (CSF) AD biomarkers [49] and plasma metabolites [50]. Using high-throughput bioassays, AI technology has been used to repurpose known drugs to treat Alzheimer’s disease [51, 52]. Though these AI applications are growing rapidly, only a few have reached the clinical stage.
We tested our approach using only whole genome data for AD (N = 981; cognitively normal older adults (CN) = 650 and AD = 331). Our approach identified the known APOE region as the most significant genetic locus for AD. Using the identified region, we made a classification model with CNN. To determine if the algorithm is comparable to traditional machine learning algorithms, we also applied XGBoost and random forest. Our classification model yielded 75.2% accuracy, which was generally compatible with a modest gain in accuracy of 3.8% and 9.6% relative to XGBoost and random forest, respectively. Our classification model yielded 75.2% accuracy over traditional machine learning methods, being 3.8% and 9.6% higher than XGBoost and random forest, respectively. Our novel deep learning–based approach can identify informative SNPs and develop a classification model for AD by combining nearby SNPs and testing their aggregation.
Materials and methods
Study participants
All individuals used in the analysis were participants of the Alzheimer’s Disease Neuroimaging Initiative (ADNI) cohort [53, 54]. The ADNI initial phase (ADNI-1) was launched in 2003 to test whether serial magnetic resonance imaging (MRI), position emission tomography (PET), other biological markers, and clinical and neuropsychological assessment could be combined to measure the progression of mild cognitive impairment (MCI) and early AD. ADNI-1 has been extended in subsequent phases (ADNI-GO, ADNI-2 and ADNI-3) for follow-up of existing participants and additional new enrollments. Demographic information, APOE and whole genome genotyping data, and clinical information are publicly available from the ADNI data repository (www.loni.usc.edu/ADNI/). Informed consent was obtained for all subjects, and the study was approved by the relevant institutional review board at each data acquisition site.
Genotyping and imputation
ADNI participants were genotyped using several Illumina genotyping platforms including Illumina Human610-Quad BeadChip, Illumina HumanOmni Express BeadChip and Illumina HumanOmni 2.5M BeadChip [54]. As ADNI used different genotyping platforms, we performed quality control procedures (QC) on each genotyping platform data separately and then imputed un-genotyped single nucleotide polymorphisms (SNPs) separately using MACH and the Haplotype Reference Consortium (HRC) data as a reference panel [55]. Before imputation, we performed QC for samples and SNPs as described previously: (1) for SNP, SNP call rate < 95%, Hardy–Weinberg P value < 1 × 10−6, and minor allele frequency (MAF) < 1% and (2) for sample, sex inconsistencies, and sample call rate < 95% [56]. Furthermore, in order to prevent spurious associations due to population stratification, we selected only non-Hispanic participants of European ancestry that clustered with HapMap CEU (Utah residents with Northern and Western European ancestry from the CEPH collection) or TSI (Toscani in Italia) populations using multidimensional scaling (MDS) analysis and the HapMap genotype data [56, 57]. After imputation, we performed standard QC on imputed genotype data as described previously [58]. Specifically, we imposed an r2 value equal to 0.30 as the threshold to accept the imputed genotypes. In the study, imputed genome-wide genotyping data from 981 ADNI non-Hispanic participants (650 cognitive normal older adults (CN) and 331 AD patients) were used with a total of 5,398,183 SNPs (minor allele frequency (MAF) > 5%).
Genome-wide association study (GWAS)
Using imputed genotypes, a GWAS for AD was conducted. For the GWAS, logistic regression with age and sex as covariates was performed using PLINK [59] to determine the association of each SNP with AD. To adjust for multiple testing, a conservative threshold for genome-wide significant association (P < 5 × 10−8) was employed based on a Bonferroni correction.
Fragmentation of whole genome data
Whole genome data for 981 participants were divided into nonoverlapping fragments of varying sizes from 10 SNPs to 200 SNPs to determine the optimal fragmentation size. The subdatasets consisting of fragments of the same size were divided into train–test–validation sets (60:20:20), and convolutional neural network (CNN) [60], long short-term memory (LSTM) [61], LSTM-CNN [62] and attention [63] algorithms were applied to each. Early stopping using a validation set was applied to prevent overfitting, followed by the measurement of training time and accuracy (ACC).
Deep learning on fragments
Deep learning is the result of continuous development such as perceptron [64, 65], which adds the concept of weight adjustment to the theory that it can behave like a human brain when neurons with on–off functions are connected in a network form [66], and Adaline [67], which uses gradient descent to update weights. These early neural nets were advanced to a multilayer perceptron, which includes hidden layers to solve the famous XOR problem [68], marking a theoretical turning point with the concept of backpropagation to update the weight of the hidden layer [69–72]. The inherent problem of backpropagation, in which vanishing gradients occur when there are many layers [73], has been alleviated through activation functions, such as sigmoid function and ReLU [74, 75], as well as optimization methods for better gradient descent methods, such as Ada-Grad [76], RMSprop [77] and Adam [78]. These developments, along with the advancement of GPU hardware, have created an era of deep learning as it is now.
Deep learning has laid the theoretical foundation for backpropagation, the application of activation functions and the development of optimizers for better gradient descent. Common deep learning algorithms, such as CNN, LSTM and attention, have a hierarchical structure that implements an enhanced version of the basic principles of deep learning. The detailed technical description of each algorithm is described extensively in the relevant paper, so here we focus on the core of the deep learning technology commonly applied to the algorithm used in the experiment.
We used ReLU as an activation function that underlies the deep learning algorithms used in our experiments. ReLU, the most used activation function in the deep learning community, replaces the given value with zero if the value is <0 and uses the given value if it is >0. Thus, if the given value is greater than zero, the derivative becomes one, and the weight can be adjusted without vanishing the gradient to the first layer through the hidden layer. We used Adam as the optimization method. Adam is currently the most popular optimization method for deep learning, as it takes advantage of momentum SGD [79] and RMSprop, which are expressed as follows: Gt is the sum of the square of the modified gradient and ε is a very small constant that prevents the equation from being divided by zero.
Backpropagation is used to calculate the initial error value from a given random weight using the least squares method and then to update the weight using a chain rule until the differential value becomes zero. Here, the differential value of zero means that the weight does not change when the gradient is subtracted from the previous weight.
If yo1 and yo2 are the output values of the output layer coming through the hidden layer and the actual values of the given data are yt1 and yt2, the partial derivative of the error ErrorYo to the weight of the output layer can be calculated using the chain rule as follows:
The partial derivative of the error ErrorYo to the weight of the hidden layer can be calculated as follows:
- (a)
- (b)
Accordingly, the weight wh of the hidden layer is updated as follows:
Calculation of phenotype influence score using deep learning
Prediction accuracy was calculated from deep learning applied to each fragment and converted to a z-score. The z-score follows a normal distribution with μ = 1 and σ = 0, under the hypothesis that there is no relationship between the variables in the population. Fragments with a z-score higher than the median were selected. An overlapping SWAT for the calculation of PIS is applied to these fragments. When the length of the fragment is w, the window is positioned w − 1 from the first SNP of the fragment and moves by one SNP and stops at the last SNP of the fragment. Each region within the SWAT is divided into a train–test–validation set (60:20:20), and early stopping using a validation set is applied to prevent overfitting. When the kth SNP is Sk, PIS is calculated as follows.
This SWAT is applied to all selected fragments, resulting in a PIS for all SNPs.
Phenotype classification using deep learning
We selected the top 100 to 10,000 SNPs based on the PIS. For the AD-CN classification, we used a CNN consisting of convolution layers with a kernel size of 5, pooling the layer with a max-pool size of 2, a fully connected layer of 64 nodes, and an output layer with a softmax activation function. Due to gradient vanishing and explosion issues caused by the repeated multiplication of the recurrent weight matrix, RNN or its variants had difficulty training. In order to compare the performance, we also applied random forest and XGBoost, which are traditionally used for tabular data classification. XGBoost is a tree-based ensemble algorithm, one of popular implementations of gradient boosting. We trained XGBoost using an ‘xgboost’ package for python (https://xgboost.readthedocs.io/). Random forest is another ensemble learning method that uses many decision trees as its classifiers [80, 81]. We trained random forest using the scikit-learn package for python by setting the number of trees as 10 and the maximum depth of each tree as 3.
Results
Our deep learning–based approach consists of three steps to select informative SNPs and develop an accurate classification model. In the first step, we divided the whole genome into nonoverlapping fragments of an optimal size. To choose an optimal fragment size and an optimal deep learning algorithm, we calculated the mean accuracy and computation time for classification of AD using various fragment sizes containing 10 to 200 SNPs and several deep learning algorithms (CNN, LSTM, LSTM-CNN, Attention). In this analysis, we used 10–200 SNPs located within a region surrounding the APOE gene, the strongest and most robust AD genetic risk locus. Figure 1 shows the average accuracy and computation time for CNN, LSTM, LSTM-CNN and attention as a function of the fragment size. As shown in Figure 1A, the analysis yielded the highest accuracy for classification of AD for a fragment size with 40 SNPs (Figure 1A). Figure 1B shows the average accuracy and time as a function of the deep learning algorithm on window size of 40 within a region surrounding the APOE gene. CNN and LSTM-CNN models had the highest accuracy for classification of AD, followed by LSTM. However, the computation time of CNN and LSTM models were 5.9 and 40.4 s, respectively. The computation time of LSTM, LSTM-CNN and attention models sharply increased compared to CNN because the fragment contains more SNPs; therefore, we chose a fragment with 40 SNPs as the optimal fragment size for the CNN and optimal deep learning algorithm, respectively. The whole genome was divided into 134,955 fragments, each with 40 SNPs. We ran CNN on each fragment to calculate z-scores based on classification accuracy and selected phenotype-associated fragments. We selected 1802 fragments with z-scores higher than a median z-score.
In the second step, using a SWAT, we ran CNN on the selected fragments to calculate the PIS of each SNP in the selected fragments and identify phenotype-associated SNPs based on the PIS, as shown in Figure 2. For each SNP, we calculated a mean accuracy of 40 windows, which is the PIS of the SNP. Using PIS values, we calculated the z-scores and one-tailed P-values. Figure 3 shows a Manhattan plot with the −log10 P-values on the y-axis against the SNP position in the genome on the x-axis. The SNP with the smallest P-value was rs5117 in the APOC1 gene (P-value = 1.04E−22) and rs429358 in the APOE gene (P-value of 1.41E-16). The genetic region including APOE/APOC1/TOMM40 genes is known as the strongest genetic risk locus for AD [30, 82–84]. The next highest genetic loci were located at SNX14, SNX16, BICD1, WDR72 and GLT1D1 genes.
In the third step, we ran CNN on the identified SNPs to develop an AD classification model. Table 1 shows the classification results of AD versus CN using subsets containing the top 100 to 10,000 SNPs based on PIS. For comparison with traditional machine learning methods, we used two popular algorithms as classifiers, XGBoost and random forest. The highest mean accuracy of 10 cross-validation in classifying AD from CN by CNN was 75.02% (area under the curve [AUC] of 0.8157) for a subset containing 4000 SNPs, which had a 6.3% higher accuracy than random forest for a subset containing 2000 SNPs and a 1.94% higher accuracy than XGBoost for a subset containing 1000 SNPs. When we calculated the classification accuracy of AD using only the number of APOE ε4 alleles, the classification accuracy was 66.7%, which was 8.3% lower than our method. Our CNN models outperformed two traditional machine learning models, random forest and XGBoost, in all cases as shown in Figure 4.
Table 1.
Top | Random forest | XGBoost | CNN | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Accuracy | STD(±) | AUC | STD(±) | Accuracy | STD(±) | AUC | STD(±) | Accuracy | STD(±) | AUC | STD(±) | |
100 | 66.46 | 7.79 | 0.7137 | 0.0576 | 70.24 | 2.80 | 0.7266 | 0.0281 | 68.29 | 2.87 | 0.7216 | 0.0603 |
200 | 67.18 | 3.88 | 0.7175 | 0.0386 | 67.99 | 1.52 | 0.7166 | 0.0245 | 69.52 | 5.08 | 0.7182 | 0.0494 |
300 | 66.26 | 3.80 | 0.7098 | 0.0377 | 68.20 | 3.32 | 0.7029 | 0.0272 | 70.64 | 2.20 | 0.7250 | 0.0585 |
400 | 67.58 | 4.67 | 0.7074 | 0.0428 | 69.42 | 3.43 | 0.7177 | 0.0234 | 67.99 | 4.65 | 0.7167 | 0.0412 |
500 | 67.59 | 7.79 | 0.7111 | 0.0457 | 71.05 | 2.56 | 0.7381 | 0.0325 | 71.56 | 6.58 | 0.7411 | 0.0614 |
1000 | 68.31 | 5.22 | 0.7178 | 0.0445 | 73.08 | 2.89 | 0.7407 | 0.0372 | 73.91 | 3.87 | 0.7741 | 0.0444 |
2000 | 68.70 | 3.13 | 0.7372 | 0.0424 | 72.48 | 2.61 | 0.7509 | 0.0365 | 73.29 | 2.77 | 0.7782 | 0.0409 |
3000 | 67.78 | 3.59 | 0.7282 | 0.0351 | 69.62 | 4.27 | 0.7376 | 0.0328 | 73.80 | 2.40 | 0.7862 | 0.0282 |
4000 | 68.19 | 4.69 | 0.7263 | 0.0469 | 71.15 | 4.07 | 0.7412 | 0.0368 | 75.02 | 3.17 | 0.8157 | 0.0261 |
5000 | 66.25 | 5.41 | 0.7105 | 0.0399 | 70.74 | 3.14 | 0.7330 | 0.0305 | 73.19 | 4.72 | 0.8003 | 0.0506 |
10 000 | 66.26 | 5.59 | 0.6919 | 0.0528 | 69.63 | 3.27 | 0.7248 | 0.0211 | 71.05 | 6.57 | 0.7083 | 0.1424 |
Notes: The table shows the number of top SNPs selected based on phenotype influence score for AD classification and the accuracy and AUC of 10-fold cross-validation. Our CNN-based approach yielded the highest accuracy and AUC of 75.02% and 0.8157, respectively, for 4000 SNPs. In all cases, our CNN models outperformed two traditional machine learning models, random forest and XGBoost
Figure 5 shows LocusZoom plots [85] for SNPs located at 300 kb upstream and downstream regions from the boundary of the APOE gene. The horizontal axis is the location of SNPs, and the vertical axis is −log10 of the P-values. Each dot represents a SNP, and the color represents the squared correlation coefficient (r2) with the most significant SNP. Figure 5A shows P-values calculated using PLINK, and the most significant SNP was rs429358 in APOE. Figure 5B shows P-values calculated using our deep learning approach, and the most significant SNP was rs5117 in APOC1. Figure 5B shows a linear increase on the left side of rs5117 and a linear decrease on the right side of rs5117, which was different from PLINK results (Figure 5A), which have no linear patterns. In addition, Figure 5B shows three strongly correlated SNPs (r2 > 0.8), with rs5117 on the left side of rs5117 but no SNPs on the right side of rs5117.
Discussion
Although deep learning has solved many real-world problems, few deep learning approaches have been used in GWAS or sequence data to identify genetic variants and for disease/risk classification due to the high dimensionality of the genomic data [22]. In this study, we propose a novel deep learning–based sliding window approach to identify and select disease-associated SNPs and develop an accurate classification model using high dimensional genomic data that we tested using the ADNI cohort (N = 981). The proposed method successfully identified significant genetic loci for AD that included the well-known APOE genetic locus and highlighted several other risk loci. Our deep learning–based approach was compatible to traditional machine learning methods for classification of AD.
The deep learning–based approach for identification of genetic variants consists of three steps. In the first step, we divided the whole genome into nonoverlapping fragments of an optimal size, creating a fragmentation and windowing approach that, to the best of our knowledge, is the first deep learning–based method for identifying genetic variants.
In the second step, we calculated a PIS of each SNP within the selected fragments by using an overlapping window and CNN algorithm. Our method calculates PIS, a novel index that is used to find disease-related variants and predict disease. Furthermore, we calculated the z-scores and one-tailed P-values using PIS, which yielded a Manhattan plot showing the most significant genetic loci in APOE/APOC1/TOMM40 genes that are known as the strongest genetic risk factors for AD. Our method also identified several novel candidate genetic loci, including sorting nexin (SNX) 14 and SNX16, located on chromosomes 6 and 8, respectively, that have not been previously identified to be associated with AD, though there may be special relevance for neurodegeneration as SNX12 [86], SNX17 [87], SNX27 [88] and SNX33 [89] are involved in neuronal survival. Bicaudal D1 (BICD1) on chromosome 12 is a susceptibility gene in chronic obstructive pulmonary disease [90] and lissencephaly [91], but there are no reports of it being associated with AD.
In the third step, we selected top SNPs based on PIS to develop classification models for AD. We selected sets of highly AD-related SNPs and classified AD from CN using CNN, as well as two popular traditional machine learning algorithms, XGBoost and random forest. We found the accuracy of classification was changed with the number of the selected SNPs and the classification algorithms. The highest mean accuracy of the classification was 75.0% when CNN was used on the top 4000 SNPs, which was comparable to two traditional machine learning algorithms. It was also 8.3% higher than the accuracy of the classification using only the number of APOE ε4 alleles. Classification is the first step toward achieving a better understanding of the genetic architecture of AD. The proposed method will benefit from future studies that use deep learning with quantitative phenotypes and baseline values to predict future disease trajectories.
We plotted the SNPs selected by PIS and PLINK for comparison using LocusZoom. We found that there were no SNPs with r2 greater than 0.8 in the PLINK results, but three strongly associated SNPs were identified using our method. This is because the PLINK method finds statistical significance SNP by SNP, whereas the deep learning approach uses multiple inputs to adjust weights through the training process. Deep learning uses adjacent SNPs to compute gradients at every epoch and uses a loss function to adjust the weights in the backpropagation process. Unlike PLINK, our method shows that SNPs related to a phenotype can be extracted by considering surrounding SNPs, which means that both methods might be complementary because they identify different variants, though notably in the same region around APOE.
In summary, our novel deep learning–based approach can identify AD-related SNPs by using genome-wide data to develop a classification model for AD. The heritability of AD is estimated to be up to 80%. Accordingly, it is important to identify novel genetic loci related to the disease. Using a modest sample size, we found a significant genetic locus and a classification accuracy of 75%. In future work, we plan to apply our method to large-scale whole genome sequencing datasets that are expected to become available soon to identify novel AD-related SNPs and develop more accurate classification models. We also plan to study early stages of disease including mild cognitive impairment and subjective cognitive decline, where there is considerable heterogeneity, and more refined classification of risk for progression to AD would be valuable. In addition, future studies will investigate use of quantitative endophenotypes that may be more informative than binary classification given their potential to elucidate genetic risk related to specific disease pathways and mechanisms.
Key Points
Although deep learning has been successfully applied to many scientific fields, deep learning has not been used in genome-wide association studies (GWAS) in practice due to the high dimensionality of genomic data.
To overcome this challenge, we propose a novel three-step approach (SWAT-CNN) for identification of genetic variants using deep learning to identify phenotype-related single nucleotide polymorphisms (SNPs) that can be applied to develop accurate disease classification models.
To accomplish this, we divided the whole genome into nonoverlapping fragments of an optimal size and ran a deep learning algorithm on each fragment to select disease-associated fragments.
We calculated phenotype influence scores (PIS) of each SNP within selected fragments to identify disease-associated significant SNPs and developed a disease classification model by using overlapping window and deep learning algorithms.
In the application of our method to Alzheimer’s disease (AD), we identified well-known significant genetic loci for AD and achieved higher classification accuracies than traditional machine learning methods.
Acknowledgements
Data collection and sharing for this project was funded by the Alzheimer’s Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01AG024904) and DODADNI (Department of Defense award number W81XWH-12-2-0012). ADNI is funded by the National Institute on Aging, the National Institute of Biomedical Imaging and Bioengineering, and through generous contributions from the following: AbbVie, Alzheimer’s Association; Alzheimer’s Drug Discovery Foundation; Araclon Biotech; BioClinica, Inc.; Biogen; Bristol-Myers Squibb Company; CereSpir, Inc.; Cogstate; Eisai Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and Company; EuroImmun; F. Hoffmann-La Roche Ltd and its affiliated company Genentech, Inc.; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda Pharmaceutical Company and Transition Therapeutics. The Canadian Institutes of Health Research is providing funds to support ADNI clinical sites in Canada. Private sector contributions are facilitated by the Foundation for the National Institutes of Health (http://www.fnih.org). The grantee organization is the Northern California Institute for Research and Education, and the study is coordinated by the Alzheimer’s Therapeutic Research Institute at the University of Southern California. ADNI data are disseminated by the Laboratory for Neuro Imaging at the University of Southern California.
Taeho Jo is an assistant research professor in the Department of Radiology and Imaging Sciences at the Indiana University School of Medicine. His research focuses on developing machine learning, deep learning methods to analyze large-scale biomedical data, including neuroimaging, genetics and other omics modalities.
Kwangsik Nho is an associate professor in the Department of Radiology and Imaging Sciences at the Indiana University School of Medicine. His main research focus is on computational strategies for integrative analysis of genomic, transcriptomic and metabolomic data in relation to multimodal MRI and PET neuroimaging, as applied to Alzheimer’s disease.
Paula Bice is an assistant research professor in the Department of Radiology and Imaging Sciences and the Indiana Alzheimer’s Disease Research Center, at the Indiana University School of Medicine. As a neuroscientist, her research focuses on studies using neuroimaging, omics and systems biology approaches.
Andrew J. Saykin is the Raymond C. Beeler Professor of Radiology and Imaging Sciences and of Medical and Molecular Genetics at the Indiana University School of Medicine where he directs the NIA-designated Indiana Alzheimer’s Disease Research Center. His research focuses on integrative analysis of multimodal imaging, omics and other biomarker data in support of precision medicine.
Contributor Information
Taeho Jo, Department of Radiology and Imaging Sciences, Center for Neuroimaging, Indiana University School of Medicine, Indianapolis, IN, USA; Indiana Alzheimer’s Disease Research Center, Indiana University School of Medicine, Indianapolis, IN, USA; Indiana University Network Science Institute, Bloomington, IN, USA.
Kwangsik Nho, Department of Radiology and Imaging Sciences, Center for Neuroimaging, Indiana University School of Medicine, Indianapolis, IN, USA; Indiana Alzheimer’s Disease Research Center, Indiana University School of Medicine, Indianapolis, IN, USA; Indiana University Network Science Institute, Bloomington, IN, USA; Center for Computational Biology and Bioinformatics, Indiana University School of Medicine, Indianapolis, IN, USA.
Paula Bice, Department of Radiology and Imaging Sciences, Center for Neuroimaging, Indiana University School of Medicine, Indianapolis, IN, USA; Indiana Alzheimer’s Disease Research Center, Indiana University School of Medicine, Indianapolis, IN, USA.
Andrew J Saykin, Department of Radiology and Imaging Sciences, Center for Neuroimaging, Indiana University School of Medicine, Indianapolis, IN, USA; Indiana Alzheimer’s Disease Research Center, Indiana University School of Medicine, Indianapolis, IN, USA; Indiana University Network Science Institute, Bloomington, IN, USA; Department of Medical and Molecular Genetics, Indiana University School of Medicine, Indianapolis, IN, USA.
Author contributions
T.J., K.N. and A.S.: Conceptualization. A.S.: Acquisition of Data and Interpretation of Results. T.J., K.N.: Data Curation. T.J.: Formal Analysis, Investigation, Methodology, Validation, Visualization. T.J., K.N., P.B. and A.S.: Writing—Original Draft Preparation. T.J., K.N., P.B. and A.S.: Review & Editing.
Funding
Data collection and sharing for this project was funded by the Alzheimer’s Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01 AG024904) and DOD ADNI (Department of Defense award number W81XWH-12-2-0012). In addition, this work was supported, in part, by grants from the National Institutes of Health (NIH) and includes the following sources: P30 AG010133, P30 AG072976, R01 AG019771, R01 AG057739, U01 AG024904, R01 LM013463, R01 AG068193, T32 AG071444, U01 AG068057, R01 LM012535 and R03 AG063250. The funders played no role in the design of the study, analysis and interpretation of the data or in writing the manuscript.
References
- 1. Wainberg M, Merico D, Delong A, et al. Deep learning in biomedicine. Nat Biotechnol 2018;36:829–38. [DOI] [PubMed] [Google Scholar]
- 2. Jo T, Nho K, Saykin AJ. Deep learning in Alzheimer’s disease: diagnostic classification and prognostic prediction using neuroimaging data. Front Aging Neurosci 2019;11:220. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Jo T, Nho K, Risacher SL, et al. Deep learning detection of informative features in tau PET for Alzheimer’s disease classification. BMC Bioinformatics 2020;21:496. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Zhang Z, Park CY, Theesfeld CL, et al. An automated framework for efficiently designing deep convolutional neural networks in genomics. Nature Machine Intelligence 2021;3:392–400. [Google Scholar]
- 5. Zhou J, Troyanskaya OG. Predicting effects of noncoding variants with deep learning–based sequence model. Nat Methods 2015;12:931–4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Alipanahi B, Delong A, Weirauch MT, et al. Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning. Nat Biotechnol 2015;33:831–8. [DOI] [PubMed] [Google Scholar]
- 7. Xiong HY, Alipanahi B, Lee LJ, et al. The human splicing code reveals new insights into the genetic determinants of disease. Science 2015;347:1254806. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Quang D, Xie X. DanQ: a hybrid convolutional and recurrent deep neural network for quantifying the function of DNA sequences. Nucleic Acids Res 2016;44:e107–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Angermueller C, Lee HJ, Reik W, et al. DeepCpG: accurate prediction of single-cell DNA methylation states using deep learning. Genome Biol 2017;18:67. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Zhang S, Hu H, Jiang T, et al. TITER: predicting translation initiation sites by deep learning. Bioinformatics 2017;33:i234–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Tasaki S, Gaiteri C, Mostafavi S, et al. Deep learning decodes the principles of differential gene expression. Nat Mach Intell 2020;2:376–86. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Zheng A, Lamkin M, Zhao H, et al. Deep neural networks identify sequence context features predictive of transcription factor binding. Nat Mach Intell 2021;3:172–80. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Scherer M, Schmidt F, Lazareva O, et al. Machine learning for deciphering cell heterogeneity and gene regulation. Nat Comput Sci 2021;1:183–91. [DOI] [PubMed] [Google Scholar]
- 14. Listgarten J, Weinstein M, Kleinstiver BP, et al. Prediction of off-target activities for the end-to-end design of CRISPR guide RNAs. Nat Biomed Eng 2018;2:38–47. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Shen MW, Arbab M, Hsu JY, et al. Predictable and precise template-free CRISPR editing of pathogenic variants. Nature 2018;563:646–51. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Leenay RT, Aghazadeh A, Hiatt J, et al. Large dataset enables prediction of repair after CRISPR–Cas9 editing in primary T cells. Nat Biotechnol 2019;37:1034–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Liu Q, He D, Xie L. Prediction of off-target specificity and cell-specific fitness of CRISPR-Cas system using attention boosted deep learning and network-based gene feature. PLoS Comput Biol 2019;15:e1007480. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Kim HK, Min S, Song M, et al. Deep learning improves prediction of CRISPR–Cpf1 guide RNA activity. Nat Biotechnol 2018;36:239–41. [DOI] [PubMed] [Google Scholar]
- 19. Ogden PJ, Kelsic ED, Sinai S, et al. Comprehensive AAV capsid fitness landscape reveals a viral gene and enables machine-guided design. Science 2019;366:1139. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Yan J, Qiu Y, Ribeiro dos Santos AM, et al. Systematic analysis of binding of transcription factors to noncoding variants. Nature 2021;591:147–51. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Buniello A, MacArthur JAL, Cerezo M, et al. The NHGRI-EBI GWAS catalog of published genome-wide association studies, targeted arrays and summary statistics 2019. Nucleic Acids Res 2019;47:D1005–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Auton A, Abecasis GR, Altshuler DM, et al. A global reference for human genetic variation. Nature 2015;526:68–74. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Li F, Yang Y, Xing EP. From Lasso regression to feature vector machine. In: Proceedings of the 18th International Conference on Neural Information Processing Systems. 2005, pp. 779–86. MIT Press, Vancouver, British Columbia, Canada. [Google Scholar]
- 24. Yamada M, Jitkrittum W, Sigal L, et al. High-dimensional feature selection by feature-wise Kernelized lasso. Neural Comput 2014;26:185–207. [DOI] [PubMed] [Google Scholar]
- 25. Xu Z, Huang G, Weinberger KQ, Zheng AX. Gradient boosted feature selection. In: Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining. 2014, pp. 522–31. Association for Computing Machinery, New York, NY, USA. [Google Scholar]
- 26. Amaldi E, Kann V. On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems. Theor Comput Sci 1998;209:237–60. [Google Scholar]
- 27. Guyon I, Elisseeff A. An introduction to variable and feature selection. J Mach Learn Res 2003;3:1157–82. [Google Scholar]
- 28. Canter RG, Penney J, Tsai L-H. The road to restoring neural circuits for the treatment of Alzheimer's disease. Nature 2016;539:187–96. [DOI] [PubMed] [Google Scholar]
- 29. Hyman BT, Phelps CH, Beach TG, et al. National Institute on Aging–Alzheimer's Association guidelines for the neuropathologic assessment of Alzheimer's disease. Alzheimers Dement 2012;8:1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Corder E, Saunders A, Strittmatter W, et al. Gene dose of apolipoprotein E type 4 allele and the risk of Alzheimer's disease in late onset families. Science 1993;261:921–3. [DOI] [PubMed] [Google Scholar]
- 31. Morris JC, Roe CM, Xiong C, et al. APOE predicts amyloid-beta but not tau Alzheimer pathology in cognitively normal aging. Ann Neurol 2010;67:122–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Farrer LA, Cupples LA, Haines JL, et al. Effects of age, sex, and ethnicity on the association between apolipoprotein E genotype and Alzheimer disease: a meta-analysis. JAMA 1997;278:1349–56. [PubMed] [Google Scholar]
- 33. Lautrup S, Sinclair DA, Mattson MP, et al. NAD+ in brain aging and neurodegenerative disorders. Cell Metab 2019;30:630–55. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Horgusluoglu E, Nudelman K, Nho K, et al. Adult neurogenesis and neurodegenerative diseases: a systems biology perspective. Am J Med Genet B Neuropsychiatr Genet 2017;174:93–112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Felsky D, Roostaei T, Nho K, et al. Neuropathological correlates and genetic architecture of microglial activation in elderly human brain. Nat Commun 2019;10:1–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Lee Y-J, Han SB, Nam S-Y, et al. Inflammation and Alzheimer’s disease. Arch Pharm Res 2010;33:1539–56. [DOI] [PubMed] [Google Scholar]
- 37. Mahoney ER, Dumitrescu L, Seto M, et al. Telomere length associations with cognition depend on Alzheimer's disease biomarkers. Alzheimers Dement Transl Res Clin Interv 2019;5:883–90. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Wong MW, Braidy N, Poljak A, et al. Dysregulation of lipids in Alzheimer's disease and their role as potential biomarkers. Alzheimers Dement 2017;13:810–27. [DOI] [PubMed] [Google Scholar]
- 39. Suzanne M, Tong M. Brain metabolic dysfunction at the core of Alzheimer's disease. Biochem Pharmacol 2014;88:548–59. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Bourdenx M, Martín-Segura A, Scrivo A, et al. Chaperone-mediated autophagy prevents collapse of the neuronal metastable proteome. Cell 2021;184:2696, e2625–714. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Liu J, Li L. Targeting autophagy for the treatment of Alzheimer’s disease: challenges and opportunities. Front Mol Neurosci 2019;12:203. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Fang EF, Hou Y, Palikaras K, et al. Mitophagy inhibits amyloid-β and tau pathology and reverses cognitive deficits in models of Alzheimer’s disease. Nat Neurosci 2019;22:401–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Kerr JS, Adriaanse BA, Greig NH, et al. Mitophagy and Alzheimer’s disease: cellular and molecular mechanisms. Trends Neurosci 2017;40:151–66. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Sevigny J, Chiao P, Bussière T, et al. The antibody aducanumab reduces Aβ plaques in Alzheimer’s disease. Nature 2016;537:50–6. [DOI] [PubMed] [Google Scholar]
- 45. Sims R, Hill M, Williams J. The multiplex model of the genetics of Alzheimer’s disease. Nat Neurosci 2020;23:311–22. [DOI] [PubMed] [Google Scholar]
- 46. Schwartzentruber J, Cooper S, Liu JZ, et al. Genome-wide meta-analysis, fine-mapping and integrative prioritization implicate new Alzheimer’s disease risk genes. Nat Genet 2021;53:392–402. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47. Chia R, Sabir MS, Bandres-Ciga S, et al. Genome sequencing analysis identifies new loci associated with Lewy body dementia and provides insights into its genetic architecture. Nat Genet 2021;53:294–303. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Ding Y, Sohn JH, Kawczynski MG, et al. A deep learning model to predict a diagnosis of Alzheimer disease by using 18F-FDG PET of the brain. Radiology 2019;290:456–64. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Stamate D, Kim M, Proitsi P, et al. A metabolite-based machine learning approach to diagnose Alzheimer-type dementia in blood: results from the European medical information framework for Alzheimer disease biomarker discovery cohort. Alzheimers Dement Transl Res Clin Interv 2019;5:933–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50. Bellomo G, Indaco A, Chiasserini D, et al. Machine learning driven profiling of cerebrospinal fluid Core biomarkers in Alzheimer’s disease and other neurological disorders. Front Neurosci 2021;15:337. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51. Zhang M, Schmitt-Ulms G, Sato C, et al. Drug repositioning for Alzheimer’s disease based on systematic ‘omics’ data mining. PloS One 2016;11:e0168812. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52. Rodriguez S, Hug C, Todorov P, et al. Machine learning identifies candidates for drug repurposing in Alzheimer’s disease. Nat Commun 2021;12:1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53. Veitch DP, Weiner MW, Aisen PS, et al. Understanding disease progression and improving Alzheimer's disease clinical trials: recent highlights from the Alzheimer's disease neuroimaging initiative. Alzheimers Dement 2019;15:106–52. [DOI] [PubMed] [Google Scholar]
- 54. Saykin AJ, Shen L, Yao X, et al. Genetic studies of quantitative MCI and AD phenotypes in ADNI: progress, opportunities, and plans. Alzheimers Dement 2015;11:792–814. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55. Park YH, Hodges A, Risacher SL, et al. Dysregulated fc gamma receptor-mediated phagocytosis pathway in Alzheimer's disease: network-based gene expression analysis. Neurobiol Aging 2020;88:24–32. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56. Horgusluoglu-Moloch E, Nho K, Risacher SL, et al. Targeted neurogenesis pathway-based gene analysis identifies ADORA2A associated with hippocampal volume in mild cognitive impairment and Alzheimer's disease. Neurobiol Aging 2017;60:92–103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57. Freedman ML, Reich D, Penney KL, et al. Assessing the impact of population stratification on genetic association studies. Nat Genet 2004;36:388–93. [DOI] [PubMed] [Google Scholar]
- 58. Park YH, Hodges A, Simmons A, et al. Association of blood-based transcriptional risk scores with biomarkers for Alzheimer disease. Neurol Genet 2020;6:e517. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59. Purcell S, Neale B, Todd-Brown K, et al. PLINK: a tool set for whole-genome association and population-based linkage analyses. Am J Hum Genet 2007;81:559–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 2012;25:1097–105. [Google Scholar]
- 61. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput 1997;9:1735–80. [DOI] [PubMed] [Google Scholar]
- 62. Zhang J, Li Y, Tian J, Li T. LSTM-CNN hybrid model for text classification. In: 2018 IEEE 3rd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC);. 2018, 1675–80.
- 63. Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need. arXiv preprint arXiv:1706.03762 2017. [Google Scholar]
- 64. Rosenblatt F. The Perceptron, A Perceiving and Recognizing Automaton. Technical Report 85–460-1, Cornell Aeronautical Laboratory, Buffalo, New York, 1957. [Google Scholar]
- 65. Rosenblatt F. The perceptron: a probabilistic model for information storage and organization in the brain. Psychol Rev 1958;65:386. [DOI] [PubMed] [Google Scholar]
- 66. McCulloch WS, Pitts W. A logical calculus of the ideas immanent in nervous activity. Bull Math Biophys 1943;5:115–33. [PubMed] [Google Scholar]
- 67. Widrow B, Hoff ME. Adaptive Switching Circuits. Stanford Univ Ca Stanford Electronics Labs, 1960. WESCON Convention Record Part IV, 96–104. [Google Scholar]
- 68. Minsky M, Papert SA. Perceptrons: An Introduction to Computational Geometry. MIT press, 1969; [Google Scholar]
- 69. Werbos PJ. Applications of advances in nonlinear sensitivity analysis. In: System Modeling and Optimization. Springer, Berlin, Heidelberg, 1982, vol 38, p 762–70. [Google Scholar]
- 70. Werbos PJ. Backwards differentiation in AD and neural nets: past links and new opportunities. In: Automatic Differentiation: Applications, Theory, and Implementations, Springer, Berlin, Heidelberg, 2006, vol 50, p 15–34.
- 71. Rumelhart DE, Hinton GE, Williams RJ. Learning representations by back-propagating errors. Nature 1986;323:533–6. [Google Scholar]
- 72. LeCun Y, Touresky D, Hinton G, Sejnowski T. A theoretical framework for back-propagation. In: Proceedings of the 1988 Connectionist Models Summer School, CMU, Pittsburg, PA, 1988: 21–8.
- 73. Goodfellow I, Bengio Y, Courville A, et al. Deep Learning. MIT press Cambridge, 2016. [Google Scholar]
- 74. Nair V, Hinton GE. Rectified linear units improve restricted boltzmann machines. Icml 2010:807–814. [Google Scholar]
- 75. Glorot X, Bordes A, Bengio Y. Deep sparse rectifier neural networks. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics; 2011;15:315–23. [Google Scholar]
- 76. Duchi J, Hazan E, Singer Y. Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res 2011;12:2121–59. [Google Scholar]
- 77. Hinton G, Srivastava N, Swersky K. Neural networks for machine learning lecture 6a overview of mini-batch gradient descent. Coursera Lect Slides. 2012;14. [Google Scholar]
- 78. Kingma DP, Adam BJ. Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980 2014.
- 79. Sutskever I, Martens J, Dahl G, Hinton G. On the importance of initialization and momentum in deep learning. In: Proceedings of the 30th International Conference on Machine Learning (ICML-13), 2013: 1139–47.
- 80. Breiman L. Random forests. Mach Learn 2001;45:5–32. [Google Scholar]
- 81. Jo T, Cheng J. Improving protein fold recognition by random forest. BMC Bioinformatics 2014;15:S14. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82. Saunders AM, Strittmatter WJ, Schmechel D, et al. Association of apolipoprotein E allele ϵ4 with late-onset familial and sporadic Alzheimer's disease. Neurology 1993;43:1467–7. [DOI] [PubMed] [Google Scholar]
- 83. Roses AD, Lutz MW, Amrine-Madsen H, et al. A TOMM40 variable-length polymorphism predicts the age of late-onset Alzheimer's disease. Pharmacogenomics J 2010;10:375–84. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84. Cervantes S, Samaranch L, Vidal-Taboada JM, et al. Genetic variation in APOE cluster region and Alzheimer's disease risk. Neurobiol Aging 2011;32:2107.e2107–17. [DOI] [PubMed] [Google Scholar]
- 85. Pruim RJ, Welch RP, Sanna S, et al. LocusZoom: regional visualization of genome-wide association scan results. Bioinformatics 2010;26:2336–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86. Zhao Y, Wang Y, Yang J, et al. Sorting nexin 12 interacts with BACE1 and regulates BACE1-mediated APP processing. Mol Neurodegener 2012;7:1–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87. Lee J, Retamal C, Cuitiño L, et al. Adaptor protein sorting nexin 17 regulates amyloid precursor protein trafficking and processing in the early endosomes. J Biol Chem 2008;283:11501–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88. Gallon M, Clairfeuille T, Steinberg F, et al. A unique PDZ domain and arrestin-like fold interaction reveals mechanistic details of endocytic recycling by SNX27-retromer. Proc Natl Acad Sci 2014;111:E3604–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89. Heiseke A, Schöbel S, Lichtenthaler SF, et al. The novel sorting nexin SNX33 interferes with cellular PrPSc formation by modulation of PrPc shedding. Traffic 2008;9:1116–29. [DOI] [PubMed] [Google Scholar]
- 90. Mercado N, Colley T, Baker JR, et al. Bicaudal D1 impairs autophagosome maturation in chronic obstructive pulmonary disease. FASEB BioAdv 2019;1:688–705. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91. Swan A, Nguyen T, Suter B. Drosophila Lissencephaly-1 functions with Bic-D and dynein in oocyte determination and nuclear positioning. Nat Cell Biol 1999;1:444–9. [DOI] [PubMed] [Google Scholar]